tests/test-merge-symlinks.t
author Na'Tosha Bard <natosha@unity3d.com>
Mon, 13 Feb 2012 18:37:07 +0100
changeset 16120 47ee41fcf42b
parent 15441 e0348815e806
child 16907 91dc9587862e
permissions -rw-r--r--
largefiles: optimize update speed by only updating changed largefiles Historically, during 'hg update', every largefile in the working copy was hashed (which is a very expensive operation on big files) and any largefiles that did not have a hash that matched their standin were updated. This patch optimizes 'hg update' by keeping track of what standins have changed between the old and new revisions, and only updating the largefiles that have changed. This saves a lot of time by avoiding the unecessary calculation of a list of sha1 hashes for big files. With this patch, the time 'hg update' takes to complete is a function of how many largefiles need to be updated and what their size is. Performance tests on a repository with about 80 largefiles ranging from a few MB to about 97 MB are shown below. The tests show how long it takes to run 'hg update' with no changes actually being updated. Mercurial 2.1 release: $ time hg update 0 files updated, 0 files merged, 0 files removed, 0 files unresolved getting changed largefiles 0 largefiles updated, 0 removed real 0m10.045s user 0m9.367s sys 0m0.674s With this patch: $ time hg update 0 files updated, 0 files merged, 0 files removed, 0 files unresolved real 0m0.965s user 0m0.845s sys 0m0.115s The same repsoitory, without the largefiles extension enabled: $ time hg update 0 files updated, 0 files merged, 0 files removed, 0 files unresolved real 0m0.799s user 0m0.684s sys 0m0.111s So before the patch, 'hg update' with no changes was approximately 9.25s slower with largefiles enabled. With this patch, it is approximately 0.165s slower.

  $ "$TESTDIR/hghave" symlink || exit 80

  $ cat > echo.py <<EOF
  > #!/usr/bin/env python
  > import os, sys
  > try:
  >     import msvcrt
  >     msvcrt.setmode(sys.stdout.fileno(), os.O_BINARY)
  >     msvcrt.setmode(sys.stderr.fileno(), os.O_BINARY)
  > except ImportError:
  >     pass
  > 
  > for k in ('HG_FILE', 'HG_MY_ISLINK', 'HG_OTHER_ISLINK', 'HG_BASE_ISLINK'):
  >     print k, os.environ[k]
  > EOF

Create 2 heads containing the same file, once as
a file, once as a link. Bundle was generated with:

# hg init t
# cd t
# echo a > a
# hg ci -qAm t0 -d '0 0'
# echo l > l
# hg ci -qAm t1 -d '1 0'
# hg up -C 0
# ln -s a l
# hg ci -qAm t2 -d '2 0'
# echo l2 > l2
# hg ci -qAm t3 -d '3 0'

  $ hg init t
  $ cd t
  $ hg -q pull "$TESTDIR/bundles/test-merge-symlinks.hg"
  $ hg up -C 3
  3 files updated, 0 files merged, 0 files removed, 0 files unresolved

Merge them and display *_ISLINK vars
merge heads

  $ hg merge --tool="python ../echo.py"
  merging l
  HG_FILE l
  HG_MY_ISLINK 1
  HG_OTHER_ISLINK 0
  HG_BASE_ISLINK 0
  0 files updated, 1 files merged, 0 files removed, 0 files unresolved
  (branch merge, don't forget to commit)

Test working directory symlink bit calculation wrt copies,
especially on non-supporting systems.
merge working directory

  $ hg up -C 2
  1 files updated, 0 files merged, 1 files removed, 0 files unresolved
  $ hg copy l l2
  $ HGMERGE="python ../echo.py" hg up 3
  merging l2
  HG_FILE l2
  HG_MY_ISLINK 1
  HG_OTHER_ISLINK 0
  HG_BASE_ISLINK 0
  0 files updated, 1 files merged, 0 files removed, 0 files unresolved