Mercurial > hg
view tests/test-filecache.py @ 18728:1e636f7b1cfe stable
largefiles: simplify cachelfiles - don't spend a lot of time checking hashes
cachelfiles jumped through loops to handle merges and modified files ... but it
did apparently no longer have a valid reason to do so. It should just always
make sure that the largefiles referenced from the standins are present - no
matter which actual largefile is stored in the working directory. If there is
no standin then there is nothing to fetch.
The old code usually verified the hash of all largefiles every time this
function was invoked - for examply by 'update'.
This change makes a trivial noop update 5-10 seconds faster on our repo (with
the other 50% spent doing another unnecessary hashing of all largefiles).
author | Mads Kiilerich <madski@unity3d.com> |
---|---|
date | Thu, 28 Feb 2013 13:45:18 +0100 |
parents | f36375576ed5 |
children | ed80cecdfc57 |
line wrap: on
line source
import sys, os, subprocess if subprocess.call(['python', '%s/hghave' % os.environ['TESTDIR'], 'cacheable']): sys.exit(80) from mercurial import util, scmutil, extensions, hg, ui filecache = scmutil.filecache class fakerepo(object): def __init__(self): self._filecache = {} def join(self, p): return p def sjoin(self, p): return p @filecache('x') def cached(self): print 'creating' def invalidate(self): for k in self._filecache: try: delattr(self, k) except AttributeError: pass def basic(repo): # file doesn't exist, calls function repo.cached repo.invalidate() # file still doesn't exist, uses cache repo.cached # create empty file f = open('x', 'w') f.close() repo.invalidate() # should recreate the object repo.cached f = open('x', 'w') f.write('a') f.close() repo.invalidate() # should recreate the object repo.cached repo.invalidate() # stats file again, nothing changed, reuses object repo.cached # atomic replace file, size doesn't change # hopefully st_mtime doesn't change as well so this doesn't use the cache # because of inode change f = scmutil.opener('.')('x', 'w', atomictemp=True) f.write('b') f.close() repo.invalidate() repo.cached def fakeuncacheable(): def wrapcacheable(orig, *args, **kwargs): return False def wrapinit(orig, *args, **kwargs): pass originit = extensions.wrapfunction(util.cachestat, '__init__', wrapinit) origcacheable = extensions.wrapfunction(util.cachestat, 'cacheable', wrapcacheable) try: os.remove('x') except OSError: pass basic(fakerepo()) util.cachestat.cacheable = origcacheable util.cachestat.__init__ = originit def test_filecache_synced(): # test old behaviour that caused filecached properties to go out of sync os.system('hg init && echo a >> a && hg ci -qAm.') repo = hg.repository(ui.ui()) # first rollback clears the filecache, but changelog to stays in __dict__ repo.rollback() repo.commit('.') # second rollback comes along and touches the changelog externally # (file is moved) repo.rollback() # but since changelog isn't under the filecache control anymore, we don't # see that it changed, and return the old changelog without reconstructing # it repo.commit('.') def setbeforeget(repo): os.remove('x') repo.cached = 0 repo.invalidate() print repo.cached repo.invalidate() f = open('x', 'w') f.write('a') f.close() print repo.cached print 'basic:' print basic(fakerepo()) print print 'fakeuncacheable:' print fakeuncacheable() test_filecache_synced() print print 'setbeforeget:' print setbeforeget(fakerepo())