Mercurial > hg
view hgext/lfs/__init__.py @ 40811:e13ab4acf555
rust: peek_mut optim for lazy ancestors
This is one of the two optimizations that are also
present in the Python code: replacing pairs of pop/push
on the BinaryHeap by single updates, hence having it
under the hood maintain its consistency (sift) only once.
On Mozilla central, the measured gain (see details below)
is around 7%.
Creating the PeekMut object by calling peek_mut() right away
instead of peek() first is less efficient (gain is only 4%, stats
not included).
Our interpretation is that its creation has a cost which is vasted
in the cases where it ends by droping the value (Peekmut::pop()
just does self.heap.pop() anyway). On the other hand, the immutable
peek() is very fast: it's just taking a reference in the
underlying vector.
The Python version still has another optimization:
if parent(current) == current-1, then the heap doesn't need
to maintain its consistency, since we already know that
it's bigger than all the others in the heap.
Rust's BinaryHeap doesn't allow us to mutate its biggest
element with no housekeeping, but we tried it anyway, with a
copy of the BinaryHeap implementation with a dedicaded added
method: it's not worth the technical debt in our opinion
(we measured only a further 1.6% improvement).
One possible explanation would be that the sift is really fast
anyway in that case, whereas it's not in the case of Python,
because it's at least partly done in slow Python code.
Still it's possible that replacing BinaryHeap by something more
dedicated to discrete ordered types could be faster.
Measurements on mozilla-central:
Three runs of 'hg perfancestors' on the parent changeset:
Moyenne des médianes: 0.100587
! wall 0.100062 comb 0.100000 user 0.100000 sys 0.000000 (best of 98)
! wall 0.135804 comb 0.130000 user 0.130000 sys 0.000000 (max of 98)
! wall 0.102864 comb 0.102755 user 0.099286 sys 0.003469 (avg of 98)
! wall 0.101486 comb 0.110000 user 0.110000 sys 0.000000 (median of 98)
! wall 0.096804 comb 0.090000 user 0.090000 sys 0.000000 (best of 100)
! wall 0.132235 comb 0.130000 user 0.120000 sys 0.010000 (max of 100)
! wall 0.100258 comb 0.100300 user 0.096000 sys 0.004300 (avg of 100)
! wall 0.098384 comb 0.100000 user 0.100000 sys 0.000000 (median of 100)
! wall 0.099925 comb 0.100000 user 0.100000 sys 0.000000 (best of 98)
! wall 0.133518 comb 0.140000 user 0.130000 sys 0.010000 (max of 98)
! wall 0.102381 comb 0.102449 user 0.098265 sys 0.004184 (avg of 98)
! wall 0.101891 comb 0.090000 user 0.090000 sys 0.000000 (median of 98)
Mean of the medians: 0.100587
On the present changeset:
! wall 0.091344 comb 0.090000 user 0.090000 sys 0.000000 (best of 100)
! wall 0.122728 comb 0.120000 user 0.110000 sys 0.010000 (max of 100)
! wall 0.093268 comb 0.093300 user 0.089300 sys 0.004000 (avg of 100)
! wall 0.092567 comb 0.100000 user 0.090000 sys 0.010000 (median of 100)
! wall 0.093294 comb 0.080000 user 0.080000 sys 0.000000 (best of 100)
! wall 0.144887 comb 0.150000 user 0.140000 sys 0.010000 (max of 100)
! wall 0.097708 comb 0.097700 user 0.093400 sys 0.004300 (avg of 100)
! wall 0.094980 comb 0.100000 user 0.090000 sys 0.010000 (median of 100)
! wall 0.091262 comb 0.090000 user 0.080000 sys 0.010000 (best of 100)
! wall 0.123772 comb 0.130000 user 0.120000 sys 0.010000 (max of 100)
! wall 0.093188 comb 0.093200 user 0.089300 sys 0.003900 (avg of 100)
! wall 0.092364 comb 0.100000 user 0.090000 sys 0.010000 (median of 100)
Mean of the medians is 0.0933
Differential Revision: https://phab.mercurial-scm.org/D5358
author | Georges Racinet <gracinet@anybox.fr> |
---|---|
date | Thu, 29 Nov 2018 09:13:13 +0000 |
parents | ef0baff11aea |
children | 84d61fdcefa5 |
line wrap: on
line source
# lfs - hash-preserving large file support using Git-LFS protocol # # Copyright 2017 Facebook, Inc. # # This software may be used and distributed according to the terms of the # GNU General Public License version 2 or any later version. """lfs - large file support (EXPERIMENTAL) This extension allows large files to be tracked outside of the normal repository storage and stored on a centralized server, similar to the ``largefiles`` extension. The ``git-lfs`` protocol is used when communicating with the server, so existing git infrastructure can be harnessed. Even though the files are stored outside of the repository, they are still integrity checked in the same manner as normal files. The files stored outside of the repository are downloaded on demand, which reduces the time to clone, and possibly the local disk usage. This changes fundamental workflows in a DVCS, so careful thought should be given before deploying it. :hg:`convert` can be used to convert LFS repositories to normal repositories that no longer require this extension, and do so without changing the commit hashes. This allows the extension to be disabled if the centralized workflow becomes burdensome. However, the pre and post convert clones will not be able to communicate with each other unless the extension is enabled on both. To start a new repository, or to add LFS files to an existing one, just create an ``.hglfs`` file as described below in the root directory of the repository. Typically, this file should be put under version control, so that the settings will propagate to other repositories with push and pull. During any commit, Mercurial will consult this file to determine if an added or modified file should be stored externally. The type of storage depends on the characteristics of the file at each commit. A file that is near a size threshold may switch back and forth between LFS and normal storage, as needed. Alternately, both normal repositories and largefile controlled repositories can be converted to LFS by using :hg:`convert` and the ``lfs.track`` config option described below. The ``.hglfs`` file should then be created and added, to control subsequent LFS selection. The hashes are also unchanged in this case. The LFS and non-LFS repositories can be distinguished because the LFS repository will abort any command if this extension is disabled. Committed LFS files are held locally, until the repository is pushed. Prior to pushing the normal repository data, the LFS files that are tracked by the outgoing commits are automatically uploaded to the configured central server. No LFS files are transferred on :hg:`pull` or :hg:`clone`. Instead, the files are downloaded on demand as they need to be read, if a cached copy cannot be found locally. Both committing and downloading an LFS file will link the file to a usercache, to speed up future access. See the `usercache` config setting described below. .hglfs:: The extension reads its configuration from a versioned ``.hglfs`` configuration file found in the root of the working directory. The ``.hglfs`` file uses the same syntax as all other Mercurial configuration files. It uses a single section, ``[track]``. The ``[track]`` section specifies which files are stored as LFS (or not). Each line is keyed by a file pattern, with a predicate value. The first file pattern match is used, so put more specific patterns first. The available predicates are ``all()``, ``none()``, and ``size()``. See "hg help filesets.size" for the latter. Example versioned ``.hglfs`` file:: [track] # No Makefile or python file, anywhere, will be LFS **Makefile = none() **.py = none() **.zip = all() **.exe = size(">1MB") # Catchall for everything not matched above ** = size(">10MB") Configs:: [lfs] # Remote endpoint. Multiple protocols are supported: # - http(s)://user:pass@example.com/path # git-lfs endpoint # - file:///tmp/path # local filesystem, usually for testing # if unset, lfs will assume the remote repository also handles blob storage # for http(s) URLs. Otherwise, lfs will prompt to set this when it must # use this value. # (default: unset) url = https://example.com/repo.git/info/lfs # Which files to track in LFS. Path tests are "**.extname" for file # extensions, and "path:under/some/directory" for path prefix. Both # are relative to the repository root. # File size can be tested with the "size()" fileset, and tests can be # joined with fileset operators. (See "hg help filesets.operators".) # # Some examples: # - all() # everything # - none() # nothing # - size(">20MB") # larger than 20MB # - !**.txt # anything not a *.txt file # - **.zip | **.tar.gz | **.7z # some types of compressed files # - path:bin # files under "bin" in the project root # - (**.php & size(">2MB")) | (**.js & size(">5MB")) | **.tar.gz # | (path:bin & !path:/bin/README) | size(">1GB") # (default: none()) # # This is ignored if there is a tracked '.hglfs' file, and this setting # will eventually be deprecated and removed. track = size(">10M") # how many times to retry before giving up on transferring an object retry = 5 # the local directory to store lfs files for sharing across local clones. # If not set, the cache is located in an OS specific cache location. usercache = /path/to/global/cache """ from __future__ import absolute_import import sys from mercurial.i18n import _ from mercurial import ( bundle2, changegroup, cmdutil, config, context, error, exchange, extensions, filelog, filesetlang, localrepo, minifileset, node, pycompat, registrar, repository, revlog, scmutil, templateutil, upgrade, util, vfs as vfsmod, wireprotoserver, wireprotov1server, ) from . import ( blobstore, wireprotolfsserver, wrapper, ) # Note for extension authors: ONLY specify testedwith = 'ships-with-hg-core' for # extensions which SHIP WITH MERCURIAL. Non-mainline extensions should # be specifying the version(s) of Mercurial they are tested with, or # leave the attribute unspecified. testedwith = 'ships-with-hg-core' configtable = {} configitem = registrar.configitem(configtable) configitem('experimental', 'lfs.serve', default=True, ) configitem('experimental', 'lfs.user-agent', default=None, ) configitem('experimental', 'lfs.disableusercache', default=False, ) configitem('experimental', 'lfs.worker-enable', default=False, ) configitem('lfs', 'url', default=None, ) configitem('lfs', 'usercache', default=None, ) # Deprecated configitem('lfs', 'threshold', default=None, ) configitem('lfs', 'track', default='none()', ) configitem('lfs', 'retry', default=5, ) cmdtable = {} command = registrar.command(cmdtable) templatekeyword = registrar.templatekeyword() filesetpredicate = registrar.filesetpredicate() lfsprocessor = ( wrapper.readfromstore, wrapper.writetostore, wrapper.bypasscheckhash, ) def featuresetup(ui, supported): # don't die on seeing a repo with the lfs requirement supported |= {'lfs'} def uisetup(ui): localrepo.featuresetupfuncs.add(featuresetup) def reposetup(ui, repo): # Nothing to do with a remote repo if not repo.local(): return repo.svfs.lfslocalblobstore = blobstore.local(repo) repo.svfs.lfsremoteblobstore = blobstore.remote(repo) class lfsrepo(repo.__class__): @localrepo.unfilteredmethod def commitctx(self, ctx, error=False): repo.svfs.options['lfstrack'] = _trackedmatcher(self) return super(lfsrepo, self).commitctx(ctx, error) repo.__class__ = lfsrepo if 'lfs' not in repo.requirements: def checkrequireslfs(ui, repo, **kwargs): if 'lfs' in repo.requirements: return 0 last = kwargs.get(r'node_last') _bin = node.bin if last: s = repo.set('%n:%n', _bin(kwargs[r'node']), _bin(last)) else: s = repo.set('%n', _bin(kwargs[r'node'])) match = repo.narrowmatch() for ctx in s: # TODO: is there a way to just walk the files in the commit? if any(ctx[f].islfs() for f in ctx.files() if f in ctx and match(f)): repo.requirements.add('lfs') repo.features.add(repository.REPO_FEATURE_LFS) repo._writerequirements() repo.prepushoutgoinghooks.add('lfs', wrapper.prepush) break ui.setconfig('hooks', 'commit.lfs', checkrequireslfs, 'lfs') ui.setconfig('hooks', 'pretxnchangegroup.lfs', checkrequireslfs, 'lfs') else: repo.prepushoutgoinghooks.add('lfs', wrapper.prepush) def _trackedmatcher(repo): """Return a function (path, size) -> bool indicating whether or not to track a given file with lfs.""" if not repo.wvfs.exists('.hglfs'): # No '.hglfs' in wdir. Fallback to config for now. trackspec = repo.ui.config('lfs', 'track') # deprecated config: lfs.threshold threshold = repo.ui.configbytes('lfs', 'threshold') if threshold: filesetlang.parse(trackspec) # make sure syntax errors are confined trackspec = "(%s) | size('>%d')" % (trackspec, threshold) return minifileset.compile(trackspec) data = repo.wvfs.tryread('.hglfs') if not data: return lambda p, s: False # Parse errors here will abort with a message that points to the .hglfs file # and line number. cfg = config.config() cfg.parse('.hglfs', data) try: rules = [(minifileset.compile(pattern), minifileset.compile(rule)) for pattern, rule in cfg.items('track')] except error.ParseError as e: # The original exception gives no indicator that the error is in the # .hglfs file, so add that. # TODO: See if the line number of the file can be made available. raise error.Abort(_('parse error in .hglfs: %s') % e) def _match(path, size): for pat, rule in rules: if pat(path, size): return rule(path, size) return False return _match def wrapfilelog(filelog): wrapfunction = extensions.wrapfunction wrapfunction(filelog, 'addrevision', wrapper.filelogaddrevision) wrapfunction(filelog, 'renamed', wrapper.filelogrenamed) wrapfunction(filelog, 'size', wrapper.filelogsize) def _resolverevlogstorevfsoptions(orig, ui, requirements, features): opts = orig(ui, requirements, features) for name, module in extensions.extensions(ui): if module is sys.modules[__name__]: if revlog.REVIDX_EXTSTORED in opts[b'flagprocessors']: msg = (_(b"cannot register multiple processors on flag '%#x'.") % revlog.REVIDX_EXTSTORED) raise error.Abort(msg) opts[b'flagprocessors'][revlog.REVIDX_EXTSTORED] = lfsprocessor break return opts def extsetup(ui): wrapfilelog(filelog.filelog) wrapfunction = extensions.wrapfunction wrapfunction(localrepo, 'makefilestorage', wrapper.localrepomakefilestorage) wrapfunction(localrepo, 'resolverevlogstorevfsoptions', _resolverevlogstorevfsoptions) wrapfunction(cmdutil, '_updatecatformatter', wrapper._updatecatformatter) wrapfunction(scmutil, 'wrapconvertsink', wrapper.convertsink) wrapfunction(upgrade, '_finishdatamigration', wrapper.upgradefinishdatamigration) wrapfunction(upgrade, 'preservedrequirements', wrapper.upgraderequirements) wrapfunction(upgrade, 'supporteddestrequirements', wrapper.upgraderequirements) wrapfunction(changegroup, 'allsupportedversions', wrapper.allsupportedversions) wrapfunction(exchange, 'push', wrapper.push) wrapfunction(wireprotov1server, '_capabilities', wrapper._capabilities) wrapfunction(wireprotoserver, 'handlewsgirequest', wireprotolfsserver.handlewsgirequest) wrapfunction(context.basefilectx, 'cmp', wrapper.filectxcmp) wrapfunction(context.basefilectx, 'isbinary', wrapper.filectxisbinary) context.basefilectx.islfs = wrapper.filectxislfs scmutil.fileprefetchhooks.add('lfs', wrapper._prefetchfiles) # Make bundle choose changegroup3 instead of changegroup2. This affects # "hg bundle" command. Note: it does not cover all bundle formats like # "packed1". Using "packed1" with lfs will likely cause trouble. exchange._bundlespeccontentopts["v2"]["cg.version"] = "03" # bundlerepo uses "vfsmod.readonlyvfs(othervfs)", we need to make sure lfs # options and blob stores are passed from othervfs to the new readonlyvfs. wrapfunction(vfsmod.readonlyvfs, '__init__', wrapper.vfsinit) # when writing a bundle via "hg bundle" command, upload related LFS blobs wrapfunction(bundle2, 'writenewbundle', wrapper.writenewbundle) @filesetpredicate('lfs()') def lfsfileset(mctx, x): """File that uses LFS storage.""" # i18n: "lfs" is a keyword filesetlang.getargs(x, 0, 0, _("lfs takes no arguments")) ctx = mctx.ctx def lfsfilep(f): return wrapper.pointerfromctx(ctx, f, removed=True) is not None return mctx.predicate(lfsfilep, predrepr='<lfs>') @templatekeyword('lfs_files', requires={'ctx'}) def lfsfiles(context, mapping): """List of strings. All files modified, added, or removed by this changeset.""" ctx = context.resource(mapping, 'ctx') pointers = wrapper.pointersfromctx(ctx, removed=True) # {path: pointer} files = sorted(pointers.keys()) def pointer(v): # In the file spec, version is first and the other keys are sorted. sortkeyfunc = lambda x: (x[0] != 'version', x) items = sorted(pointers[v].iteritems(), key=sortkeyfunc) return util.sortdict(items) makemap = lambda v: { 'file': v, 'lfsoid': pointers[v].oid() if pointers[v] else None, 'lfspointer': templateutil.hybriddict(pointer(v)), } # TODO: make the separator ', '? f = templateutil._showcompatlist(context, mapping, 'lfs_file', files) return templateutil.hybrid(f, files, makemap, pycompat.identity) @command('debuglfsupload', [('r', 'rev', [], _('upload large files introduced by REV'))]) def debuglfsupload(ui, repo, **opts): """upload lfs blobs added by the working copy parent or given revisions""" revs = opts.get(r'rev', []) pointers = wrapper.extractpointers(repo, scmutil.revrange(repo, revs)) wrapper.uploadblobs(repo, pointers)