Mercurial > hg
view hgext/infinitepush/__init__.py @ 40083:48a0ce67d67a
status: intersect matcher with narrow matcher instead of filtering afterwards
I seem to have done a very naive move of the code from the narrow
extension into core in e411774a2e0f (narrow: move status-filtering to
core and to ctx, 2018-08-02). It seems obvious that a better way is to
intersect the matchers.
Note that this means that when requesting status for the working
directory in a narrow repo, we now pass the narrow matcher (possibly
intersected with a user-provided matcher) into _buildstatus() and then
into dirstate.status() and dirstate.walk(), which will the intersect
it again with the narrow matcher. That's functionally fine, but
wasteful. I hope to later remove the dirstate wrapping that adds the
second layer of matcher intersection.
Differential Revision: https://phab.mercurial-scm.org/D4897
author | Martin von Zweigbergk <martinvonz@google.com> |
---|---|
date | Thu, 27 Sep 2018 23:01:26 -0700 |
parents | c31ce080eb75 |
children | 090e5f3900b7 |
line wrap: on
line source
# Infinite push # # Copyright 2016 Facebook, Inc. # # This software may be used and distributed according to the terms of the # GNU General Public License version 2 or any later version. """ store some pushes in a remote blob store on the server (EXPERIMENTAL) [infinitepush] # Server-side and client-side option. Pattern of the infinitepush bookmark branchpattern = PATTERN # Server or client server = False # Server-side option. Possible values: 'disk' or 'sql'. Fails if not set indextype = disk # Server-side option. Used only if indextype=sql. # Format: 'IP:PORT:DB_NAME:USER:PASSWORD' sqlhost = IP:PORT:DB_NAME:USER:PASSWORD # Server-side option. Used only if indextype=disk. # Filesystem path to the index store indexpath = PATH # Server-side option. Possible values: 'disk' or 'external' # Fails if not set storetype = disk # Server-side option. # Path to the binary that will save bundle to the bundlestore # Formatted cmd line will be passed to it (see `put_args`) put_binary = put # Serser-side option. Used only if storetype=external. # Format cmd-line string for put binary. Placeholder: {filename} put_args = {filename} # Server-side option. # Path to the binary that get bundle from the bundlestore. # Formatted cmd line will be passed to it (see `get_args`) get_binary = get # Serser-side option. Used only if storetype=external. # Format cmd-line string for get binary. Placeholders: {filename} {handle} get_args = {filename} {handle} # Server-side option logfile = FIlE # Server-side option loglevel = DEBUG # Server-side option. Used only if indextype=sql. # Sets mysql wait_timeout option. waittimeout = 300 # Server-side option. Used only if indextype=sql. # Sets mysql innodb_lock_wait_timeout option. locktimeout = 120 # Server-side option. Used only if indextype=sql. # Name of the repository reponame = '' # Client-side option. Used by --list-remote option. List of remote scratch # patterns to list if no patterns are specified. defaultremotepatterns = ['*'] # Instructs infinitepush to forward all received bundle2 parts to the # bundle for storage. Defaults to False. storeallparts = True # routes each incoming push to the bundlestore. defaults to False pushtobundlestore = True [remotenames] # Client-side option # This option should be set only if remotenames extension is enabled. # Whether remote bookmarks are tracked by remotenames extension. bookmarks = True """ from __future__ import absolute_import import collections import contextlib import errno import functools import logging import os import random import re import socket import subprocess import time from mercurial.node import ( bin, hex, ) from mercurial.i18n import _ from mercurial.utils import ( procutil, stringutil, ) from mercurial import ( bundle2, changegroup, commands, discovery, encoding, error, exchange, extensions, hg, localrepo, phases, pushkey, pycompat, registrar, util, wireprototypes, wireprotov1peer, wireprotov1server, ) from . import ( bundleparts, common, ) # Note for extension authors: ONLY specify testedwith = 'ships-with-hg-core' for # extensions which SHIP WITH MERCURIAL. Non-mainline extensions should # be specifying the version(s) of Mercurial they are tested with, or # leave the attribute unspecified. testedwith = 'ships-with-hg-core' configtable = {} configitem = registrar.configitem(configtable) configitem('infinitepush', 'server', default=False, ) configitem('infinitepush', 'storetype', default='', ) configitem('infinitepush', 'indextype', default='', ) configitem('infinitepush', 'indexpath', default='', ) configitem('infinitepush', 'storeallparts', default=False, ) configitem('infinitepush', 'reponame', default='', ) configitem('scratchbranch', 'storepath', default='', ) configitem('infinitepush', 'branchpattern', default='', ) configitem('infinitepush', 'pushtobundlestore', default=False, ) configitem('experimental', 'server-bundlestore-bookmark', default='', ) configitem('experimental', 'infinitepush-scratchpush', default=False, ) experimental = 'experimental' configbookmark = 'server-bundlestore-bookmark' configscratchpush = 'infinitepush-scratchpush' scratchbranchparttype = bundleparts.scratchbranchparttype revsetpredicate = registrar.revsetpredicate() templatekeyword = registrar.templatekeyword() _scratchbranchmatcher = lambda x: False _maybehash = re.compile(r'^[a-f0-9]+$').search def _buildexternalbundlestore(ui): put_args = ui.configlist('infinitepush', 'put_args', []) put_binary = ui.config('infinitepush', 'put_binary') if not put_binary: raise error.Abort('put binary is not specified') get_args = ui.configlist('infinitepush', 'get_args', []) get_binary = ui.config('infinitepush', 'get_binary') if not get_binary: raise error.Abort('get binary is not specified') from . import store return store.externalbundlestore(put_binary, put_args, get_binary, get_args) def _buildsqlindex(ui): sqlhost = ui.config('infinitepush', 'sqlhost') if not sqlhost: raise error.Abort(_('please set infinitepush.sqlhost')) host, port, db, user, password = sqlhost.split(':') reponame = ui.config('infinitepush', 'reponame') if not reponame: raise error.Abort(_('please set infinitepush.reponame')) logfile = ui.config('infinitepush', 'logfile', '') waittimeout = ui.configint('infinitepush', 'waittimeout', 300) locktimeout = ui.configint('infinitepush', 'locktimeout', 120) from . import sqlindexapi return sqlindexapi.sqlindexapi( reponame, host, port, db, user, password, logfile, _getloglevel(ui), waittimeout=waittimeout, locktimeout=locktimeout) def _getloglevel(ui): loglevel = ui.config('infinitepush', 'loglevel', 'DEBUG') numeric_loglevel = getattr(logging, loglevel.upper(), None) if not isinstance(numeric_loglevel, int): raise error.Abort(_('invalid log level %s') % loglevel) return numeric_loglevel def _tryhoist(ui, remotebookmark): '''returns a bookmarks with hoisted part removed Remotenames extension has a 'hoist' config that allows to use remote bookmarks without specifying remote path. For example, 'hg update master' works as well as 'hg update remote/master'. We want to allow the same in infinitepush. ''' if common.isremotebooksenabled(ui): hoist = ui.config('remotenames', 'hoistedpeer') + '/' if remotebookmark.startswith(hoist): return remotebookmark[len(hoist):] return remotebookmark class bundlestore(object): def __init__(self, repo): self._repo = repo storetype = self._repo.ui.config('infinitepush', 'storetype') if storetype == 'disk': from . import store self.store = store.filebundlestore(self._repo.ui, self._repo) elif storetype == 'external': self.store = _buildexternalbundlestore(self._repo.ui) else: raise error.Abort( _('unknown infinitepush store type specified %s') % storetype) indextype = self._repo.ui.config('infinitepush', 'indextype') if indextype == 'disk': from . import fileindexapi self.index = fileindexapi.fileindexapi(self._repo) elif indextype == 'sql': self.index = _buildsqlindex(self._repo.ui) else: raise error.Abort( _('unknown infinitepush index type specified %s') % indextype) def _isserver(ui): return ui.configbool('infinitepush', 'server') def reposetup(ui, repo): if _isserver(ui) and repo.local(): repo.bundlestore = bundlestore(repo) def extsetup(ui): commonsetup(ui) if _isserver(ui): serverextsetup(ui) else: clientextsetup(ui) def commonsetup(ui): wireprotov1server.commands['listkeyspatterns'] = ( wireprotolistkeyspatterns, 'namespace patterns') scratchbranchpat = ui.config('infinitepush', 'branchpattern') if scratchbranchpat: global _scratchbranchmatcher kind, pat, _scratchbranchmatcher = \ stringutil.stringmatcher(scratchbranchpat) def serverextsetup(ui): origpushkeyhandler = bundle2.parthandlermapping['pushkey'] def newpushkeyhandler(*args, **kwargs): bundle2pushkey(origpushkeyhandler, *args, **kwargs) newpushkeyhandler.params = origpushkeyhandler.params bundle2.parthandlermapping['pushkey'] = newpushkeyhandler orighandlephasehandler = bundle2.parthandlermapping['phase-heads'] newphaseheadshandler = lambda *args, **kwargs: \ bundle2handlephases(orighandlephasehandler, *args, **kwargs) newphaseheadshandler.params = orighandlephasehandler.params bundle2.parthandlermapping['phase-heads'] = newphaseheadshandler extensions.wrapfunction(localrepo.localrepository, 'listkeys', localrepolistkeys) wireprotov1server.commands['lookup'] = ( _lookupwrap(wireprotov1server.commands['lookup'][0]), 'key') extensions.wrapfunction(exchange, 'getbundlechunks', getbundlechunks) extensions.wrapfunction(bundle2, 'processparts', processparts) def clientextsetup(ui): entry = extensions.wrapcommand(commands.table, 'push', _push) entry[1].append( ('', 'bundle-store', None, _('force push to go to bundle store (EXPERIMENTAL)'))) extensions.wrapcommand(commands.table, 'pull', _pull) extensions.wrapfunction(discovery, 'checkheads', _checkheads) wireprotov1peer.wirepeer.listkeyspatterns = listkeyspatterns partorder = exchange.b2partsgenorder index = partorder.index('changeset') partorder.insert( index, partorder.pop(partorder.index(scratchbranchparttype))) def _checkheads(orig, pushop): if pushop.ui.configbool(experimental, configscratchpush, False): return return orig(pushop) def wireprotolistkeyspatterns(repo, proto, namespace, patterns): patterns = wireprototypes.decodelist(patterns) d = repo.listkeys(encoding.tolocal(namespace), patterns).iteritems() return pushkey.encodekeys(d) def localrepolistkeys(orig, self, namespace, patterns=None): if namespace == 'bookmarks' and patterns: index = self.bundlestore.index results = {} bookmarks = orig(self, namespace) for pattern in patterns: results.update(index.getbookmarks(pattern)) if pattern.endswith('*'): pattern = 're:^' + pattern[:-1] + '.*' kind, pat, matcher = stringutil.stringmatcher(pattern) for bookmark, node in bookmarks.iteritems(): if matcher(bookmark): results[bookmark] = node return results else: return orig(self, namespace) @wireprotov1peer.batchable def listkeyspatterns(self, namespace, patterns): if not self.capable('pushkey'): yield {}, None f = wireprotov1peer.future() self.ui.debug('preparing listkeys for "%s" with pattern "%s"\n' % (namespace, patterns)) yield { 'namespace': encoding.fromlocal(namespace), 'patterns': wireprototypes.encodelist(patterns) }, f d = f.value self.ui.debug('received listkey for "%s": %i bytes\n' % (namespace, len(d))) yield pushkey.decodekeys(d) def _readbundlerevs(bundlerepo): return list(bundlerepo.revs('bundle()')) def _includefilelogstobundle(bundlecaps, bundlerepo, bundlerevs, ui): '''Tells remotefilelog to include all changed files to the changegroup By default remotefilelog doesn't include file content to the changegroup. But we need to include it if we are fetching from bundlestore. ''' changedfiles = set() cl = bundlerepo.changelog for r in bundlerevs: # [3] means changed files changedfiles.update(cl.read(r)[3]) if not changedfiles: return bundlecaps changedfiles = '\0'.join(changedfiles) newcaps = [] appended = False for cap in (bundlecaps or []): if cap.startswith('excludepattern='): newcaps.append('\0'.join((cap, changedfiles))) appended = True else: newcaps.append(cap) if not appended: # Not found excludepattern cap. Just append it newcaps.append('excludepattern=' + changedfiles) return newcaps def _rebundle(bundlerepo, bundleroots, unknownhead): ''' Bundle may include more revision then user requested. For example, if user asks for revision but bundle also consists its descendants. This function will filter out all revision that user is not requested. ''' parts = [] version = '02' outgoing = discovery.outgoing(bundlerepo, commonheads=bundleroots, missingheads=[unknownhead]) cgstream = changegroup.makestream(bundlerepo, outgoing, version, 'pull') cgstream = util.chunkbuffer(cgstream).read() cgpart = bundle2.bundlepart('changegroup', data=cgstream) cgpart.addparam('version', version) parts.append(cgpart) return parts def _getbundleroots(oldrepo, bundlerepo, bundlerevs): cl = bundlerepo.changelog bundleroots = [] for rev in bundlerevs: node = cl.node(rev) parents = cl.parents(node) for parent in parents: # include all revs that exist in the main repo # to make sure that bundle may apply client-side if parent in oldrepo: bundleroots.append(parent) return bundleroots def _needsrebundling(head, bundlerepo): bundleheads = list(bundlerepo.revs('heads(bundle())')) return not (len(bundleheads) == 1 and bundlerepo[bundleheads[0]].node() == head) def _generateoutputparts(head, bundlerepo, bundleroots, bundlefile): '''generates bundle that will be send to the user returns tuple with raw bundle string and bundle type ''' parts = [] if not _needsrebundling(head, bundlerepo): with util.posixfile(bundlefile, "rb") as f: unbundler = exchange.readbundle(bundlerepo.ui, f, bundlefile) if isinstance(unbundler, changegroup.cg1unpacker): part = bundle2.bundlepart('changegroup', data=unbundler._stream.read()) part.addparam('version', '01') parts.append(part) elif isinstance(unbundler, bundle2.unbundle20): haschangegroup = False for part in unbundler.iterparts(): if part.type == 'changegroup': haschangegroup = True newpart = bundle2.bundlepart(part.type, data=part.read()) for key, value in part.params.iteritems(): newpart.addparam(key, value) parts.append(newpart) if not haschangegroup: raise error.Abort( 'unexpected bundle without changegroup part, ' + 'head: %s' % hex(head), hint='report to administrator') else: raise error.Abort('unknown bundle type') else: parts = _rebundle(bundlerepo, bundleroots, head) return parts def getbundlechunks(orig, repo, source, heads=None, bundlecaps=None, **kwargs): heads = heads or [] # newheads are parents of roots of scratch bundles that were requested newphases = {} scratchbundles = [] newheads = [] scratchheads = [] nodestobundle = {} allbundlestocleanup = [] try: for head in heads: if head not in repo.changelog.nodemap: if head not in nodestobundle: newbundlefile = common.downloadbundle(repo, head) bundlepath = "bundle:%s+%s" % (repo.root, newbundlefile) bundlerepo = hg.repository(repo.ui, bundlepath) allbundlestocleanup.append((bundlerepo, newbundlefile)) bundlerevs = set(_readbundlerevs(bundlerepo)) bundlecaps = _includefilelogstobundle( bundlecaps, bundlerepo, bundlerevs, repo.ui) cl = bundlerepo.changelog bundleroots = _getbundleroots(repo, bundlerepo, bundlerevs) for rev in bundlerevs: node = cl.node(rev) newphases[hex(node)] = str(phases.draft) nodestobundle[node] = (bundlerepo, bundleroots, newbundlefile) scratchbundles.append( _generateoutputparts(head, *nodestobundle[head])) newheads.extend(bundleroots) scratchheads.append(head) finally: for bundlerepo, bundlefile in allbundlestocleanup: bundlerepo.close() try: os.unlink(bundlefile) except (IOError, OSError): # if we can't cleanup the file then just ignore the error, # no need to fail pass pullfrombundlestore = bool(scratchbundles) wrappedchangegrouppart = False wrappedlistkeys = False oldchangegrouppart = exchange.getbundle2partsmapping['changegroup'] try: def _changegrouppart(bundler, *args, **kwargs): # Order is important here. First add non-scratch part # and only then add parts with scratch bundles because # non-scratch part contains parents of roots of scratch bundles. result = oldchangegrouppart(bundler, *args, **kwargs) for bundle in scratchbundles: for part in bundle: bundler.addpart(part) return result exchange.getbundle2partsmapping['changegroup'] = _changegrouppart wrappedchangegrouppart = True def _listkeys(orig, self, namespace): origvalues = orig(self, namespace) if namespace == 'phases' and pullfrombundlestore: if origvalues.get('publishing') == 'True': # Make repo non-publishing to preserve draft phase del origvalues['publishing'] origvalues.update(newphases) return origvalues extensions.wrapfunction(localrepo.localrepository, 'listkeys', _listkeys) wrappedlistkeys = True heads = list((set(newheads) | set(heads)) - set(scratchheads)) result = orig(repo, source, heads=heads, bundlecaps=bundlecaps, **kwargs) finally: if wrappedchangegrouppart: exchange.getbundle2partsmapping['changegroup'] = oldchangegrouppart if wrappedlistkeys: extensions.unwrapfunction(localrepo.localrepository, 'listkeys', _listkeys) return result def _lookupwrap(orig): def _lookup(repo, proto, key): localkey = encoding.tolocal(key) if isinstance(localkey, str) and _scratchbranchmatcher(localkey): scratchnode = repo.bundlestore.index.getnode(localkey) if scratchnode: return "%d %s\n" % (1, scratchnode) else: return "%d %s\n" % (0, 'scratch branch %s not found' % localkey) else: try: r = hex(repo.lookup(localkey)) return "%d %s\n" % (1, r) except Exception as inst: if repo.bundlestore.index.getbundle(localkey): return "%d %s\n" % (1, localkey) else: r = stringutil.forcebytestr(inst) return "%d %s\n" % (0, r) return _lookup def _pull(orig, ui, repo, source="default", **opts): opts = pycompat.byteskwargs(opts) # Copy paste from `pull` command source, branches = hg.parseurl(ui.expandpath(source), opts.get('branch')) scratchbookmarks = {} unfi = repo.unfiltered() unknownnodes = [] for rev in opts.get('rev', []): if rev not in unfi: unknownnodes.append(rev) if opts.get('bookmark'): bookmarks = [] revs = opts.get('rev') or [] for bookmark in opts.get('bookmark'): if _scratchbranchmatcher(bookmark): # rev is not known yet # it will be fetched with listkeyspatterns next scratchbookmarks[bookmark] = 'REVTOFETCH' else: bookmarks.append(bookmark) if scratchbookmarks: other = hg.peer(repo, opts, source) fetchedbookmarks = other.listkeyspatterns( 'bookmarks', patterns=scratchbookmarks) for bookmark in scratchbookmarks: if bookmark not in fetchedbookmarks: raise error.Abort('remote bookmark %s not found!' % bookmark) scratchbookmarks[bookmark] = fetchedbookmarks[bookmark] revs.append(fetchedbookmarks[bookmark]) opts['bookmark'] = bookmarks opts['rev'] = revs if scratchbookmarks or unknownnodes: # Set anyincoming to True extensions.wrapfunction(discovery, 'findcommonincoming', _findcommonincoming) try: # Remote scratch bookmarks will be deleted because remotenames doesn't # know about them. Let's save it before pull and restore after remotescratchbookmarks = _readscratchremotebookmarks(ui, repo, source) result = orig(ui, repo, source, **pycompat.strkwargs(opts)) # TODO(stash): race condition is possible # if scratch bookmarks was updated right after orig. # But that's unlikely and shouldn't be harmful. if common.isremotebooksenabled(ui): remotescratchbookmarks.update(scratchbookmarks) _saveremotebookmarks(repo, remotescratchbookmarks, source) else: _savelocalbookmarks(repo, scratchbookmarks) return result finally: if scratchbookmarks: extensions.unwrapfunction(discovery, 'findcommonincoming') def _readscratchremotebookmarks(ui, repo, other): if common.isremotebooksenabled(ui): remotenamesext = extensions.find('remotenames') remotepath = remotenamesext.activepath(repo.ui, other) result = {} # Let's refresh remotenames to make sure we have it up to date # Seems that `repo.names['remotebookmarks']` may return stale bookmarks # and it results in deleting scratch bookmarks. Our best guess how to # fix it is to use `clearnames()` repo._remotenames.clearnames() for remotebookmark in repo.names['remotebookmarks'].listnames(repo): path, bookname = remotenamesext.splitremotename(remotebookmark) if path == remotepath and _scratchbranchmatcher(bookname): nodes = repo.names['remotebookmarks'].nodes(repo, remotebookmark) if nodes: result[bookname] = hex(nodes[0]) return result else: return {} def _saveremotebookmarks(repo, newbookmarks, remote): remotenamesext = extensions.find('remotenames') remotepath = remotenamesext.activepath(repo.ui, remote) branches = collections.defaultdict(list) bookmarks = {} remotenames = remotenamesext.readremotenames(repo) for hexnode, nametype, remote, rname in remotenames: if remote != remotepath: continue if nametype == 'bookmarks': if rname in newbookmarks: # It's possible if we have a normal bookmark that matches # scratch branch pattern. In this case just use the current # bookmark node del newbookmarks[rname] bookmarks[rname] = hexnode elif nametype == 'branches': # saveremotenames expects 20 byte binary nodes for branches branches[rname].append(bin(hexnode)) for bookmark, hexnode in newbookmarks.iteritems(): bookmarks[bookmark] = hexnode remotenamesext.saveremotenames(repo, remotepath, branches, bookmarks) def _savelocalbookmarks(repo, bookmarks): if not bookmarks: return with repo.wlock(), repo.lock(), repo.transaction('bookmark') as tr: changes = [] for scratchbook, node in bookmarks.iteritems(): changectx = repo[node] changes.append((scratchbook, changectx.node())) repo._bookmarks.applychanges(repo, tr, changes) def _findcommonincoming(orig, *args, **kwargs): common, inc, remoteheads = orig(*args, **kwargs) return common, True, remoteheads def _push(orig, ui, repo, dest=None, *args, **opts): bookmark = opts.get(r'bookmark') # we only support pushing one infinitepush bookmark at once if len(bookmark) == 1: bookmark = bookmark[0] else: bookmark = '' oldphasemove = None overrides = {(experimental, configbookmark): bookmark} with ui.configoverride(overrides, 'infinitepush'): scratchpush = opts.get('bundle_store') if _scratchbranchmatcher(bookmark): scratchpush = True # bundle2 can be sent back after push (for example, bundle2 # containing `pushkey` part to update bookmarks) ui.setconfig(experimental, 'bundle2.pushback', True) if scratchpush: # this is an infinitepush, we don't want the bookmark to be applied # rather that should be stored in the bundlestore opts[r'bookmark'] = [] ui.setconfig(experimental, configscratchpush, True) oldphasemove = extensions.wrapfunction(exchange, '_localphasemove', _phasemove) # Copy-paste from `push` command path = ui.paths.getpath(dest, default=('default-push', 'default')) if not path: raise error.Abort(_('default repository not configured!'), hint=_("see 'hg help config.paths'")) destpath = path.pushloc or path.loc # Remote scratch bookmarks will be deleted because remotenames doesn't # know about them. Let's save it before push and restore after remotescratchbookmarks = _readscratchremotebookmarks(ui, repo, destpath) result = orig(ui, repo, dest, *args, **opts) if common.isremotebooksenabled(ui): if bookmark and scratchpush: other = hg.peer(repo, opts, destpath) fetchedbookmarks = other.listkeyspatterns('bookmarks', patterns=[bookmark]) remotescratchbookmarks.update(fetchedbookmarks) _saveremotebookmarks(repo, remotescratchbookmarks, destpath) if oldphasemove: exchange._localphasemove = oldphasemove return result def _deleteinfinitepushbookmarks(ui, repo, path, names): """Prune remote names by removing the bookmarks we don't want anymore, then writing the result back to disk """ remotenamesext = extensions.find('remotenames') # remotename format is: # (node, nametype ("branches" or "bookmarks"), remote, name) nametype_idx = 1 remote_idx = 2 name_idx = 3 remotenames = [remotename for remotename in \ remotenamesext.readremotenames(repo) \ if remotename[remote_idx] == path] remote_bm_names = [remotename[name_idx] for remotename in \ remotenames if remotename[nametype_idx] == "bookmarks"] for name in names: if name not in remote_bm_names: raise error.Abort(_("infinitepush bookmark '{}' does not exist " "in path '{}'").format(name, path)) bookmarks = {} branches = collections.defaultdict(list) for node, nametype, remote, name in remotenames: if nametype == "bookmarks" and name not in names: bookmarks[name] = node elif nametype == "branches": # saveremotenames wants binary nodes for branches branches[name].append(bin(node)) remotenamesext.saveremotenames(repo, path, branches, bookmarks) def _phasemove(orig, pushop, nodes, phase=phases.public): """prevent commits from being marked public Since these are going to a scratch branch, they aren't really being published.""" if phase != phases.public: orig(pushop, nodes, phase) @exchange.b2partsgenerator(scratchbranchparttype) def partgen(pushop, bundler): bookmark = pushop.ui.config(experimental, configbookmark) scratchpush = pushop.ui.configbool(experimental, configscratchpush) if 'changesets' in pushop.stepsdone or not scratchpush: return if scratchbranchparttype not in bundle2.bundle2caps(pushop.remote): return pushop.stepsdone.add('changesets') if not pushop.outgoing.missing: pushop.ui.status(_('no changes found\n')) pushop.cgresult = 0 return # This parameter tells the server that the following bundle is an # infinitepush. This let's it switch the part processing to our infinitepush # code path. bundler.addparam("infinitepush", "True") scratchparts = bundleparts.getscratchbranchparts(pushop.repo, pushop.remote, pushop.outgoing, pushop.ui, bookmark) for scratchpart in scratchparts: bundler.addpart(scratchpart) def handlereply(op): # server either succeeds or aborts; no code to read pushop.cgresult = 1 return handlereply bundle2.capabilities[bundleparts.scratchbranchparttype] = () def _getrevs(bundle, oldnode, force, bookmark): 'extracts and validates the revs to be imported' revs = [bundle[r] for r in bundle.revs('sort(bundle())')] # new bookmark if oldnode is None: return revs # Fast forward update if oldnode in bundle and list(bundle.set('bundle() & %s::', oldnode)): return revs return revs @contextlib.contextmanager def logservicecall(logger, service, **kwargs): start = time.time() logger(service, eventtype='start', **kwargs) try: yield logger(service, eventtype='success', elapsedms=(time.time() - start) * 1000, **kwargs) except Exception as e: logger(service, eventtype='failure', elapsedms=(time.time() - start) * 1000, errormsg=str(e), **kwargs) raise def _getorcreateinfinitepushlogger(op): logger = op.records['infinitepushlogger'] if not logger: ui = op.repo.ui try: username = procutil.getuser() except Exception: username = 'unknown' # Generate random request id to be able to find all logged entries # for the same request. Since requestid is pseudo-generated it may # not be unique, but we assume that (hostname, username, requestid) # is unique. random.seed() requestid = random.randint(0, 2000000000) hostname = socket.gethostname() logger = functools.partial(ui.log, 'infinitepush', user=username, requestid=requestid, hostname=hostname, reponame=ui.config('infinitepush', 'reponame')) op.records.add('infinitepushlogger', logger) else: logger = logger[0] return logger def storetobundlestore(orig, repo, op, unbundler): """stores the incoming bundle coming from push command to the bundlestore instead of applying on the revlogs""" repo.ui.status(_("storing changesets on the bundlestore\n")) bundler = bundle2.bundle20(repo.ui) # processing each part and storing it in bundler with bundle2.partiterator(repo, op, unbundler) as parts: for part in parts: bundlepart = None if part.type == 'replycaps': # This configures the current operation to allow reply parts. bundle2._processpart(op, part) else: bundlepart = bundle2.bundlepart(part.type, data=part.read()) for key, value in part.params.iteritems(): bundlepart.addparam(key, value) # Certain parts require a response if part.type in ('pushkey', 'changegroup'): if op.reply is not None: rpart = op.reply.newpart('reply:%s' % part.type) rpart.addparam('in-reply-to', str(part.id), mandatory=False) rpart.addparam('return', '1', mandatory=False) op.records.add(part.type, { 'return': 1, }) if bundlepart: bundler.addpart(bundlepart) # storing the bundle in the bundlestore buf = util.chunkbuffer(bundler.getchunks()) fd, bundlefile = pycompat.mkstemp() try: try: fp = os.fdopen(fd, r'wb') fp.write(buf.read()) finally: fp.close() storebundle(op, {}, bundlefile) finally: try: os.unlink(bundlefile) except Exception: # we would rather see the original exception pass def processparts(orig, repo, op, unbundler): # make sure we don't wrap processparts in case of `hg unbundle` if op.source == 'unbundle': return orig(repo, op, unbundler) # this server routes each push to bundle store if repo.ui.configbool('infinitepush', 'pushtobundlestore'): return storetobundlestore(orig, repo, op, unbundler) if unbundler.params.get('infinitepush') != 'True': return orig(repo, op, unbundler) handleallparts = repo.ui.configbool('infinitepush', 'storeallparts') bundler = bundle2.bundle20(repo.ui) cgparams = None with bundle2.partiterator(repo, op, unbundler) as parts: for part in parts: bundlepart = None if part.type == 'replycaps': # This configures the current operation to allow reply parts. bundle2._processpart(op, part) elif part.type == bundleparts.scratchbranchparttype: # Scratch branch parts need to be converted to normal # changegroup parts, and the extra parameters stored for later # when we upload to the store. Eventually those parameters will # be put on the actual bundle instead of this part, then we can # send a vanilla changegroup instead of the scratchbranch part. cgversion = part.params.get('cgversion', '01') bundlepart = bundle2.bundlepart('changegroup', data=part.read()) bundlepart.addparam('version', cgversion) cgparams = part.params # If we're not dumping all parts into the new bundle, we need to # alert the future pushkey and phase-heads handler to skip # the part. if not handleallparts: op.records.add(scratchbranchparttype + '_skippushkey', True) op.records.add(scratchbranchparttype + '_skipphaseheads', True) else: if handleallparts: # Ideally we would not process any parts, and instead just # forward them to the bundle for storage, but since this # differs from previous behavior, we need to put it behind a # config flag for incremental rollout. bundlepart = bundle2.bundlepart(part.type, data=part.read()) for key, value in part.params.iteritems(): bundlepart.addparam(key, value) # Certain parts require a response if part.type == 'pushkey': if op.reply is not None: rpart = op.reply.newpart('reply:pushkey') rpart.addparam('in-reply-to', str(part.id), mandatory=False) rpart.addparam('return', '1', mandatory=False) else: bundle2._processpart(op, part) if handleallparts: op.records.add(part.type, { 'return': 1, }) if bundlepart: bundler.addpart(bundlepart) # If commits were sent, store them if cgparams: buf = util.chunkbuffer(bundler.getchunks()) fd, bundlefile = pycompat.mkstemp() try: try: fp = os.fdopen(fd, r'wb') fp.write(buf.read()) finally: fp.close() storebundle(op, cgparams, bundlefile) finally: try: os.unlink(bundlefile) except Exception: # we would rather see the original exception pass def storebundle(op, params, bundlefile): log = _getorcreateinfinitepushlogger(op) parthandlerstart = time.time() log(scratchbranchparttype, eventtype='start') index = op.repo.bundlestore.index store = op.repo.bundlestore.store op.records.add(scratchbranchparttype + '_skippushkey', True) bundle = None try: # guards bundle bundlepath = "bundle:%s+%s" % (op.repo.root, bundlefile) bundle = hg.repository(op.repo.ui, bundlepath) bookmark = params.get('bookmark') bookprevnode = params.get('bookprevnode', '') force = params.get('force') if bookmark: oldnode = index.getnode(bookmark) else: oldnode = None bundleheads = bundle.revs('heads(bundle())') if bookmark and len(bundleheads) > 1: raise error.Abort( _('cannot push more than one head to a scratch branch')) revs = _getrevs(bundle, oldnode, force, bookmark) # Notify the user of what is being pushed plural = 's' if len(revs) > 1 else '' op.repo.ui.warn(_("pushing %d commit%s:\n") % (len(revs), plural)) maxoutput = 10 for i in range(0, min(len(revs), maxoutput)): firstline = bundle[revs[i]].description().split('\n')[0][:50] op.repo.ui.warn((" %s %s\n") % (revs[i], firstline)) if len(revs) > maxoutput + 1: op.repo.ui.warn((" ...\n")) firstline = bundle[revs[-1]].description().split('\n')[0][:50] op.repo.ui.warn((" %s %s\n") % (revs[-1], firstline)) nodesctx = [bundle[rev] for rev in revs] inindex = lambda rev: bool(index.getbundle(bundle[rev].hex())) if bundleheads: newheadscount = sum(not inindex(rev) for rev in bundleheads) else: newheadscount = 0 # If there's a bookmark specified, there should be only one head, # so we choose the last node, which will be that head. # If a bug or malicious client allows there to be a bookmark # with multiple heads, we will place the bookmark on the last head. bookmarknode = nodesctx[-1].hex() if nodesctx else None key = None if newheadscount: with open(bundlefile, 'rb') as f: bundledata = f.read() with logservicecall(log, 'bundlestore', bundlesize=len(bundledata)): bundlesizelimit = 100 * 1024 * 1024 # 100 MB if len(bundledata) > bundlesizelimit: error_msg = ('bundle is too big: %d bytes. ' + 'max allowed size is 100 MB') raise error.Abort(error_msg % (len(bundledata),)) key = store.write(bundledata) with logservicecall(log, 'index', newheadscount=newheadscount), index: if key: index.addbundle(key, nodesctx) if bookmark: index.addbookmark(bookmark, bookmarknode) _maybeaddpushbackpart(op, bookmark, bookmarknode, bookprevnode, params) log(scratchbranchparttype, eventtype='success', elapsedms=(time.time() - parthandlerstart) * 1000) except Exception as e: log(scratchbranchparttype, eventtype='failure', elapsedms=(time.time() - parthandlerstart) * 1000, errormsg=str(e)) raise finally: if bundle: bundle.close() @bundle2.parthandler(scratchbranchparttype, ('bookmark', 'bookprevnode', 'force', 'pushbackbookmarks', 'cgversion')) def bundle2scratchbranch(op, part): '''unbundle a bundle2 part containing a changegroup to store''' bundler = bundle2.bundle20(op.repo.ui) cgversion = part.params.get('cgversion', '01') cgpart = bundle2.bundlepart('changegroup', data=part.read()) cgpart.addparam('version', cgversion) bundler.addpart(cgpart) buf = util.chunkbuffer(bundler.getchunks()) fd, bundlefile = pycompat.mkstemp() try: try: fp = os.fdopen(fd, r'wb') fp.write(buf.read()) finally: fp.close() storebundle(op, part.params, bundlefile) finally: try: os.unlink(bundlefile) except OSError as e: if e.errno != errno.ENOENT: raise return 1 def _maybeaddpushbackpart(op, bookmark, newnode, oldnode, params): if params.get('pushbackbookmarks'): if op.reply and 'pushback' in op.reply.capabilities: params = { 'namespace': 'bookmarks', 'key': bookmark, 'new': newnode, 'old': oldnode, } op.reply.newpart('pushkey', mandatoryparams=params.iteritems()) def bundle2pushkey(orig, op, part): '''Wrapper of bundle2.handlepushkey() The only goal is to skip calling the original function if flag is set. It's set if infinitepush push is happening. ''' if op.records[scratchbranchparttype + '_skippushkey']: if op.reply is not None: rpart = op.reply.newpart('reply:pushkey') rpart.addparam('in-reply-to', str(part.id), mandatory=False) rpart.addparam('return', '1', mandatory=False) return 1 return orig(op, part) def bundle2handlephases(orig, op, part): '''Wrapper of bundle2.handlephases() The only goal is to skip calling the original function if flag is set. It's set if infinitepush push is happening. ''' if op.records[scratchbranchparttype + '_skipphaseheads']: return return orig(op, part) def _asyncsavemetadata(root, nodes): '''starts a separate process that fills metadata for the nodes This function creates a separate process and doesn't wait for it's completion. This was done to avoid slowing down pushes ''' maxnodes = 50 if len(nodes) > maxnodes: return nodesargs = [] for node in nodes: nodesargs.append('--node') nodesargs.append(node) with open(os.devnull, 'w+b') as devnull: cmdline = [util.hgexecutable(), 'debugfillinfinitepushmetadata', '-R', root] + nodesargs # Process will run in background. We don't care about the return code subprocess.Popen(pycompat.rapply(procutil.tonativestr, cmdline), close_fds=True, shell=False, stdin=devnull, stdout=devnull, stderr=devnull)