Mercurial > hg
view mercurial/revset.py @ 51424:3cee8706f53b
phases: directly update the phase sets in advanceboundary
This is similar to what we do in retractboundary. There is no need to invalidate
the cache if we have everything at hand to update it.
author | Pierre-Yves David <pierre-yves.david@octobus.net> |
---|---|
date | Fri, 23 Feb 2024 06:37:25 +0100 |
parents | 18c8c18993f0 |
children | de5bf3fe0233 |
line wrap: on
line source
# revset.py - revision set queries for mercurial # # Copyright 2010 Olivia Mackall <olivia@selenic.com> # # This software may be used and distributed according to the terms of the # GNU General Public License version 2 or any later version. import binascii import functools import random import re from .i18n import _ from .node import ( bin, nullrev, wdirrev, ) from . import ( dagop, destutil, diffutil, encoding, error, grep as grepmod, hbisect, match as matchmod, obsolete as obsmod, obsutil, pathutil, phases, pycompat, registrar, repoview, revsetlang, scmutil, smartset, stack as stackmod, util, ) from .utils import ( dateutil, stringutil, urlutil, ) # helpers for processing parsed tree getsymbol = revsetlang.getsymbol getstring = revsetlang.getstring getinteger = revsetlang.getinteger getboolean = revsetlang.getboolean getlist = revsetlang.getlist getintrange = revsetlang.getintrange getargs = revsetlang.getargs getargsdict = revsetlang.getargsdict baseset = smartset.baseset generatorset = smartset.generatorset spanset = smartset.spanset fullreposet = smartset.fullreposet # revisions not included in all(), but populated if specified _virtualrevs = (nullrev, wdirrev) # Constants for ordering requirement, used in getset(): # # If 'define', any nested functions and operations MAY change the ordering of # the entries in the set (but if changes the ordering, it MUST ALWAYS change # it). If 'follow', any nested functions and operations MUST take the ordering # specified by the first operand to the '&' operator. # # For instance, # # X & (Y | Z) # ^ ^^^^^^^ # | follow # define # # will be evaluated as 'or(y(x()), z(x()))', where 'x()' can change the order # of the entries in the set, but 'y()', 'z()' and 'or()' shouldn't. # # 'any' means the order doesn't matter. For instance, # # (X & !Y) | ancestors(Z) # ^ ^ # any any # # For 'X & !Y', 'X' decides the order and 'Y' is subtracted from 'X', so the # order of 'Y' does not matter. For 'ancestors(Z)', Z's order does not matter # since 'ancestors' does not care about the order of its argument. # # Currently, most revsets do not care about the order, so 'define' is # equivalent to 'follow' for them, and the resulting order is based on the # 'subset' parameter passed down to them: # # m = revset.match(...) # m(repo, subset, order=defineorder) # ^^^^^^ # For most revsets, 'define' means using the order this subset provides # # There are a few revsets that always redefine the order if 'define' is # specified: 'sort(X)', 'reverse(X)', 'x:y'. anyorder = b'any' # don't care the order, could be even random-shuffled defineorder = b'define' # ALWAYS redefine, or ALWAYS follow the current order followorder = b'follow' # MUST follow the current order # helpers def getset(repo, subset, x, order=defineorder): if not x: raise error.ParseError(_(b"missing argument")) return methods[x[0]](repo, subset, *x[1:], order=order) def _getrevsource(repo, r): extra = repo[r].extra() for label in (b'source', b'transplant_source', b'rebase_source'): if label in extra: try: return repo[extra[label]].rev() except error.RepoLookupError: pass return None def _sortedb(xs): return sorted(pycompat.rapply(pycompat.maybebytestr, xs)) # operator methods def stringset(repo, subset, x, order): if not x: raise error.ParseError(_(b"empty string is not a valid revision")) x = scmutil.intrev(scmutil.revsymbol(repo, x)) if x in subset or x in _virtualrevs and isinstance(subset, fullreposet): return baseset([x]) return baseset() def rawsmartset(repo, subset, x, order): """argument is already a smartset, use that directly""" if order == followorder: return subset & x else: return x & subset def rangeset(repo, subset, x, y, order): m = getset(repo, fullreposet(repo), x) n = getset(repo, fullreposet(repo), y) if not m or not n: return baseset() return _makerangeset(repo, subset, m.first(), n.last(), order) def rangeall(repo, subset, x, order): assert x is None return _makerangeset(repo, subset, 0, repo.changelog.tiprev(), order) def rangepre(repo, subset, y, order): # ':y' can't be rewritten to '0:y' since '0' may be hidden n = getset(repo, fullreposet(repo), y) if not n: return baseset() return _makerangeset(repo, subset, 0, n.last(), order) def rangepost(repo, subset, x, order): m = getset(repo, fullreposet(repo), x) if not m: return baseset() return _makerangeset( repo, subset, m.first(), repo.changelog.tiprev(), order ) def _makerangeset(repo, subset, m, n, order): if m == n: r = baseset([m]) elif n == wdirrev: r = spanset(repo, m, len(repo)) + baseset([n]) elif m == wdirrev: r = baseset([m]) + spanset(repo, repo.changelog.tiprev(), n - 1) elif m < n: r = spanset(repo, m, n + 1) else: r = spanset(repo, m, n - 1) if order == defineorder: return r & subset else: # carrying the sorting over when possible would be more efficient return subset & r def dagrange(repo, subset, x, y, order): r = fullreposet(repo) xs = dagop.reachableroots( repo, getset(repo, r, x), getset(repo, r, y), includepath=True ) return subset & xs def andset(repo, subset, x, y, order): if order == anyorder: yorder = anyorder else: yorder = followorder return getset(repo, getset(repo, subset, x, order), y, yorder) def andsmallyset(repo, subset, x, y, order): # 'andsmally(x, y)' is equivalent to 'and(x, y)', but faster when y is small if order == anyorder: yorder = anyorder else: yorder = followorder return getset(repo, getset(repo, subset, y, yorder), x, order) def differenceset(repo, subset, x, y, order): return getset(repo, subset, x, order) - getset(repo, subset, y, anyorder) def _orsetlist(repo, subset, xs, order): assert xs if len(xs) == 1: return getset(repo, subset, xs[0], order) p = len(xs) // 2 a = _orsetlist(repo, subset, xs[:p], order) b = _orsetlist(repo, subset, xs[p:], order) return a + b def orset(repo, subset, x, order): xs = getlist(x) if not xs: return baseset() if order == followorder: # slow path to take the subset order return subset & _orsetlist(repo, fullreposet(repo), xs, anyorder) else: return _orsetlist(repo, subset, xs, order) def notset(repo, subset, x, order): return subset - getset(repo, subset, x, anyorder) def relationset(repo, subset, x, y, order): # this is pretty basic implementation of 'x#y' operator, still # experimental so undocumented. see the wiki for further ideas. # https://www.mercurial-scm.org/wiki/RevsetOperatorPlan rel = getsymbol(y) if rel in relations: return relations[rel](repo, subset, x, rel, order) relnames = [r for r in relations.keys() if len(r) > 1] raise error.UnknownIdentifier(rel, relnames) def _splitrange(a, b): """Split range with bounds a and b into two ranges at 0 and return two tuples of numbers for use as startdepth and stopdepth arguments of revancestors and revdescendants. >>> _splitrange(-10, -5) # [-10:-5] ((5, 11), (None, None)) >>> _splitrange(5, 10) # [5:10] ((None, None), (5, 11)) >>> _splitrange(-10, 10) # [-10:10] ((0, 11), (0, 11)) >>> _splitrange(-10, 0) # [-10:0] ((0, 11), (None, None)) >>> _splitrange(0, 10) # [0:10] ((None, None), (0, 11)) >>> _splitrange(0, 0) # [0:0] ((0, 1), (None, None)) >>> _splitrange(1, -1) # [1:-1] ((None, None), (None, None)) """ ancdepths = (None, None) descdepths = (None, None) if a == b == 0: ancdepths = (0, 1) if a < 0: ancdepths = (-min(b, 0), -a + 1) if b > 0: descdepths = (max(a, 0), b + 1) return ancdepths, descdepths def generationsrel(repo, subset, x, rel, order): z = (b'rangeall', None) return generationssubrel(repo, subset, x, rel, z, order) def generationssubrel(repo, subset, x, rel, z, order): # TODO: rewrite tests, and drop startdepth argument from ancestors() and # descendants() predicates a, b = getintrange( z, _(b'relation subscript must be an integer or a range'), _(b'relation subscript bounds must be integers'), deffirst=-(dagop.maxlogdepth - 1), deflast=+(dagop.maxlogdepth - 1), ) (ancstart, ancstop), (descstart, descstop) = _splitrange(a, b) if ancstart is None and descstart is None: return baseset() revs = getset(repo, fullreposet(repo), x) if not revs: return baseset() if ancstart is not None and descstart is not None: s = dagop.revancestors(repo, revs, False, ancstart, ancstop) s += dagop.revdescendants(repo, revs, False, descstart, descstop) elif ancstart is not None: s = dagop.revancestors(repo, revs, False, ancstart, ancstop) elif descstart is not None: s = dagop.revdescendants(repo, revs, False, descstart, descstop) return subset & s def relsubscriptset(repo, subset, x, y, z, order): # this is pretty basic implementation of 'x#y[z]' operator, still # experimental so undocumented. see the wiki for further ideas. # https://www.mercurial-scm.org/wiki/RevsetOperatorPlan rel = getsymbol(y) if rel in subscriptrelations: return subscriptrelations[rel](repo, subset, x, rel, z, order) relnames = [r for r in subscriptrelations.keys() if len(r) > 1] raise error.UnknownIdentifier(rel, relnames) def subscriptset(repo, subset, x, y, order): raise error.ParseError(_(b"can't use a subscript in this context")) def listset(repo, subset, *xs, **opts): raise error.ParseError( _(b"can't use a list in this context"), hint=_(b'see \'hg help "revsets.x or y"\''), ) def keyvaluepair(repo, subset, k, v, order): raise error.ParseError(_(b"can't use a key-value pair in this context")) def func(repo, subset, a, b, order): f = getsymbol(a) if f in symbols: func = symbols[f] if getattr(func, '_takeorder', False): return func(repo, subset, b, order) return func(repo, subset, b) keep = lambda fn: getattr(fn, '__doc__', None) is not None syms = [s for (s, fn) in symbols.items() if keep(fn)] raise error.UnknownIdentifier(f, syms) # functions # symbols are callables like: # fn(repo, subset, x) # with: # repo - current repository instance # subset - of revisions to be examined # x - argument in tree form symbols = revsetlang.symbols # symbols which can't be used for a DoS attack for any given input # (e.g. those which accept regexes as plain strings shouldn't be included) # functions that just return a lot of changesets (like all) don't count here safesymbols = set() predicate = registrar.revsetpredicate() @predicate(b'_destupdate') def _destupdate(repo, subset, x): # experimental revset for update destination args = getargsdict(x, b'limit', b'clean') return subset & baseset( [destutil.destupdate(repo, **pycompat.strkwargs(args))[0]] ) @predicate(b'_destmerge') def _destmerge(repo, subset, x): # experimental revset for merge destination sourceset = None if x is not None: sourceset = getset(repo, fullreposet(repo), x) return subset & baseset([destutil.destmerge(repo, sourceset=sourceset)]) @predicate(b'adds(pattern)', safe=True, weight=30) def adds(repo, subset, x): """Changesets that add a file matching pattern. The pattern without explicit kind like ``glob:`` is expected to be relative to the current directory and match against a file or a directory. """ # i18n: "adds" is a keyword pat = getstring(x, _(b"adds requires a pattern")) return checkstatus(repo, subset, pat, 'added') @predicate(b'ancestor(*changeset)', safe=True, weight=0.5) def ancestor(repo, subset, x): """A greatest common ancestor of the changesets. Accepts 0 or more changesets. Will return empty list when passed no args. Greatest common ancestor of a single changeset is that changeset. """ reviter = iter(orset(repo, fullreposet(repo), x, order=anyorder)) try: anc = repo[next(reviter)] except StopIteration: return baseset() for r in reviter: anc = anc.ancestor(repo[r]) r = scmutil.intrev(anc) if r in subset: return baseset([r]) return baseset() def _ancestors( repo, subset, x, followfirst=False, startdepth=None, stopdepth=None ): heads = getset(repo, fullreposet(repo), x) if not heads: return baseset() s = dagop.revancestors(repo, heads, followfirst, startdepth, stopdepth) return subset & s @predicate(b'ancestors(set[, depth])', safe=True) def ancestors(repo, subset, x): """Changesets that are ancestors of changesets in set, including the given changesets themselves. If depth is specified, the result only includes changesets up to the specified generation. """ # startdepth is for internal use only until we can decide the UI args = getargsdict(x, b'ancestors', b'set depth startdepth') if b'set' not in args: # i18n: "ancestors" is a keyword raise error.ParseError(_(b'ancestors takes at least 1 argument')) startdepth = stopdepth = None if b'startdepth' in args: n = getinteger( args[b'startdepth'], b"ancestors expects an integer startdepth" ) if n < 0: raise error.ParseError(b"negative startdepth") startdepth = n if b'depth' in args: # i18n: "ancestors" is a keyword n = getinteger(args[b'depth'], _(b"ancestors expects an integer depth")) if n < 0: raise error.ParseError(_(b"negative depth")) stopdepth = n + 1 return _ancestors( repo, subset, args[b'set'], startdepth=startdepth, stopdepth=stopdepth ) @predicate(b'_firstancestors', safe=True) def _firstancestors(repo, subset, x): # ``_firstancestors(set)`` # Like ``ancestors(set)`` but follows only the first parents. return _ancestors(repo, subset, x, followfirst=True) def _childrenspec(repo, subset, x, n, order): """Changesets that are the Nth child of a changeset in set. """ cs = set() for r in getset(repo, fullreposet(repo), x): for i in range(n): c = repo[r].children() if len(c) == 0: break if len(c) > 1: raise error.RepoLookupError( _(b"revision in set has more than one child") ) r = c[0].rev() else: cs.add(r) return subset & cs def ancestorspec(repo, subset, x, n, order): """``set~n`` Changesets that are the Nth ancestor (first parents only) of a changeset in set. """ n = getinteger(n, _(b"~ expects a number")) if n < 0: # children lookup return _childrenspec(repo, subset, x, -n, order) ps = set() cl = repo.changelog for r in getset(repo, fullreposet(repo), x): for i in range(n): try: r = cl.parentrevs(r)[0] except error.WdirUnsupported: r = repo[r].p1().rev() ps.add(r) return subset & ps @predicate(b'author(string)', safe=True, weight=10) def author(repo, subset, x): """Alias for ``user(string)``.""" # i18n: "author" is a keyword n = getstring(x, _(b"author requires a string")) kind, pattern, matcher = _substringmatcher(n, casesensitive=False) return subset.filter( lambda x: matcher(repo[x].user()), condrepr=(b'<user %r>', n) ) @predicate(b'bisect(string)', safe=True) def bisect(repo, subset, x): """Changesets marked in the specified bisect status: - ``good``, ``bad``, ``skip``: csets explicitly marked as good/bad/skip - ``goods``, ``bads`` : csets topologically good/bad - ``range`` : csets taking part in the bisection - ``pruned`` : csets that are goods, bads or skipped - ``untested`` : csets whose fate is yet unknown - ``ignored`` : csets ignored due to DAG topology - ``current`` : the cset currently being bisected """ # i18n: "bisect" is a keyword status = getstring(x, _(b"bisect requires a string")).lower() state = set(hbisect.get(repo, status)) return subset & state # Backward-compatibility # - no help entry so that we do not advertise it any more @predicate(b'bisected', safe=True) def bisected(repo, subset, x): return bisect(repo, subset, x) @predicate(b'bookmark([name])', safe=True) def bookmark(repo, subset, x): """The named bookmark or all bookmarks. Pattern matching is supported for `name`. See :hg:`help revisions.patterns`. """ # i18n: "bookmark" is a keyword args = getargs(x, 0, 1, _(b'bookmark takes one or no arguments')) if args: bm = getstring( args[0], # i18n: "bookmark" is a keyword _(b'the argument to bookmark must be a string'), ) kind, pattern, matcher = stringutil.stringmatcher(bm) bms = set() if kind == b'literal': if bm == pattern: pattern = repo._bookmarks.expandname(pattern) bmrev = repo._bookmarks.get(pattern, None) if not bmrev: raise error.RepoLookupError( _(b"bookmark '%s' does not exist") % pattern ) bms.add(repo[bmrev].rev()) else: matchrevs = set() for name, bmrev in repo._bookmarks.items(): if matcher(name): matchrevs.add(bmrev) for bmrev in matchrevs: bms.add(repo[bmrev].rev()) else: bms = {repo[r].rev() for r in repo._bookmarks.values()} bms -= {nullrev} return subset & bms @predicate(b'branch(string or set)', safe=True, weight=10) def branch(repo, subset, x): """ All changesets belonging to the given branch or the branches of the given changesets. Pattern matching is supported for `string`. See :hg:`help revisions.patterns`. """ getbi = repo.revbranchcache().branchinfo def getbranch(r): try: return getbi(r)[0] except error.WdirUnsupported: return repo[r].branch() try: b = getstring(x, b'') except error.ParseError: # not a string, but another revspec, e.g. tip() pass else: kind, pattern, matcher = stringutil.stringmatcher(b) if kind == b'literal': # note: falls through to the revspec case if no branch with # this name exists and pattern kind is not specified explicitly if repo.branchmap().hasbranch(pattern): return subset.filter( lambda r: matcher(getbranch(r)), condrepr=(b'<branch %r>', b), ) if b.startswith(b'literal:'): raise error.RepoLookupError( _(b"branch '%s' does not exist") % pattern ) else: return subset.filter( lambda r: matcher(getbranch(r)), condrepr=(b'<branch %r>', b) ) s = getset(repo, fullreposet(repo), x) b = set() for r in s: b.add(getbranch(r)) c = s.__contains__ return subset.filter( lambda r: c(r) or getbranch(r) in b, condrepr=lambda: b'<branch %r>' % _sortedb(b), ) @predicate(b'phasedivergent()', safe=True) def phasedivergent(repo, subset, x): """Mutable changesets marked as successors of public changesets. Only non-public and non-obsolete changesets can be `phasedivergent`. (EXPERIMENTAL) """ # i18n: "phasedivergent" is a keyword getargs(x, 0, 0, _(b"phasedivergent takes no arguments")) phasedivergent = obsmod.getrevs(repo, b'phasedivergent') return subset & phasedivergent @predicate(b'bundle()', safe=True) def bundle(repo, subset, x): """Changesets in the bundle. Bundle must be specified by the -R option.""" try: bundlerevs = repo.changelog.bundlerevs except AttributeError: raise error.Abort(_(b"no bundle provided - specify with -R")) return subset & bundlerevs def checkstatus(repo, subset, pat, field): """Helper for status-related revsets (adds, removes, modifies). The field parameter says which kind is desired. """ hasset = matchmod.patkind(pat) == b'set' mcache = [None] def matches(x): c = repo[x] if not mcache[0] or hasset: mcache[0] = matchmod.match(repo.root, repo.getcwd(), [pat], ctx=c) m = mcache[0] fname = None assert m is not None # help pytype if not m.anypats() and len(m.files()) == 1: fname = m.files()[0] if fname is not None: if fname not in c.files(): return False else: if not any(m(f) for f in c.files()): return False files = getattr(repo.status(c.p1().node(), c.node()), field) if fname is not None: if fname in files: return True else: if any(m(f) for f in files): return True return subset.filter( matches, condrepr=(b'<status.%s %r>', pycompat.sysbytes(field), pat) ) def _children(repo, subset, parentset): if not parentset: return baseset() cs = set() pr = repo.changelog.parentrevs minrev = parentset.min() for r in subset: if r <= minrev: continue p1, p2 = pr(r) if p1 in parentset: cs.add(r) if p2 != nullrev and p2 in parentset: cs.add(r) return baseset(cs) @predicate(b'children(set)', safe=True) def children(repo, subset, x): """Child changesets of changesets in set.""" s = getset(repo, fullreposet(repo), x) cs = _children(repo, subset, s) return subset & cs @predicate(b'closed()', safe=True, weight=10) def closed(repo, subset, x): """Changeset is closed.""" # i18n: "closed" is a keyword getargs(x, 0, 0, _(b"closed takes no arguments")) return subset.filter( lambda r: repo[r].closesbranch(), condrepr=b'<branch closed>' ) # for internal use @predicate(b'_commonancestorheads(set)', safe=True) def _commonancestorheads(repo, subset, x): # This is an internal method is for quickly calculating "heads(::x and # ::y)" # These greatest common ancestors are the same ones that the consensus bid # merge will find. startrevs = getset(repo, fullreposet(repo), x, order=anyorder) ancs = repo.changelog._commonancestorsheads(*list(startrevs)) return subset & baseset(ancs) @predicate(b'commonancestors(set)', safe=True) def commonancestors(repo, subset, x): """Changesets that are ancestors of every changeset in set.""" startrevs = getset(repo, fullreposet(repo), x, order=anyorder) if not startrevs: return baseset() for r in startrevs: subset &= dagop.revancestors(repo, baseset([r])) return subset @predicate(b'conflictlocal()', safe=True) def conflictlocal(repo, subset, x): """The local side of the merge, if currently in an unresolved merge. "merge" here includes merge conflicts from e.g. 'hg rebase' or 'hg graft'. """ getargs(x, 0, 0, _(b"conflictlocal takes no arguments")) from . import mergestate as mergestatemod mergestate = mergestatemod.mergestate.read(repo) if mergestate.active() and repo.changelog.hasnode(mergestate.local): return subset & {repo.changelog.rev(mergestate.local)} return baseset() @predicate(b'conflictother()', safe=True) def conflictother(repo, subset, x): """The other side of the merge, if currently in an unresolved merge. "merge" here includes merge conflicts from e.g. 'hg rebase' or 'hg graft'. """ getargs(x, 0, 0, _(b"conflictother takes no arguments")) from . import mergestate as mergestatemod mergestate = mergestatemod.mergestate.read(repo) if mergestate.active() and repo.changelog.hasnode(mergestate.other): return subset & {repo.changelog.rev(mergestate.other)} return baseset() @predicate(b'contains(pattern)', weight=100) def contains(repo, subset, x): """The revision's manifest contains a file matching pattern (but might not modify it). See :hg:`help patterns` for information about file patterns. The pattern without explicit kind like ``glob:`` is expected to be relative to the current directory and match against a file exactly for efficiency. """ # i18n: "contains" is a keyword pat = getstring(x, _(b"contains requires a pattern")) def matches(x): if not matchmod.patkind(pat): pats = pathutil.canonpath(repo.root, repo.getcwd(), pat) if pats in repo[x]: return True else: c = repo[x] m = matchmod.match(repo.root, repo.getcwd(), [pat], ctx=c) for f in c.manifest(): if m(f): return True return False return subset.filter(matches, condrepr=(b'<contains %r>', pat)) @predicate(b'converted([id])', safe=True) def converted(repo, subset, x): """Changesets converted from the given identifier in the old repository if present, or all converted changesets if no identifier is specified. """ # There is exactly no chance of resolving the revision, so do a simple # string compare and hope for the best rev = None # i18n: "converted" is a keyword l = getargs(x, 0, 1, _(b'converted takes one or no arguments')) if l: # i18n: "converted" is a keyword rev = getstring(l[0], _(b'converted requires a revision')) def _matchvalue(r): source = repo[r].extra().get(b'convert_revision', None) return source is not None and (rev is None or source.startswith(rev)) return subset.filter( lambda r: _matchvalue(r), condrepr=(b'<converted %r>', rev) ) @predicate(b'date(interval)', safe=True, weight=10) def date(repo, subset, x): """Changesets within the interval, see :hg:`help dates`.""" # i18n: "date" is a keyword ds = getstring(x, _(b"date requires a string")) dm = dateutil.matchdate(ds) return subset.filter( lambda x: dm(repo[x].date()[0]), condrepr=(b'<date %r>', ds) ) @predicate(b'desc(string)', safe=True, weight=10) def desc(repo, subset, x): """Search commit message for string. The match is case-insensitive. Pattern matching is supported for `string`. See :hg:`help revisions.patterns`. """ # i18n: "desc" is a keyword ds = getstring(x, _(b"desc requires a string")) kind, pattern, matcher = _substringmatcher(ds, casesensitive=False) return subset.filter( lambda r: matcher(repo[r].description()), condrepr=(b'<desc %r>', ds) ) def _descendants( repo, subset, x, followfirst=False, startdepth=None, stopdepth=None ): roots = getset(repo, fullreposet(repo), x) if not roots: return baseset() s = dagop.revdescendants(repo, roots, followfirst, startdepth, stopdepth) return subset & s @predicate(b'descendants(set[, depth])', safe=True) def descendants(repo, subset, x): """Changesets which are descendants of changesets in set, including the given changesets themselves. If depth is specified, the result only includes changesets up to the specified generation. """ # startdepth is for internal use only until we can decide the UI args = getargsdict(x, b'descendants', b'set depth startdepth') if b'set' not in args: # i18n: "descendants" is a keyword raise error.ParseError(_(b'descendants takes at least 1 argument')) startdepth = stopdepth = None if b'startdepth' in args: n = getinteger( args[b'startdepth'], b"descendants expects an integer startdepth" ) if n < 0: raise error.ParseError(b"negative startdepth") startdepth = n if b'depth' in args: # i18n: "descendants" is a keyword n = getinteger( args[b'depth'], _(b"descendants expects an integer depth") ) if n < 0: raise error.ParseError(_(b"negative depth")) stopdepth = n + 1 return _descendants( repo, subset, args[b'set'], startdepth=startdepth, stopdepth=stopdepth ) @predicate(b'_firstdescendants', safe=True) def _firstdescendants(repo, subset, x): # ``_firstdescendants(set)`` # Like ``descendants(set)`` but follows only the first parents. return _descendants(repo, subset, x, followfirst=True) @predicate(b'destination([set])', safe=True, weight=10) def destination(repo, subset, x): """Changesets that were created by a graft, transplant or rebase operation, with the given revisions specified as the source. Omitting the optional set is the same as passing all(). """ if x is not None: sources = getset(repo, fullreposet(repo), x) else: sources = fullreposet(repo) dests = set() # subset contains all of the possible destinations that can be returned, so # iterate over them and see if their source(s) were provided in the arg set. # Even if the immediate src of r is not in the arg set, src's source (or # further back) may be. Scanning back further than the immediate src allows # transitive transplants and rebases to yield the same results as transitive # grafts. for r in subset: src = _getrevsource(repo, r) lineage = None while src is not None: if lineage is None: lineage = list() lineage.append(r) # The visited lineage is a match if the current source is in the arg # set. Since every candidate dest is visited by way of iterating # subset, any dests further back in the lineage will be tested by a # different iteration over subset. Likewise, if the src was already # selected, the current lineage can be selected without going back # further. if src in sources or src in dests: dests.update(lineage) break r = src src = _getrevsource(repo, r) return subset.filter( dests.__contains__, condrepr=lambda: b'<destination %r>' % _sortedb(dests), ) @predicate(b'diffcontains(pattern)', weight=110) def diffcontains(repo, subset, x): """Search revision differences for when the pattern was added or removed. The pattern may be a substring literal or a regular expression. See :hg:`help revisions.patterns`. """ args = getargsdict(x, b'diffcontains', b'pattern') if b'pattern' not in args: # i18n: "diffcontains" is a keyword raise error.ParseError(_(b'diffcontains takes at least 1 argument')) pattern = getstring( args[b'pattern'], _(b'diffcontains requires a string pattern') ) regexp = stringutil.substringregexp(pattern, re.M) # TODO: add support for file pattern and --follow. For example, # diffcontains(pattern[, set]) where set may be file(pattern) or # follow(pattern), and we'll eventually add a support for narrowing # files by revset? fmatch = matchmod.always() def makefilematcher(ctx): return fmatch # TODO: search in a windowed way searcher = grepmod.grepsearcher(repo.ui, repo, regexp, diff=True) def testdiff(rev): # consume the generator to discard revfiles/matches cache found = False for fn, ctx, pstates, states in searcher.searchfiles( baseset([rev]), makefilematcher ): if next(grepmod.difflinestates(pstates, states), None): found = True return found return subset.filter(testdiff, condrepr=(b'<diffcontains %r>', pattern)) @predicate(b'contentdivergent()', safe=True) def contentdivergent(repo, subset, x): """ Final successors of changesets with an alternative set of final successors. (EXPERIMENTAL) """ # i18n: "contentdivergent" is a keyword getargs(x, 0, 0, _(b"contentdivergent takes no arguments")) contentdivergent = obsmod.getrevs(repo, b'contentdivergent') return subset & contentdivergent @predicate(b'expectsize(set[, size])', safe=True, takeorder=True) def expectsize(repo, subset, x, order): """Return the given revset if size matches the revset size. Abort if the revset doesn't expect given size. size can either be an integer range or an integer. For example, ``expectsize(0:1, 3:5)`` will abort as revset size is 2 and 2 is not between 3 and 5 inclusive.""" args = getargsdict(x, b'expectsize', b'set size') minsize = 0 maxsize = len(repo) + 1 err = b'' if b'size' not in args or b'set' not in args: raise error.ParseError(_(b'invalid set of arguments')) minsize, maxsize = getintrange( args[b'size'], _(b'expectsize requires a size range or a positive integer'), _(b'size range bounds must be integers'), minsize, maxsize, ) if minsize < 0 or maxsize < 0: raise error.ParseError(_(b'negative size')) rev = getset(repo, fullreposet(repo), args[b'set'], order=order) if minsize != maxsize and (len(rev) < minsize or len(rev) > maxsize): err = _(b'revset size mismatch. expected between %d and %d, got %d') % ( minsize, maxsize, len(rev), ) elif minsize == maxsize and len(rev) != minsize: err = _(b'revset size mismatch. expected %d, got %d') % ( minsize, len(rev), ) if err: raise error.RepoLookupError(err) if order == followorder: return subset & rev else: return rev & subset @predicate(b'extdata(source)', safe=False, weight=100) def extdata(repo, subset, x): """Changesets in the specified extdata source. (EXPERIMENTAL)""" # i18n: "extdata" is a keyword args = getargsdict(x, b'extdata', b'source') source = getstring( args.get(b'source'), # i18n: "extdata" is a keyword _(b'extdata takes at least 1 string argument'), ) data = scmutil.extdatasource(repo, source) return subset & baseset(data) @predicate(b'extinct()', safe=True) def extinct(repo, subset, x): """Obsolete changesets with obsolete descendants only. (EXPERIMENTAL)""" # i18n: "extinct" is a keyword getargs(x, 0, 0, _(b"extinct takes no arguments")) extincts = obsmod.getrevs(repo, b'extinct') return subset & extincts @predicate(b'extra(label, [value])', safe=True) def extra(repo, subset, x): """Changesets with the given label in the extra metadata, with the given optional value. Pattern matching is supported for `value`. See :hg:`help revisions.patterns`. """ args = getargsdict(x, b'extra', b'label value') if b'label' not in args: # i18n: "extra" is a keyword raise error.ParseError(_(b'extra takes at least 1 argument')) # i18n: "extra" is a keyword label = getstring( args[b'label'], _(b'first argument to extra must be a string') ) value = None if b'value' in args: # i18n: "extra" is a keyword value = getstring( args[b'value'], _(b'second argument to extra must be a string') ) kind, value, matcher = stringutil.stringmatcher(value) def _matchvalue(r): extra = repo[r].extra() return label in extra and (value is None or matcher(extra[label])) return subset.filter( lambda r: _matchvalue(r), condrepr=(b'<extra[%r] %r>', label, value) ) @predicate(b'filelog(pattern)', safe=True) def filelog(repo, subset, x): """Changesets connected to the specified filelog. For performance reasons, visits only revisions mentioned in the file-level filelog, rather than filtering through all changesets (much faster, but doesn't include deletes or duplicate changes). For a slower, more accurate result, use ``file()``. The pattern without explicit kind like ``glob:`` is expected to be relative to the current directory and match against a file exactly for efficiency. """ # i18n: "filelog" is a keyword pat = getstring(x, _(b"filelog requires a pattern")) s = set() cl = repo.changelog if not matchmod.patkind(pat): f = pathutil.canonpath(repo.root, repo.getcwd(), pat) files = [f] else: m = matchmod.match(repo.root, repo.getcwd(), [pat], ctx=repo[None]) files = (f for f in repo[None] if m(f)) for f in files: fl = repo.file(f) known = {} scanpos = 0 for fr in list(fl): fn = fl.node(fr) if fn in known: s.add(known[fn]) continue lr = fl.linkrev(fr) if lr in cl: s.add(lr) elif scanpos is not None: # lowest matching changeset is filtered, scan further # ahead in changelog start = max(lr, scanpos) + 1 scanpos = None for r in cl.revs(start): # minimize parsing of non-matching entries if f in cl.revision(r) and f in cl.readfiles(r): try: # try to use manifest delta fastpath n = repo[r].filenode(f) if n not in known: if n == fn: s.add(r) scanpos = r break else: known[n] = r except error.ManifestLookupError: # deletion in changelog continue return subset & s @predicate(b'first(set, [n])', safe=True, takeorder=True, weight=0) def first(repo, subset, x, order): """An alias for limit().""" return limit(repo, subset, x, order) def _follow(repo, subset, x, name, followfirst=False): args = getargsdict(x, name, b'file startrev') revs = None if b'startrev' in args: revs = getset(repo, fullreposet(repo), args[b'startrev']) if b'file' in args: x = getstring(args[b'file'], _(b"%s expected a pattern") % name) if revs is None: revs = [None] fctxs = [] for r in revs: ctx = mctx = repo[r] if r is None: ctx = repo[b'.'] m = matchmod.match( repo.root, repo.getcwd(), [x], ctx=mctx, default=b'path' ) fctxs.extend(ctx[f].introfilectx() for f in ctx.manifest().walk(m)) s = dagop.filerevancestors(fctxs, followfirst) else: if revs is None: revs = baseset([repo[b'.'].rev()]) s = dagop.revancestors(repo, revs, followfirst) return subset & s @predicate(b'follow([file[, startrev]])', safe=True) def follow(repo, subset, x): """ An alias for ``::.`` (ancestors of the working directory's first parent). If file pattern is specified, the histories of files matching given pattern in the revision given by startrev are followed, including copies. """ return _follow(repo, subset, x, b'follow') @predicate(b'_followfirst', safe=True) def _followfirst(repo, subset, x): # ``followfirst([file[, startrev]])`` # Like ``follow([file[, startrev]])`` but follows only the first parent # of every revisions or files revisions. return _follow(repo, subset, x, b'_followfirst', followfirst=True) @predicate( b'followlines(file, fromline:toline[, startrev=., descend=False])', safe=True, ) def followlines(repo, subset, x): """Changesets modifying `file` in line range ('fromline', 'toline'). Line range corresponds to 'file' content at 'startrev' and should hence be consistent with file size. If startrev is not specified, working directory's parent is used. By default, ancestors of 'startrev' are returned. If 'descend' is True, descendants of 'startrev' are returned though renames are (currently) not followed in this direction. """ args = getargsdict(x, b'followlines', b'file *lines startrev descend') if len(args[b'lines']) != 1: raise error.ParseError(_(b"followlines requires a line range")) rev = b'.' if b'startrev' in args: revs = getset(repo, fullreposet(repo), args[b'startrev']) if len(revs) != 1: raise error.ParseError( # i18n: "followlines" is a keyword _(b"followlines expects exactly one revision") ) rev = revs.last() pat = getstring(args[b'file'], _(b"followlines requires a pattern")) # i18n: "followlines" is a keyword msg = _(b"followlines expects exactly one file") fname = scmutil.parsefollowlinespattern(repo, rev, pat, msg) fromline, toline = util.processlinerange( *getintrange( args[b'lines'][0], # i18n: "followlines" is a keyword _(b"followlines expects a line number or a range"), _(b"line range bounds must be integers"), ) ) fctx = repo[rev].filectx(fname) descend = False if b'descend' in args: descend = getboolean( args[b'descend'], # i18n: "descend" is a keyword _(b"descend argument must be a boolean"), ) if descend: rs = generatorset( ( c.rev() for c, _linerange in dagop.blockdescendants( fctx, fromline, toline ) ), iterasc=True, ) else: rs = generatorset( ( c.rev() for c, _linerange in dagop.blockancestors( fctx, fromline, toline ) ), iterasc=False, ) return subset & rs @predicate(b'nodefromfile(path)') def nodefromfile(repo, subset, x): """Read a list of nodes from the file at `path`. This applies `id(LINE)` to each line of the file. This is useful when the amount of nodes you need to specify gets too large for the command line. """ path = getstring(x, _(b"nodefromfile require a file path")) listed_rev = set() try: with pycompat.open(path, 'rb') as f: for line in f: n = line.strip() rn = _node(repo, n) if rn is not None: listed_rev.add(rn) except IOError as exc: m = _(b'cannot open nodes file "%s": %s') m %= (path, encoding.strtolocal(exc.strerror)) raise error.Abort(m) return subset & baseset(listed_rev) @predicate(b'all()', safe=True) def getall(repo, subset, x): """All changesets, the same as ``0:tip``.""" # i18n: "all" is a keyword getargs(x, 0, 0, _(b"all takes no arguments")) return subset & spanset(repo) # drop "null" if any @predicate(b'grep(regex)', weight=10) def grep(repo, subset, x): """Like ``keyword(string)`` but accepts a regex. Use ``grep(r'...')`` to ensure special escape characters are handled correctly. Unlike ``keyword(string)``, the match is case-sensitive. """ try: # i18n: "grep" is a keyword gr = re.compile(getstring(x, _(b"grep requires a string"))) except re.error as e: raise error.ParseError( _(b'invalid match pattern: %s') % stringutil.forcebytestr(e) ) def matches(x): c = repo[x] for e in c.files() + [c.user(), c.description()]: if gr.search(e): return True return False return subset.filter(matches, condrepr=(b'<grep %r>', gr.pattern)) @predicate(b'_matchfiles', safe=True) def _matchfiles(repo, subset, x): # _matchfiles takes a revset list of prefixed arguments: # # [p:foo, i:bar, x:baz] # # builds a match object from them and filters subset. Allowed # prefixes are 'p:' for regular patterns, 'i:' for include # patterns and 'x:' for exclude patterns. Use 'r:' prefix to pass # a revision identifier, or the empty string to reference the # working directory, from which the match object is # initialized. Use 'd:' to set the default matching mode, default # to 'glob'. At most one 'r:' and 'd:' argument can be passed. l = getargs(x, 1, -1, b"_matchfiles requires at least one argument") pats, inc, exc = [], [], [] rev, default = None, None for arg in l: s = getstring(arg, b"_matchfiles requires string arguments") prefix, value = s[:2], s[2:] if prefix == b'p:': pats.append(value) elif prefix == b'i:': inc.append(value) elif prefix == b'x:': exc.append(value) elif prefix == b'r:': if rev is not None: raise error.ParseError( b'_matchfiles expected at most one revision' ) if value == b'': # empty means working directory rev = wdirrev else: rev = value elif prefix == b'd:': if default is not None: raise error.ParseError( b'_matchfiles expected at most one default mode' ) default = value else: raise error.ParseError(b'invalid _matchfiles prefix: %s' % prefix) if not default: default = b'glob' hasset = any(matchmod.patkind(p) == b'set' for p in pats + inc + exc) mcache = [None] # This directly read the changelog data as creating changectx for all # revisions is quite expensive. getfiles = repo.changelog.readfiles def matches(x): if x == wdirrev: files = repo[x].files() else: files = getfiles(x) if not mcache[0] or (hasset and rev is None): r = x if rev is None else rev mcache[0] = matchmod.match( repo.root, repo.getcwd(), pats, include=inc, exclude=exc, ctx=repo[r], default=default, ) m = mcache[0] for f in files: if m(f): return True return False return subset.filter( matches, condrepr=( b'<matchfiles patterns=%r, include=%r ' b'exclude=%r, default=%r, rev=%r>', pats, inc, exc, default, rev, ), ) @predicate(b'file(pattern)', safe=True, weight=10) def hasfile(repo, subset, x): """Changesets affecting files matched by pattern. For a faster but less accurate result, consider using ``filelog()`` instead. This predicate uses ``glob:`` as the default kind of pattern. """ # i18n: "file" is a keyword pat = getstring(x, _(b"file requires a pattern")) return _matchfiles(repo, subset, (b'string', b'p:' + pat)) @predicate(b'head()', safe=True) def head(repo, subset, x): """Changeset is a named branch head.""" # i18n: "head" is a keyword getargs(x, 0, 0, _(b"head takes no arguments")) hs = set() cl = repo.changelog for ls in repo.branchmap().iterheads(): hs.update(cl.rev(h) for h in ls) return subset & baseset(hs) @predicate(b'heads(set)', safe=True, takeorder=True) def heads(repo, subset, x, order): """Members of set with no children in set.""" # argument set should never define order if order == defineorder: order = followorder inputset = getset(repo, fullreposet(repo), x, order=order) wdirparents = None if wdirrev in inputset: # a bit slower, but not common so good enough for now wdirparents = [p.rev() for p in repo[None].parents()] inputset = set(inputset) inputset.discard(wdirrev) heads = repo.changelog.headrevs(inputset) if wdirparents is not None: heads.difference_update(wdirparents) heads.add(wdirrev) heads = baseset(heads) return subset & heads @predicate(b'hidden()', safe=True) def hidden(repo, subset, x): """Hidden changesets.""" # i18n: "hidden" is a keyword getargs(x, 0, 0, _(b"hidden takes no arguments")) hiddenrevs = repoview.filterrevs(repo, b'visible') return subset & hiddenrevs @predicate(b'keyword(string)', safe=True, weight=10) def keyword(repo, subset, x): """Search commit message, user name, and names of changed files for string. The match is case-insensitive. For a regular expression or case sensitive search of these fields, use ``grep(regex)``. """ # i18n: "keyword" is a keyword kw = encoding.lower(getstring(x, _(b"keyword requires a string"))) def matches(r): c = repo[r] return any( kw in encoding.lower(t) for t in c.files() + [c.user(), c.description()] ) return subset.filter(matches, condrepr=(b'<keyword %r>', kw)) @predicate(b'limit(set[, n[, offset]])', safe=True, takeorder=True, weight=0) def limit(repo, subset, x, order): """First n members of set, defaulting to 1, starting from offset.""" args = getargsdict(x, b'limit', b'set n offset') if b'set' not in args: # i18n: "limit" is a keyword raise error.ParseError(_(b"limit requires one to three arguments")) # i18n: "limit" is a keyword lim = getinteger(args.get(b'n'), _(b"limit expects a number"), default=1) if lim < 0: raise error.ParseError(_(b"negative number to select")) # i18n: "limit" is a keyword ofs = getinteger( args.get(b'offset'), _(b"limit expects a number"), default=0 ) if ofs < 0: raise error.ParseError(_(b"negative offset")) os = getset(repo, fullreposet(repo), args[b'set']) ls = os.slice(ofs, ofs + lim) if order == followorder and lim > 1: return subset & ls return ls & subset @predicate(b'last(set, [n])', safe=True, takeorder=True) def last(repo, subset, x, order): """Last n members of set, defaulting to 1.""" # i18n: "last" is a keyword l = getargs(x, 1, 2, _(b"last requires one or two arguments")) lim = 1 if len(l) == 2: # i18n: "last" is a keyword lim = getinteger(l[1], _(b"last expects a number")) if lim < 0: raise error.ParseError(_(b"negative number to select")) os = getset(repo, fullreposet(repo), l[0]) os.reverse() ls = os.slice(0, lim) if order == followorder and lim > 1: return subset & ls ls.reverse() return ls & subset @predicate(b'max(set)', safe=True) def maxrev(repo, subset, x): """Changeset with highest revision number in set.""" os = getset(repo, fullreposet(repo), x) try: m = os.max() if m in subset: return baseset([m], datarepr=(b'<max %r, %r>', subset, os)) except ValueError: # os.max() throws a ValueError when the collection is empty. # Same as python's max(). pass return baseset(datarepr=(b'<max %r, %r>', subset, os)) @predicate(b'merge()', safe=True) def merge(repo, subset, x): """Changeset is a merge changeset.""" # i18n: "merge" is a keyword getargs(x, 0, 0, _(b"merge takes no arguments")) cl = repo.changelog def ismerge(r): try: return cl.parentrevs(r)[1] != nullrev except error.WdirUnsupported: return bool(repo[r].p2()) return subset.filter(ismerge, condrepr=b'<merge>') @predicate(b'branchpoint()', safe=True) def branchpoint(repo, subset, x): """Changesets with more than one child.""" # i18n: "branchpoint" is a keyword getargs(x, 0, 0, _(b"branchpoint takes no arguments")) cl = repo.changelog if not subset: return baseset() # XXX this should be 'parentset.min()' assuming 'parentset' is a smartset # (and if it is not, it should.) baserev = min(subset) parentscount = [0] * (len(repo) - baserev) for r in cl.revs(start=baserev + 1): for p in cl.parentrevs(r): if p >= baserev: parentscount[p - baserev] += 1 return subset.filter( lambda r: parentscount[r - baserev] > 1, condrepr=b'<branchpoint>' ) @predicate(b'min(set)', safe=True) def minrev(repo, subset, x): """Changeset with lowest revision number in set.""" os = getset(repo, fullreposet(repo), x) try: m = os.min() if m in subset: return baseset([m], datarepr=(b'<min %r, %r>', subset, os)) except ValueError: # os.min() throws a ValueError when the collection is empty. # Same as python's min(). pass return baseset(datarepr=(b'<min %r, %r>', subset, os)) @predicate(b'modifies(pattern)', safe=True, weight=30) def modifies(repo, subset, x): """Changesets modifying files matched by pattern. The pattern without explicit kind like ``glob:`` is expected to be relative to the current directory and match against a file or a directory. """ # i18n: "modifies" is a keyword pat = getstring(x, _(b"modifies requires a pattern")) return checkstatus(repo, subset, pat, 'modified') @predicate(b'named(namespace)') def named(repo, subset, x): """The changesets in a given namespace. Pattern matching is supported for `namespace`. See :hg:`help revisions.patterns`. """ # i18n: "named" is a keyword args = getargs(x, 1, 1, _(b'named requires a namespace argument')) ns = getstring( args[0], # i18n: "named" is a keyword _(b'the argument to named must be a string'), ) kind, pattern, matcher = stringutil.stringmatcher(ns) namespaces = set() if kind == b'literal': if pattern not in repo.names: raise error.RepoLookupError( _(b"namespace '%s' does not exist") % ns ) namespaces.add(repo.names[pattern]) else: for name, ns in repo.names.items(): if matcher(name): namespaces.add(ns) names = set() for ns in namespaces: for name in ns.listnames(repo): if name not in ns.deprecated: names.update(repo[n].rev() for n in ns.nodes(repo, name)) names -= {nullrev} return subset & names def _node(repo, n): """process a node input""" rn = None if len(n) == 2 * repo.nodeconstants.nodelen: try: rn = repo.changelog.rev(bin(n)) except error.WdirUnsupported: rn = wdirrev except (binascii.Error, LookupError): rn = None else: try: pm = scmutil.resolvehexnodeidprefix(repo, n) if pm is not None: rn = repo.changelog.rev(pm) except LookupError: pass except error.WdirUnsupported: rn = wdirrev return rn @predicate(b'id(string)', safe=True) def node_(repo, subset, x): """Revision non-ambiguously specified by the given hex string prefix.""" # i18n: "id" is a keyword l = getargs(x, 1, 1, _(b"id requires one argument")) # i18n: "id" is a keyword n = getstring(l[0], _(b"id requires a string")) rn = _node(repo, n) if rn is None: return baseset() result = baseset([rn]) return result & subset @predicate(b'none()', safe=True) def none(repo, subset, x): """No changesets.""" # i18n: "none" is a keyword getargs(x, 0, 0, _(b"none takes no arguments")) return baseset() @predicate(b'obsolete()', safe=True) def obsolete(repo, subset, x): """Mutable changeset with a newer version. (EXPERIMENTAL)""" # i18n: "obsolete" is a keyword getargs(x, 0, 0, _(b"obsolete takes no arguments")) obsoletes = obsmod.getrevs(repo, b'obsolete') return subset & obsoletes @predicate(b'only(set, [set])', safe=True) def only(repo, subset, x): """Changesets that are ancestors of the first set that are not ancestors of any other head in the repo. If a second set is specified, the result is ancestors of the first set that are not ancestors of the second set (i.e. ::<set1> - ::<set2>). """ cl = repo.changelog # i18n: "only" is a keyword args = getargs(x, 1, 2, _(b'only takes one or two arguments')) include = getset(repo, fullreposet(repo), args[0]) if len(args) == 1: if not include: return baseset() descendants = set(dagop.revdescendants(repo, include, False)) exclude = [ rev for rev in cl.headrevs() if not rev in descendants and not rev in include ] else: exclude = getset(repo, fullreposet(repo), args[1]) results = set(cl.findmissingrevs(common=exclude, heads=include)) # XXX we should turn this into a baseset instead of a set, smartset may do # some optimizations from the fact this is a baseset. return subset & results @predicate(b'origin([set])', safe=True) def origin(repo, subset, x): """ Changesets that were specified as a source for the grafts, transplants or rebases that created the given revisions. Omitting the optional set is the same as passing all(). If a changeset created by these operations is itself specified as a source for one of these operations, only the source changeset for the first operation is selected. """ if x is not None: dests = getset(repo, fullreposet(repo), x) else: dests = fullreposet(repo) def _firstsrc(rev): src = _getrevsource(repo, rev) if src is None: return None while True: prev = _getrevsource(repo, src) if prev is None: return src src = prev o = {_firstsrc(r) for r in dests} o -= {None} # XXX we should turn this into a baseset instead of a set, smartset may do # some optimizations from the fact this is a baseset. return subset & o @predicate(b'outgoing([path])', safe=False, weight=10) def outgoing(repo, subset, x): """Changesets not found in the specified destination repository, or the default push location. If the location resolve to multiple repositories, the union of all outgoing changeset will be used. """ # Avoid cycles. from . import ( discovery, hg, ) # i18n: "outgoing" is a keyword l = getargs(x, 0, 1, _(b"outgoing takes one or no arguments")) # i18n: "outgoing" is a keyword dest = ( l and getstring(l[0], _(b"outgoing requires a repository path")) or b'' ) if dest: dests = [dest] else: dests = [] missing = set() for path in urlutil.get_push_paths(repo, repo.ui, dests): branches = path.branch, [] revs, checkout = hg.addbranchrevs(repo, repo, branches, []) if revs: revs = [repo.lookup(rev) for rev in revs] other = hg.peer(repo, {}, path) try: with repo.ui.silent(): outgoing = discovery.findcommonoutgoing( repo, other, onlyheads=revs ) finally: other.close() missing.update(outgoing.missing) cl = repo.changelog o = {cl.rev(r) for r in missing} return subset & o @predicate(b'p1([set])', safe=True) def p1(repo, subset, x): """First parent of changesets in set, or the working directory.""" if x is None: p = repo[x].p1().rev() if p >= 0: return subset & baseset([p]) return baseset() ps = set() cl = repo.changelog for r in getset(repo, fullreposet(repo), x): try: ps.add(cl.parentrevs(r)[0]) except error.WdirUnsupported: ps.add(repo[r].p1().rev()) ps -= {nullrev} # XXX we should turn this into a baseset instead of a set, smartset may do # some optimizations from the fact this is a baseset. return subset & ps @predicate(b'p2([set])', safe=True) def p2(repo, subset, x): """Second parent of changesets in set, or the working directory.""" if x is None: ps = repo[x].parents() try: p = ps[1].rev() if p >= 0: return subset & baseset([p]) return baseset() except IndexError: return baseset() ps = set() cl = repo.changelog for r in getset(repo, fullreposet(repo), x): try: ps.add(cl.parentrevs(r)[1]) except error.WdirUnsupported: parents = repo[r].parents() if len(parents) == 2: ps.add(parents[1]) ps -= {nullrev} # XXX we should turn this into a baseset instead of a set, smartset may do # some optimizations from the fact this is a baseset. return subset & ps def parentpost(repo, subset, x, order): return p1(repo, subset, x) @predicate(b'parents([set])', safe=True) def parents(repo, subset, x): """ The set of all parents for all changesets in set, or the working directory. """ if x is None: ps = {p.rev() for p in repo[x].parents()} else: ps = set() cl = repo.changelog up = ps.update parentrevs = cl.parentrevs for r in getset(repo, fullreposet(repo), x): try: up(parentrevs(r)) except error.WdirUnsupported: up(p.rev() for p in repo[r].parents()) ps -= {nullrev} return subset & ps def _phase(repo, subset, *targets): """helper to select all rev in <targets> phases""" return repo._phasecache.getrevset(repo, targets, subset) @predicate(b'_internal()', safe=True) def _internal(repo, subset, x): getargs(x, 0, 0, _(b"_internal takes no arguments")) return _phase(repo, subset, *phases.all_internal_phases) @predicate(b'_phase(idx)', safe=True) def phase(repo, subset, x): l = getargs(x, 1, 1, b"_phase requires one argument") target = getinteger(l[0], b"_phase expects a number") return _phase(repo, subset, target) @predicate(b'draft()', safe=True) def draft(repo, subset, x): """Changeset in draft phase.""" # i18n: "draft" is a keyword getargs(x, 0, 0, _(b"draft takes no arguments")) target = phases.draft return _phase(repo, subset, target) @predicate(b'secret()', safe=True) def secret(repo, subset, x): """Changeset in secret phase.""" # i18n: "secret" is a keyword getargs(x, 0, 0, _(b"secret takes no arguments")) target = phases.secret return _phase(repo, subset, target) @predicate(b'stack([revs])', safe=True) def stack(repo, subset, x): """Experimental revset for the stack of changesets or working directory parent. (EXPERIMENTAL) """ if x is None: stacks = stackmod.getstack(repo) else: stacks = smartset.baseset([]) for revision in getset(repo, fullreposet(repo), x): currentstack = stackmod.getstack(repo, revision) stacks = stacks + currentstack return subset & stacks def parentspec(repo, subset, x, n, order): """``set^0`` The set. ``set^1`` (or ``set^``), ``set^2`` First or second parent, respectively, of all changesets in set. """ try: n = int(n[1]) if n not in (0, 1, 2): raise ValueError except (TypeError, ValueError): raise error.ParseError(_(b"^ expects a number 0, 1, or 2")) ps = set() cl = repo.changelog for r in getset(repo, fullreposet(repo), x): if n == 0: ps.add(r) elif n == 1: try: ps.add(cl.parentrevs(r)[0]) except error.WdirUnsupported: ps.add(repo[r].p1().rev()) else: try: parents = cl.parentrevs(r) if parents[1] != nullrev: ps.add(parents[1]) except error.WdirUnsupported: parents = repo[r].parents() if len(parents) == 2: ps.add(parents[1].rev()) return subset & ps @predicate(b'present(set)', safe=True, takeorder=True) def present(repo, subset, x, order): """An empty set, if any revision in set isn't found; otherwise, all revisions in set. If any of specified revisions is not present in the local repository, the query is normally aborted. But this predicate allows the query to continue even in such cases. """ try: return getset(repo, subset, x, order) except error.RepoLookupError: return baseset() # for internal use @predicate(b'_notpublic', safe=True) def _notpublic(repo, subset, x): getargs(x, 0, 0, b"_notpublic takes no arguments") return _phase(repo, subset, *phases.not_public_phases) # for internal use @predicate(b'_phaseandancestors(phasename, set)', safe=True) def _phaseandancestors(repo, subset, x): # equivalent to (phasename() & ancestors(set)) but more efficient # phasename could be one of 'draft', 'secret', or '_notpublic' args = getargs(x, 2, 2, b"_phaseandancestors requires two arguments") phasename = getsymbol(args[0]) s = getset(repo, fullreposet(repo), args[1]) draft = phases.draft secret = phases.secret phasenamemap = { b'_notpublic': draft, b'draft': draft, # follow secret's ancestors b'secret': secret, } if phasename not in phasenamemap: raise error.ParseError(b'%r is not a valid phasename' % phasename) minimalphase = phasenamemap[phasename] getphase = repo._phasecache.phase def cutfunc(rev): return getphase(repo, rev) < minimalphase revs = dagop.revancestors(repo, s, cutfunc=cutfunc) if phasename == b'draft': # need to remove secret changesets revs = revs.filter(lambda r: getphase(repo, r) == draft) return subset & revs @predicate(b'public()', safe=True) def public(repo, subset, x): """Changeset in public phase.""" # i18n: "public" is a keyword getargs(x, 0, 0, _(b"public takes no arguments")) return _phase(repo, subset, phases.public) @predicate(b'remote([id [,path]])', safe=False) def remote(repo, subset, x): """Local revision that corresponds to the given identifier in a remote repository, if present. Here, the '.' identifier is a synonym for the current local branch. """ from . import hg # avoid start-up nasties # i18n: "remote" is a keyword l = getargs(x, 0, 2, _(b"remote takes zero, one, or two arguments")) q = b'.' if len(l) > 0: # i18n: "remote" is a keyword q = getstring(l[0], _(b"remote requires a string id")) if q == b'.': q = repo[b'.'].branch() dest = b'' if len(l) > 1: # i18n: "remote" is a keyword dest = getstring(l[1], _(b"remote requires a repository path")) if not dest: dest = b'default' path = urlutil.get_unique_pull_path_obj(b'remote', repo.ui, dest) other = hg.peer(repo, {}, path) n = other.lookup(q) if n in repo: r = repo[n].rev() if r in subset: return baseset([r]) return baseset() @predicate(b'removes(pattern)', safe=True, weight=30) def removes(repo, subset, x): """Changesets which remove files matching pattern. The pattern without explicit kind like ``glob:`` is expected to be relative to the current directory and match against a file or a directory. """ # i18n: "removes" is a keyword pat = getstring(x, _(b"removes requires a pattern")) return checkstatus(repo, subset, pat, 'removed') @predicate(b'rev(number)', safe=True) def rev(repo, subset, x): """Revision with the given numeric identifier.""" try: return _rev(repo, subset, x) except error.RepoLookupError: return baseset() @predicate(b'_rev(number)', safe=True) def _rev(repo, subset, x): # internal version of "rev(x)" that raise error if "x" is invalid # i18n: "rev" is a keyword l = getargs(x, 1, 1, _(b"rev requires one argument")) try: # i18n: "rev" is a keyword l = int(getstring(l[0], _(b"rev requires a number"))) except (TypeError, ValueError): # i18n: "rev" is a keyword raise error.ParseError(_(b"rev expects a number")) if l not in _virtualrevs: try: repo.changelog.node(l) # check that the rev exists except IndexError: raise error.RepoLookupError(_(b"unknown revision '%d'") % l) return subset & baseset([l]) @predicate(b'revset(set)', safe=True, takeorder=True) def revsetpredicate(repo, subset, x, order): """Strictly interpret the content as a revset. The content of this special predicate will be strictly interpreted as a revset. For example, ``revset(id(0))`` will be interpreted as "id(0)" without possible ambiguity with a "id(0)" bookmark or tag. """ return getset(repo, subset, x, order) @predicate(b'matching(revision [, field])', safe=True) def matching(repo, subset, x): """Changesets in which a given set of fields match the set of fields in the selected revision or set. To match more than one field pass the list of fields to match separated by spaces (e.g. ``author description``). Valid fields are most regular revision fields and some special fields. Regular revision fields are ``description``, ``author``, ``branch``, ``date``, ``files``, ``phase``, ``parents``, ``substate``, ``user`` and ``diff``. Note that ``author`` and ``user`` are synonyms. ``diff`` refers to the contents of the revision. Two revisions matching their ``diff`` will also match their ``files``. Special fields are ``summary`` and ``metadata``: ``summary`` matches the first line of the description. ``metadata`` is equivalent to matching ``description user date`` (i.e. it matches the main metadata fields). ``metadata`` is the default field which is used when no fields are specified. You can match more than one field at a time. """ # i18n: "matching" is a keyword l = getargs(x, 1, 2, _(b"matching takes 1 or 2 arguments")) revs = getset(repo, fullreposet(repo), l[0]) fieldlist = [b'metadata'] if len(l) > 1: fieldlist = getstring( l[1], # i18n: "matching" is a keyword _(b"matching requires a string as its second argument"), ).split() # Make sure that there are no repeated fields, # expand the 'special' 'metadata' field type # and check the 'files' whenever we check the 'diff' fields = [] for field in fieldlist: if field == b'metadata': fields += [b'user', b'description', b'date'] elif field == b'diff': # a revision matching the diff must also match the files # since matching the diff is very costly, make sure to # also match the files first fields += [b'files', b'diff'] else: if field == b'author': field = b'user' fields.append(field) fields = set(fields) if b'summary' in fields and b'description' in fields: # If a revision matches its description it also matches its summary fields.discard(b'summary') # We may want to match more than one field # Not all fields take the same amount of time to be matched # Sort the selected fields in order of increasing matching cost fieldorder = [ b'phase', b'parents', b'user', b'date', b'branch', b'summary', b'files', b'description', b'substate', b'diff', ] def fieldkeyfunc(f): try: return fieldorder.index(f) except ValueError: # assume an unknown field is very costly return len(fieldorder) fields = list(fields) fields.sort(key=fieldkeyfunc) # Each field will be matched with its own "getfield" function # which will be added to the getfieldfuncs array of functions getfieldfuncs = [] _funcs = { b'user': lambda r: repo[r].user(), b'branch': lambda r: repo[r].branch(), b'date': lambda r: repo[r].date(), b'description': lambda r: repo[r].description(), b'files': lambda r: repo[r].files(), b'parents': lambda r: repo[r].parents(), b'phase': lambda r: repo[r].phase(), b'substate': lambda r: repo[r].substate, b'summary': lambda r: repo[r].description().splitlines()[0], b'diff': lambda r: list( repo[r].diff(opts=diffutil.diffallopts(repo.ui, {b'git': True})) ), } for info in fields: getfield = _funcs.get(info, None) if getfield is None: raise error.ParseError( # i18n: "matching" is a keyword _(b"unexpected field name passed to matching: %s") % info ) getfieldfuncs.append(getfield) # convert the getfield array of functions into a "getinfo" function # which returns an array of field values (or a single value if there # is only one field to match) getinfo = lambda r: [f(r) for f in getfieldfuncs] def matches(x): for rev in revs: target = getinfo(rev) match = True for n, f in enumerate(getfieldfuncs): if target[n] != f(x): match = False if match: return True return False return subset.filter(matches, condrepr=(b'<matching%r %r>', fields, revs)) @predicate(b'reverse(set)', safe=True, takeorder=True, weight=0) def reverse(repo, subset, x, order): """Reverse order of set.""" l = getset(repo, subset, x, order) if order == defineorder: l.reverse() return l @predicate(b'roots(set)', safe=True) def roots(repo, subset, x): """Changesets in set with no parent changeset in set.""" s = getset(repo, fullreposet(repo), x) parents = repo.changelog.parentrevs def filter(r): try: for p in parents(r): if 0 <= p and p in s: return False except error.WdirUnsupported: for p in repo[None].parents(): if p.rev() in s: return False return True return subset & s.filter(filter, condrepr=b'<roots>') MAXINT = (1 << 31) - 1 MININT = -MAXINT - 1 def pick_random(c, gen=random): # exists as its own function to make it possible to overwrite the seed return gen.randint(MININT, MAXINT) _sortkeyfuncs = { b'rev': scmutil.intrev, b'branch': lambda c: c.branch(), b'desc': lambda c: c.description(), b'user': lambda c: c.user(), b'author': lambda c: c.user(), b'date': lambda c: c.date()[0], b'node': scmutil.binnode, b'random': pick_random, } def _getsortargs(x): """Parse sort options into (set, [(key, reverse)], opts)""" args = getargsdict( x, b'sort', b'set keys topo.firstbranch random.seed', ) if b'set' not in args: # i18n: "sort" is a keyword raise error.ParseError(_(b'sort requires one or two arguments')) keys = b"rev" if b'keys' in args: # i18n: "sort" is a keyword keys = getstring(args[b'keys'], _(b"sort spec must be a string")) keyflags = [] for k in keys.split(): fk = k reverse = k.startswith(b'-') if reverse: k = k[1:] if k not in _sortkeyfuncs and k != b'topo': raise error.ParseError( _(b"unknown sort key %r") % pycompat.bytestr(fk) ) keyflags.append((k, reverse)) if len(keyflags) > 1 and any(k == b'topo' for k, reverse in keyflags): # i18n: "topo" is a keyword raise error.ParseError( _(b'topo sort order cannot be combined with other sort keys') ) opts = {} if b'topo.firstbranch' in args: if any(k == b'topo' for k, reverse in keyflags): opts[b'topo.firstbranch'] = args[b'topo.firstbranch'] else: # i18n: "topo" and "topo.firstbranch" are keywords raise error.ParseError( _( b'topo.firstbranch can only be used ' b'when using the topo sort key' ) ) if b'random.seed' in args: if any(k == b'random' for k, reverse in keyflags): s = args[b'random.seed'] seed = getstring(s, _(b"random.seed must be a string")) opts[b'random.seed'] = seed else: # i18n: "random" and "random.seed" are keywords raise error.ParseError( _( b'random.seed can only be used ' b'when using the random sort key' ) ) return args[b'set'], keyflags, opts @predicate( b'sort(set[, [-]key... [, ...]])', safe=True, takeorder=True, weight=10 ) def sort(repo, subset, x, order): """Sort set by keys. The default sort order is ascending, specify a key as ``-key`` to sort in descending order. The keys can be: - ``rev`` for the revision number, - ``branch`` for the branch name, - ``desc`` for the commit message (description), - ``user`` for user name (``author`` can be used as an alias), - ``date`` for the commit date - ``topo`` for a reverse topographical sort - ``node`` the nodeid of the revision - ``random`` randomly shuffle revisions The ``topo`` sort order cannot be combined with other sort keys. This sort takes one optional argument, ``topo.firstbranch``, which takes a revset that specifies what topographical branches to prioritize in the sort. The ``random`` sort takes one optional ``random.seed`` argument to control the pseudo-randomness of the result. """ s, keyflags, opts = _getsortargs(x) revs = getset(repo, subset, s, order) if not keyflags or order != defineorder: return revs if len(keyflags) == 1 and keyflags[0][0] == b"rev": revs.sort(reverse=keyflags[0][1]) return revs elif keyflags[0][0] == b"topo": firstbranch = () parentrevs = repo.changelog.parentrevs parentsfunc = parentrevs if wdirrev in revs: def parentsfunc(r): try: return parentrevs(r) except error.WdirUnsupported: return [p.rev() for p in repo[None].parents()] if b'topo.firstbranch' in opts: firstbranch = getset(repo, subset, opts[b'topo.firstbranch']) revs = baseset( dagop.toposort(revs, parentsfunc, firstbranch), istopo=True, ) if keyflags[0][1]: revs.reverse() return revs # sort() is guaranteed to be stable ctxs = [repo[r] for r in revs] for k, reverse in reversed(keyflags): func = _sortkeyfuncs[k] if k == b'random' and b'random.seed' in opts: seed = opts[b'random.seed'] r = random.Random(seed) func = functools.partial(func, gen=r) ctxs.sort(key=func, reverse=reverse) return baseset([c.rev() for c in ctxs]) @predicate(b'subrepo([pattern])') def subrepo(repo, subset, x): """Changesets that add, modify or remove the given subrepo. If no subrepo pattern is named, any subrepo changes are returned. """ # i18n: "subrepo" is a keyword args = getargs(x, 0, 1, _(b'subrepo takes at most one argument')) pat = None if len(args) != 0: pat = getstring(args[0], _(b"subrepo requires a pattern")) m = matchmod.exact([b'.hgsubstate']) def submatches(names): k, p, m = stringutil.stringmatcher(pat) for name in names: if m(name): yield name def matches(x): c = repo[x] s = repo.status(c.p1().node(), c.node(), match=m) if pat is None: return s.added or s.modified or s.removed if s.added: return any(submatches(c.substate.keys())) if s.modified: subs = set(c.p1().substate.keys()) subs.update(c.substate.keys()) for path in submatches(subs): if c.p1().substate.get(path) != c.substate.get(path): return True if s.removed: return any(submatches(c.p1().substate.keys())) return False return subset.filter(matches, condrepr=(b'<subrepo %r>', pat)) def _mapbynodefunc(repo, s, f): """(repo, smartset, [node] -> [node]) -> smartset Helper method to map a smartset to another smartset given a function only talking about nodes. Handles converting between rev numbers and nodes, and filtering. """ cl = repo.unfiltered().changelog torev = cl.index.get_rev tonode = cl.node result = {torev(n) for n in f(tonode(r) for r in s)} result.discard(None) return smartset.baseset(result - repo.changelog.filteredrevs) @predicate(b'successors(set)', safe=True) def successors(repo, subset, x): """All successors for set, including the given set themselves. (EXPERIMENTAL)""" s = getset(repo, fullreposet(repo), x) f = lambda nodes: obsutil.allsuccessors(repo.obsstore, nodes) d = _mapbynodefunc(repo, s, f) return subset & d def _substringmatcher(pattern, casesensitive=True): kind, pattern, matcher = stringutil.stringmatcher( pattern, casesensitive=casesensitive ) if kind == b'literal': if not casesensitive: pattern = encoding.lower(pattern) matcher = lambda s: pattern in encoding.lower(s) else: matcher = lambda s: pattern in s return kind, pattern, matcher @predicate(b'tag([name])', safe=True) def tag(repo, subset, x): """The specified tag by name, or all tagged revisions if no name is given. Pattern matching is supported for `name`. See :hg:`help revisions.patterns`. """ # i18n: "tag" is a keyword args = getargs(x, 0, 1, _(b"tag takes one or no arguments")) cl = repo.changelog if args: pattern = getstring( args[0], # i18n: "tag" is a keyword _(b'the argument to tag must be a string'), ) kind, pattern, matcher = stringutil.stringmatcher(pattern) if kind == b'literal': # avoid resolving all tags tn = repo._tagscache.tags.get(pattern, None) if tn is None: raise error.RepoLookupError( _(b"tag '%s' does not exist") % pattern ) s = {repo[tn].rev()} else: s = {cl.rev(n) for t, n in repo.tagslist() if matcher(t)} else: s = {cl.rev(n) for t, n in repo.tagslist() if t != b'tip'} return subset & s @predicate(b'tagged', safe=True) def tagged(repo, subset, x): return tag(repo, subset, x) @predicate(b'orphan()', safe=True) def orphan(repo, subset, x): """Non-obsolete changesets with obsolete ancestors. (EXPERIMENTAL)""" # i18n: "orphan" is a keyword getargs(x, 0, 0, _(b"orphan takes no arguments")) orphan = obsmod.getrevs(repo, b'orphan') return subset & orphan @predicate(b'unstable()', safe=True) def unstable(repo, subset, x): """Changesets with instabilities. (EXPERIMENTAL)""" # i18n: "unstable" is a keyword getargs(x, 0, 0, b'unstable takes no arguments') _unstable = set() _unstable.update(obsmod.getrevs(repo, b'orphan')) _unstable.update(obsmod.getrevs(repo, b'phasedivergent')) _unstable.update(obsmod.getrevs(repo, b'contentdivergent')) return subset & baseset(_unstable) @predicate(b'user(string)', safe=True, weight=10) def user(repo, subset, x): """User name contains string. The match is case-insensitive. Pattern matching is supported for `string`. See :hg:`help revisions.patterns`. """ return author(repo, subset, x) @predicate(b'wdir()', safe=True, weight=0) def wdir(repo, subset, x): """Working directory. (EXPERIMENTAL)""" # i18n: "wdir" is a keyword getargs(x, 0, 0, _(b"wdir takes no arguments")) if wdirrev in subset or isinstance(subset, fullreposet): return baseset([wdirrev]) return baseset() def _orderedlist(repo, subset, x): s = getstring(x, b"internal error") if not s: return baseset() # remove duplicates here. it's difficult for caller to deduplicate sets # because different symbols can point to the same rev. cl = repo.changelog ls = [] seen = set() for t in s.split(b'\0'): try: # fast path for integer revision r = int(t) if (b'%d' % r) != t or r not in cl: raise ValueError revs = [r] except ValueError: revs = stringset(repo, subset, t, defineorder) for r in revs: if r in seen: continue if ( r in subset or r in _virtualrevs and isinstance(subset, fullreposet) ): ls.append(r) seen.add(r) return baseset(ls) # for internal use @predicate(b'_list', safe=True, takeorder=True) def _list(repo, subset, x, order): if order == followorder: # slow path to take the subset order return subset & _orderedlist(repo, fullreposet(repo), x) else: return _orderedlist(repo, subset, x) def _orderedintlist(repo, subset, x): s = getstring(x, b"internal error") if not s: return baseset() ls = [int(r) for r in s.split(b'\0')] s = subset return baseset([r for r in ls if r in s]) # for internal use @predicate(b'_intlist', safe=True, takeorder=True, weight=0) def _intlist(repo, subset, x, order): if order == followorder: # slow path to take the subset order return subset & _orderedintlist(repo, fullreposet(repo), x) else: return _orderedintlist(repo, subset, x) def _orderedhexlist(repo, subset, x): s = getstring(x, b"internal error") if not s: return baseset() cl = repo.changelog ls = [cl.rev(bin(r)) for r in s.split(b'\0')] s = subset return baseset([r for r in ls if r in s]) # for internal use @predicate(b'_hexlist', safe=True, takeorder=True) def _hexlist(repo, subset, x, order): if order == followorder: # slow path to take the subset order return subset & _orderedhexlist(repo, fullreposet(repo), x) else: return _orderedhexlist(repo, subset, x) methods = { b"range": rangeset, b"rangeall": rangeall, b"rangepre": rangepre, b"rangepost": rangepost, b"dagrange": dagrange, b"string": stringset, b"symbol": stringset, b"and": andset, b"andsmally": andsmallyset, b"or": orset, b"not": notset, b"difference": differenceset, b"relation": relationset, b"relsubscript": relsubscriptset, b"subscript": subscriptset, b"list": listset, b"keyvalue": keyvaluepair, b"func": func, b"ancestor": ancestorspec, b"parent": parentspec, b"parentpost": parentpost, b"smartset": rawsmartset, } relations = { b"g": generationsrel, b"generations": generationsrel, } subscriptrelations = { b"g": generationssubrel, b"generations": generationssubrel, } def lookupfn(repo): def fn(symbol): try: return scmutil.isrevsymbol(repo, symbol) except error.AmbiguousPrefixLookupError: raise error.InputError( b'ambiguous revision identifier: %s' % symbol ) return fn def match(ui, spec, lookup=None): """Create a matcher for a single revision spec""" return matchany(ui, [spec], lookup=lookup) def matchany(ui, specs, lookup=None, localalias=None): """Create a matcher that will include any revisions matching one of the given specs If lookup function is not None, the parser will first attempt to handle old-style ranges, which may contain operator characters. If localalias is not None, it is a dict {name: definitionstring}. It takes precedence over [revsetalias] config section. """ if not specs: def mfunc(repo, subset=None): return baseset() return mfunc if not all(specs): raise error.ParseError(_(b"empty query")) if len(specs) == 1: tree = revsetlang.parse(specs[0], lookup) else: tree = ( b'or', (b'list',) + tuple(revsetlang.parse(s, lookup) for s in specs), ) aliases = [] warn = None if ui: aliases.extend(ui.configitems(b'revsetalias')) warn = ui.warn if localalias: aliases.extend(localalias.items()) if aliases: tree = revsetlang.expandaliases(tree, aliases, warn=warn) tree = revsetlang.foldconcat(tree) tree = revsetlang.analyze(tree) tree = revsetlang.optimize(tree) return makematcher(tree) def makematcher(tree): """Create a matcher from an evaluatable tree""" def mfunc(repo, subset=None, order=None): if order is None: if subset is None: order = defineorder # 'x' else: order = followorder # 'subset & x' if subset is None: subset = fullreposet(repo) return getset(repo, subset, tree, order) return mfunc def loadpredicate(ui, extname, registrarobj): """Load revset predicates from specified registrarobj""" for name, func in registrarobj._table.items(): symbols[name] = func if func._safe: safesymbols.add(name) # load built-in predicates explicitly to setup safesymbols loadpredicate(None, None, predicate) # tell hggettext to extract docstrings from these functions: i18nfunctions = symbols.values()