Mercurial > hg-stable
view mercurial/context.py @ 34963:5e27afeddaee stable
subrepo: add config option to reject any subrepo operations (SEC)
This is an alternative workaround for the issue5730.
Perhaps this is the simplest way of disabling subrepo operations. It does
nothing clever, but just aborts if Mercurial starts accessing to a subrepo.
I think Greg's patch is more useful since it allows us to at least check
out the parent repository. However, that would be confusing if the default
is flipped to checkout=False and subrepos are silently ignored.
I don't like the config name 'allowed', but I couldn't get any better name.
author | Yuya Nishihara <yuya@tcha.org> |
---|---|
date | Sun, 05 Nov 2017 21:22:07 +0900 |
parents | f7e4d6c20095 |
children | a9454beb9dd8 |
line wrap: on
line source
# context.py - changeset and file context objects for mercurial # # Copyright 2006, 2007 Matt Mackall <mpm@selenic.com> # # This software may be used and distributed according to the terms of the # GNU General Public License version 2 or any later version. from __future__ import absolute_import import errno import filecmp import os import re import stat from .i18n import _ from .node import ( addednodeid, bin, hex, modifiednodeid, nullid, nullrev, short, wdirid, wdirnodes, wdirrev, ) from .thirdparty import ( attr, ) from . import ( encoding, error, fileset, match as matchmod, mdiff, obsolete as obsmod, patch, pathutil, phases, pycompat, repoview, revlog, scmutil, sparse, subrepo, util, ) propertycache = util.propertycache nonascii = re.compile(r'[^\x21-\x7f]').search class basectx(object): """A basectx object represents the common logic for its children: changectx: read-only context that is already present in the repo, workingctx: a context that represents the working directory and can be committed, memctx: a context that represents changes in-memory and can also be committed.""" def __new__(cls, repo, changeid='', *args, **kwargs): if isinstance(changeid, basectx): return changeid o = super(basectx, cls).__new__(cls) o._repo = repo o._rev = nullrev o._node = nullid return o def __bytes__(self): return short(self.node()) __str__ = encoding.strmethod(__bytes__) def __int__(self): return self.rev() def __repr__(self): return r"<%s %s>" % (type(self).__name__, str(self)) def __eq__(self, other): try: return type(self) == type(other) and self._rev == other._rev except AttributeError: return False def __ne__(self, other): return not (self == other) def __contains__(self, key): return key in self._manifest def __getitem__(self, key): return self.filectx(key) def __iter__(self): return iter(self._manifest) def _buildstatusmanifest(self, status): """Builds a manifest that includes the given status results, if this is a working copy context. For non-working copy contexts, it just returns the normal manifest.""" return self.manifest() def _matchstatus(self, other, match): """This internal method provides a way for child objects to override the match operator. """ return match def _buildstatus(self, other, s, match, listignored, listclean, listunknown): """build a status with respect to another context""" # Load earliest manifest first for caching reasons. More specifically, # if you have revisions 1000 and 1001, 1001 is probably stored as a # delta against 1000. Thus, if you read 1000 first, we'll reconstruct # 1000 and cache it so that when you read 1001, we just need to apply a # delta to what's in the cache. So that's one full reconstruction + one # delta application. mf2 = None if self.rev() is not None and self.rev() < other.rev(): mf2 = self._buildstatusmanifest(s) mf1 = other._buildstatusmanifest(s) if mf2 is None: mf2 = self._buildstatusmanifest(s) modified, added = [], [] removed = [] clean = [] deleted, unknown, ignored = s.deleted, s.unknown, s.ignored deletedset = set(deleted) d = mf1.diff(mf2, match=match, clean=listclean) for fn, value in d.iteritems(): if fn in deletedset: continue if value is None: clean.append(fn) continue (node1, flag1), (node2, flag2) = value if node1 is None: added.append(fn) elif node2 is None: removed.append(fn) elif flag1 != flag2: modified.append(fn) elif node2 not in wdirnodes: # When comparing files between two commits, we save time by # not comparing the file contents when the nodeids differ. # Note that this means we incorrectly report a reverted change # to a file as a modification. modified.append(fn) elif self[fn].cmp(other[fn]): modified.append(fn) else: clean.append(fn) if removed: # need to filter files if they are already reported as removed unknown = [fn for fn in unknown if fn not in mf1 and (not match or match(fn))] ignored = [fn for fn in ignored if fn not in mf1 and (not match or match(fn))] # if they're deleted, don't report them as removed removed = [fn for fn in removed if fn not in deletedset] return scmutil.status(modified, added, removed, deleted, unknown, ignored, clean) @propertycache def substate(self): return subrepo.state(self, self._repo.ui) def subrev(self, subpath): return self.substate[subpath][1] def rev(self): return self._rev def node(self): return self._node def hex(self): return hex(self.node()) def manifest(self): return self._manifest def manifestctx(self): return self._manifestctx def repo(self): return self._repo def phasestr(self): return phases.phasenames[self.phase()] def mutable(self): return self.phase() > phases.public def getfileset(self, expr): return fileset.getfileset(self, expr) def obsolete(self): """True if the changeset is obsolete""" return self.rev() in obsmod.getrevs(self._repo, 'obsolete') def extinct(self): """True if the changeset is extinct""" return self.rev() in obsmod.getrevs(self._repo, 'extinct') def unstable(self): msg = ("'context.unstable' is deprecated, " "use 'context.orphan'") self._repo.ui.deprecwarn(msg, '4.4') return self.orphan() def orphan(self): """True if the changeset is not obsolete but it's ancestor are""" return self.rev() in obsmod.getrevs(self._repo, 'orphan') def bumped(self): msg = ("'context.bumped' is deprecated, " "use 'context.phasedivergent'") self._repo.ui.deprecwarn(msg, '4.4') return self.phasedivergent() def phasedivergent(self): """True if the changeset try to be a successor of a public changeset Only non-public and non-obsolete changesets may be bumped. """ return self.rev() in obsmod.getrevs(self._repo, 'phasedivergent') def divergent(self): msg = ("'context.divergent' is deprecated, " "use 'context.contentdivergent'") self._repo.ui.deprecwarn(msg, '4.4') return self.contentdivergent() def contentdivergent(self): """Is a successors of a changeset with multiple possible successors set Only non-public and non-obsolete changesets may be divergent. """ return self.rev() in obsmod.getrevs(self._repo, 'contentdivergent') def troubled(self): msg = ("'context.troubled' is deprecated, " "use 'context.isunstable'") self._repo.ui.deprecwarn(msg, '4.4') return self.isunstable() def isunstable(self): """True if the changeset is either unstable, bumped or divergent""" return self.orphan() or self.phasedivergent() or self.contentdivergent() def troubles(self): """Keep the old version around in order to avoid breaking extensions about different return values. """ msg = ("'context.troubles' is deprecated, " "use 'context.instabilities'") self._repo.ui.deprecwarn(msg, '4.4') troubles = [] if self.orphan(): troubles.append('orphan') if self.phasedivergent(): troubles.append('bumped') if self.contentdivergent(): troubles.append('divergent') return troubles def instabilities(self): """return the list of instabilities affecting this changeset. Instabilities are returned as strings. possible values are: - orphan, - phase-divergent, - content-divergent. """ instabilities = [] if self.orphan(): instabilities.append('orphan') if self.phasedivergent(): instabilities.append('phase-divergent') if self.contentdivergent(): instabilities.append('content-divergent') return instabilities def parents(self): """return contexts for each parent changeset""" return self._parents def p1(self): return self._parents[0] def p2(self): parents = self._parents if len(parents) == 2: return parents[1] return changectx(self._repo, nullrev) def _fileinfo(self, path): if r'_manifest' in self.__dict__: try: return self._manifest[path], self._manifest.flags(path) except KeyError: raise error.ManifestLookupError(self._node, path, _('not found in manifest')) if r'_manifestdelta' in self.__dict__ or path in self.files(): if path in self._manifestdelta: return (self._manifestdelta[path], self._manifestdelta.flags(path)) mfl = self._repo.manifestlog try: node, flag = mfl[self._changeset.manifest].find(path) except KeyError: raise error.ManifestLookupError(self._node, path, _('not found in manifest')) return node, flag def filenode(self, path): return self._fileinfo(path)[0] def flags(self, path): try: return self._fileinfo(path)[1] except error.LookupError: return '' def sub(self, path, allowcreate=True): '''return a subrepo for the stored revision of path, never wdir()''' return subrepo.subrepo(self, path, allowcreate=allowcreate) def nullsub(self, path, pctx): return subrepo.nullsubrepo(self, path, pctx) def workingsub(self, path): '''return a subrepo for the stored revision, or wdir if this is a wdir context. ''' return subrepo.subrepo(self, path, allowwdir=True) def match(self, pats=None, include=None, exclude=None, default='glob', listsubrepos=False, badfn=None): r = self._repo return matchmod.match(r.root, r.getcwd(), pats, include, exclude, default, auditor=r.nofsauditor, ctx=self, listsubrepos=listsubrepos, badfn=badfn) def diff(self, ctx2=None, match=None, **opts): """Returns a diff generator for the given contexts and matcher""" if ctx2 is None: ctx2 = self.p1() if ctx2 is not None: ctx2 = self._repo[ctx2] diffopts = patch.diffopts(self._repo.ui, opts) return patch.diff(self._repo, ctx2, self, match=match, opts=diffopts) def dirs(self): return self._manifest.dirs() def hasdir(self, dir): return self._manifest.hasdir(dir) def status(self, other=None, match=None, listignored=False, listclean=False, listunknown=False, listsubrepos=False): """return status of files between two nodes or node and working directory. If other is None, compare this node with working directory. returns (modified, added, removed, deleted, unknown, ignored, clean) """ ctx1 = self ctx2 = self._repo[other] # This next code block is, admittedly, fragile logic that tests for # reversing the contexts and wouldn't need to exist if it weren't for # the fast (and common) code path of comparing the working directory # with its first parent. # # What we're aiming for here is the ability to call: # # workingctx.status(parentctx) # # If we always built the manifest for each context and compared those, # then we'd be done. But the special case of the above call means we # just copy the manifest of the parent. reversed = False if (not isinstance(ctx1, changectx) and isinstance(ctx2, changectx)): reversed = True ctx1, ctx2 = ctx2, ctx1 match = match or matchmod.always(self._repo.root, self._repo.getcwd()) match = ctx2._matchstatus(ctx1, match) r = scmutil.status([], [], [], [], [], [], []) r = ctx2._buildstatus(ctx1, r, match, listignored, listclean, listunknown) if reversed: # Reverse added and removed. Clear deleted, unknown and ignored as # these make no sense to reverse. r = scmutil.status(r.modified, r.removed, r.added, [], [], [], r.clean) if listsubrepos: for subpath, sub in scmutil.itersubrepos(ctx1, ctx2): try: rev2 = ctx2.subrev(subpath) except KeyError: # A subrepo that existed in node1 was deleted between # node1 and node2 (inclusive). Thus, ctx2's substate # won't contain that subpath. The best we can do ignore it. rev2 = None submatch = matchmod.subdirmatcher(subpath, match) s = sub.status(rev2, match=submatch, ignored=listignored, clean=listclean, unknown=listunknown, listsubrepos=True) for rfiles, sfiles in zip(r, s): rfiles.extend("%s/%s" % (subpath, f) for f in sfiles) for l in r: l.sort() return r def _filterederror(repo, changeid): """build an exception to be raised about a filtered changeid This is extracted in a function to help extensions (eg: evolve) to experiment with various message variants.""" if repo.filtername.startswith('visible'): msg = _("hidden revision '%s'") % changeid hint = _('use --hidden to access hidden revisions') return error.FilteredRepoLookupError(msg, hint=hint) msg = _("filtered revision '%s' (not in '%s' subset)") msg %= (changeid, repo.filtername) return error.FilteredRepoLookupError(msg) class changectx(basectx): """A changecontext object makes access to data related to a particular changeset convenient. It represents a read-only context already present in the repo.""" def __init__(self, repo, changeid=''): """changeid is a revision number, node, or tag""" # since basectx.__new__ already took care of copying the object, we # don't need to do anything in __init__, so we just exit here if isinstance(changeid, basectx): return if changeid == '': changeid = '.' self._repo = repo try: if isinstance(changeid, int): self._node = repo.changelog.node(changeid) self._rev = changeid return if not pycompat.ispy3 and isinstance(changeid, long): changeid = str(changeid) if changeid == 'null': self._node = nullid self._rev = nullrev return if changeid == 'tip': self._node = repo.changelog.tip() self._rev = repo.changelog.rev(self._node) return if (changeid == '.' or repo.local() and changeid == repo.dirstate.p1()): # this is a hack to delay/avoid loading obsmarkers # when we know that '.' won't be hidden self._node = repo.dirstate.p1() self._rev = repo.unfiltered().changelog.rev(self._node) return if len(changeid) == 20: try: self._node = changeid self._rev = repo.changelog.rev(changeid) return except error.FilteredRepoLookupError: raise except LookupError: pass try: r = int(changeid) if '%d' % r != changeid: raise ValueError l = len(repo.changelog) if r < 0: r += l if r < 0 or r >= l and r != wdirrev: raise ValueError self._rev = r self._node = repo.changelog.node(r) return except error.FilteredIndexError: raise except (ValueError, OverflowError, IndexError): pass if len(changeid) == 40: try: self._node = bin(changeid) self._rev = repo.changelog.rev(self._node) return except error.FilteredLookupError: raise except (TypeError, LookupError): pass # lookup bookmarks through the name interface try: self._node = repo.names.singlenode(repo, changeid) self._rev = repo.changelog.rev(self._node) return except KeyError: pass except error.FilteredRepoLookupError: raise except error.RepoLookupError: pass self._node = repo.unfiltered().changelog._partialmatch(changeid) if self._node is not None: self._rev = repo.changelog.rev(self._node) return # lookup failed # check if it might have come from damaged dirstate # # XXX we could avoid the unfiltered if we had a recognizable # exception for filtered changeset access if (repo.local() and changeid in repo.unfiltered().dirstate.parents()): msg = _("working directory has unknown parent '%s'!") raise error.Abort(msg % short(changeid)) try: if len(changeid) == 20 and nonascii(changeid): changeid = hex(changeid) except TypeError: pass except (error.FilteredIndexError, error.FilteredLookupError, error.FilteredRepoLookupError): raise _filterederror(repo, changeid) except IndexError: pass raise error.RepoLookupError( _("unknown revision '%s'") % changeid) def __hash__(self): try: return hash(self._rev) except AttributeError: return id(self) def __nonzero__(self): return self._rev != nullrev __bool__ = __nonzero__ @propertycache def _changeset(self): return self._repo.changelog.changelogrevision(self.rev()) @propertycache def _manifest(self): return self._manifestctx.read() @property def _manifestctx(self): return self._repo.manifestlog[self._changeset.manifest] @propertycache def _manifestdelta(self): return self._manifestctx.readdelta() @propertycache def _parents(self): repo = self._repo p1, p2 = repo.changelog.parentrevs(self._rev) if p2 == nullrev: return [changectx(repo, p1)] return [changectx(repo, p1), changectx(repo, p2)] def changeset(self): c = self._changeset return ( c.manifest, c.user, c.date, c.files, c.description, c.extra, ) def manifestnode(self): return self._changeset.manifest def user(self): return self._changeset.user def date(self): return self._changeset.date def files(self): return self._changeset.files def description(self): return self._changeset.description def branch(self): return encoding.tolocal(self._changeset.extra.get("branch")) def closesbranch(self): return 'close' in self._changeset.extra def extra(self): return self._changeset.extra def tags(self): return self._repo.nodetags(self._node) def bookmarks(self): return self._repo.nodebookmarks(self._node) def phase(self): return self._repo._phasecache.phase(self._repo, self._rev) def hidden(self): return self._rev in repoview.filterrevs(self._repo, 'visible') def isinmemory(self): return False def children(self): """return contexts for each child changeset""" c = self._repo.changelog.children(self._node) return [changectx(self._repo, x) for x in c] def ancestors(self): for a in self._repo.changelog.ancestors([self._rev]): yield changectx(self._repo, a) def descendants(self): for d in self._repo.changelog.descendants([self._rev]): yield changectx(self._repo, d) def filectx(self, path, fileid=None, filelog=None): """get a file context from this changeset""" if fileid is None: fileid = self.filenode(path) return filectx(self._repo, path, fileid=fileid, changectx=self, filelog=filelog) def ancestor(self, c2, warn=False): """return the "best" ancestor context of self and c2 If there are multiple candidates, it will show a message and check merge.preferancestor configuration before falling back to the revlog ancestor.""" # deal with workingctxs n2 = c2._node if n2 is None: n2 = c2._parents[0]._node cahs = self._repo.changelog.commonancestorsheads(self._node, n2) if not cahs: anc = nullid elif len(cahs) == 1: anc = cahs[0] else: # experimental config: merge.preferancestor for r in self._repo.ui.configlist('merge', 'preferancestor'): try: ctx = changectx(self._repo, r) except error.RepoLookupError: continue anc = ctx.node() if anc in cahs: break else: anc = self._repo.changelog.ancestor(self._node, n2) if warn: self._repo.ui.status( (_("note: using %s as ancestor of %s and %s\n") % (short(anc), short(self._node), short(n2))) + ''.join(_(" alternatively, use --config " "merge.preferancestor=%s\n") % short(n) for n in sorted(cahs) if n != anc)) return changectx(self._repo, anc) def descendant(self, other): """True if other is descendant of this changeset""" return self._repo.changelog.descendant(self._rev, other._rev) def walk(self, match): '''Generates matching file names.''' # Wrap match.bad method to have message with nodeid def bad(fn, msg): # The manifest doesn't know about subrepos, so don't complain about # paths into valid subrepos. if any(fn == s or fn.startswith(s + '/') for s in self.substate): return match.bad(fn, _('no such file in rev %s') % self) m = matchmod.badmatch(match, bad) return self._manifest.walk(m) def matches(self, match): return self.walk(match) class basefilectx(object): """A filecontext object represents the common logic for its children: filectx: read-only access to a filerevision that is already present in the repo, workingfilectx: a filecontext that represents files from the working directory, memfilectx: a filecontext that represents files in-memory, overlayfilectx: duplicate another filecontext with some fields overridden. """ @propertycache def _filelog(self): return self._repo.file(self._path) @propertycache def _changeid(self): if r'_changeid' in self.__dict__: return self._changeid elif r'_changectx' in self.__dict__: return self._changectx.rev() elif r'_descendantrev' in self.__dict__: # this file context was created from a revision with a known # descendant, we can (lazily) correct for linkrev aliases return self._adjustlinkrev(self._descendantrev) else: return self._filelog.linkrev(self._filerev) @propertycache def _filenode(self): if r'_fileid' in self.__dict__: return self._filelog.lookup(self._fileid) else: return self._changectx.filenode(self._path) @propertycache def _filerev(self): return self._filelog.rev(self._filenode) @propertycache def _repopath(self): return self._path def __nonzero__(self): try: self._filenode return True except error.LookupError: # file is missing return False __bool__ = __nonzero__ def __bytes__(self): try: return "%s@%s" % (self.path(), self._changectx) except error.LookupError: return "%s@???" % self.path() __str__ = encoding.strmethod(__bytes__) def __repr__(self): return "<%s %s>" % (type(self).__name__, str(self)) def __hash__(self): try: return hash((self._path, self._filenode)) except AttributeError: return id(self) def __eq__(self, other): try: return (type(self) == type(other) and self._path == other._path and self._filenode == other._filenode) except AttributeError: return False def __ne__(self, other): return not (self == other) def filerev(self): return self._filerev def filenode(self): return self._filenode @propertycache def _flags(self): return self._changectx.flags(self._path) def flags(self): return self._flags def filelog(self): return self._filelog def rev(self): return self._changeid def linkrev(self): return self._filelog.linkrev(self._filerev) def node(self): return self._changectx.node() def hex(self): return self._changectx.hex() def user(self): return self._changectx.user() def date(self): return self._changectx.date() def files(self): return self._changectx.files() def description(self): return self._changectx.description() def branch(self): return self._changectx.branch() def extra(self): return self._changectx.extra() def phase(self): return self._changectx.phase() def phasestr(self): return self._changectx.phasestr() def manifest(self): return self._changectx.manifest() def changectx(self): return self._changectx def renamed(self): return self._copied def repo(self): return self._repo def size(self): return len(self.data()) def path(self): return self._path def isbinary(self): try: return util.binary(self.data()) except IOError: return False def isexec(self): return 'x' in self.flags() def islink(self): return 'l' in self.flags() def isabsent(self): """whether this filectx represents a file not in self._changectx This is mainly for merge code to detect change/delete conflicts. This is expected to be True for all subclasses of basectx.""" return False _customcmp = False def cmp(self, fctx): """compare with other file context returns True if different than fctx. """ if fctx._customcmp: return fctx.cmp(self) if (fctx._filenode is None and (self._repo._encodefilterpats # if file data starts with '\1\n', empty metadata block is # prepended, which adds 4 bytes to filelog.size(). or self.size() - 4 == fctx.size()) or self.size() == fctx.size()): return self._filelog.cmp(self._filenode, fctx.data()) return True def _adjustlinkrev(self, srcrev, inclusive=False): """return the first ancestor of <srcrev> introducing <fnode> If the linkrev of the file revision does not point to an ancestor of srcrev, we'll walk down the ancestors until we find one introducing this file revision. :srcrev: the changeset revision we search ancestors from :inclusive: if true, the src revision will also be checked """ repo = self._repo cl = repo.unfiltered().changelog mfl = repo.manifestlog # fetch the linkrev lkr = self.linkrev() # hack to reuse ancestor computation when searching for renames memberanc = getattr(self, '_ancestrycontext', None) iteranc = None if srcrev is None: # wctx case, used by workingfilectx during mergecopy revs = [p.rev() for p in self._repo[None].parents()] inclusive = True # we skipped the real (revless) source else: revs = [srcrev] if memberanc is None: memberanc = iteranc = cl.ancestors(revs, lkr, inclusive=inclusive) # check if this linkrev is an ancestor of srcrev if lkr not in memberanc: if iteranc is None: iteranc = cl.ancestors(revs, lkr, inclusive=inclusive) fnode = self._filenode path = self._path for a in iteranc: ac = cl.read(a) # get changeset data (we avoid object creation) if path in ac[3]: # checking the 'files' field. # The file has been touched, check if the content is # similar to the one we search for. if fnode == mfl[ac[0]].readfast().get(path): return a # In theory, we should never get out of that loop without a result. # But if manifest uses a buggy file revision (not children of the # one it replaces) we could. Such a buggy situation will likely # result is crash somewhere else at to some point. return lkr def introrev(self): """return the rev of the changeset which introduced this file revision This method is different from linkrev because it take into account the changeset the filectx was created from. It ensures the returned revision is one of its ancestors. This prevents bugs from 'linkrev-shadowing' when a file revision is used by multiple changesets. """ lkr = self.linkrev() attrs = vars(self) noctx = not ('_changeid' in attrs or '_changectx' in attrs) if noctx or self.rev() == lkr: return self.linkrev() return self._adjustlinkrev(self.rev(), inclusive=True) def _parentfilectx(self, path, fileid, filelog): """create parent filectx keeping ancestry info for _adjustlinkrev()""" fctx = filectx(self._repo, path, fileid=fileid, filelog=filelog) if '_changeid' in vars(self) or '_changectx' in vars(self): # If self is associated with a changeset (probably explicitly # fed), ensure the created filectx is associated with a # changeset that is an ancestor of self.changectx. # This lets us later use _adjustlinkrev to get a correct link. fctx._descendantrev = self.rev() fctx._ancestrycontext = getattr(self, '_ancestrycontext', None) elif '_descendantrev' in vars(self): # Otherwise propagate _descendantrev if we have one associated. fctx._descendantrev = self._descendantrev fctx._ancestrycontext = getattr(self, '_ancestrycontext', None) return fctx def parents(self): _path = self._path fl = self._filelog parents = self._filelog.parents(self._filenode) pl = [(_path, node, fl) for node in parents if node != nullid] r = fl.renamed(self._filenode) if r: # - In the simple rename case, both parent are nullid, pl is empty. # - In case of merge, only one of the parent is null id and should # be replaced with the rename information. This parent is -always- # the first one. # # As null id have always been filtered out in the previous list # comprehension, inserting to 0 will always result in "replacing # first nullid parent with rename information. pl.insert(0, (r[0], r[1], self._repo.file(r[0]))) return [self._parentfilectx(path, fnode, l) for path, fnode, l in pl] def p1(self): return self.parents()[0] def p2(self): p = self.parents() if len(p) == 2: return p[1] return filectx(self._repo, self._path, fileid=-1, filelog=self._filelog) def annotate(self, follow=False, linenumber=False, skiprevs=None, diffopts=None): '''returns a list of tuples of ((ctx, number), line) for each line in the file, where ctx is the filectx of the node where that line was last changed; if linenumber parameter is true, number is the line number at the first appearance in the managed file, otherwise, number has a fixed value of False. ''' def lines(text): if text.endswith("\n"): return text.count("\n") return text.count("\n") + int(bool(text)) if linenumber: def decorate(text, rev): return ([annotateline(fctx=rev, lineno=i) for i in xrange(1, lines(text) + 1)], text) else: def decorate(text, rev): return ([annotateline(fctx=rev)] * lines(text), text) getlog = util.lrucachefunc(lambda x: self._repo.file(x)) def parents(f): # Cut _descendantrev here to mitigate the penalty of lazy linkrev # adjustment. Otherwise, p._adjustlinkrev() would walk changelog # from the topmost introrev (= srcrev) down to p.linkrev() if it # isn't an ancestor of the srcrev. f._changeid pl = f.parents() # Don't return renamed parents if we aren't following. if not follow: pl = [p for p in pl if p.path() == f.path()] # renamed filectx won't have a filelog yet, so set it # from the cache to save time for p in pl: if not '_filelog' in p.__dict__: p._filelog = getlog(p.path()) return pl # use linkrev to find the first changeset where self appeared base = self introrev = self.introrev() if self.rev() != introrev: base = self.filectx(self.filenode(), changeid=introrev) if getattr(base, '_ancestrycontext', None) is None: cl = self._repo.changelog if introrev is None: # wctx is not inclusive, but works because _ancestrycontext # is used to test filelog revisions ac = cl.ancestors([p.rev() for p in base.parents()], inclusive=True) else: ac = cl.ancestors([introrev], inclusive=True) base._ancestrycontext = ac # This algorithm would prefer to be recursive, but Python is a # bit recursion-hostile. Instead we do an iterative # depth-first search. # 1st DFS pre-calculates pcache and needed visit = [base] pcache = {} needed = {base: 1} while visit: f = visit.pop() if f in pcache: continue pl = parents(f) pcache[f] = pl for p in pl: needed[p] = needed.get(p, 0) + 1 if p not in pcache: visit.append(p) # 2nd DFS does the actual annotate visit[:] = [base] hist = {} while visit: f = visit[-1] if f in hist: visit.pop() continue ready = True pl = pcache[f] for p in pl: if p not in hist: ready = False visit.append(p) if ready: visit.pop() curr = decorate(f.data(), f) skipchild = False if skiprevs is not None: skipchild = f._changeid in skiprevs curr = _annotatepair([hist[p] for p in pl], f, curr, skipchild, diffopts) for p in pl: if needed[p] == 1: del hist[p] del needed[p] else: needed[p] -= 1 hist[f] = curr del pcache[f] return zip(hist[base][0], hist[base][1].splitlines(True)) def ancestors(self, followfirst=False): visit = {} c = self if followfirst: cut = 1 else: cut = None while True: for parent in c.parents()[:cut]: visit[(parent.linkrev(), parent.filenode())] = parent if not visit: break c = visit.pop(max(visit)) yield c def decodeddata(self): """Returns `data()` after running repository decoding filters. This is often equivalent to how the data would be expressed on disk. """ return self._repo.wwritedata(self.path(), self.data()) @attr.s(slots=True, frozen=True) class annotateline(object): fctx = attr.ib() lineno = attr.ib(default=False) # Whether this annotation was the result of a skip-annotate. skip = attr.ib(default=False) def _annotatepair(parents, childfctx, child, skipchild, diffopts): r''' Given parent and child fctxes and annotate data for parents, for all lines in either parent that match the child, annotate the child with the parent's data. Additionally, if `skipchild` is True, replace all other lines with parent annotate data as well such that child is never blamed for any lines. See test-annotate.py for unit tests. ''' pblocks = [(parent, mdiff.allblocks(parent[1], child[1], opts=diffopts)) for parent in parents] if skipchild: # Need to iterate over the blocks twice -- make it a list pblocks = [(p, list(blocks)) for (p, blocks) in pblocks] # Mercurial currently prefers p2 over p1 for annotate. # TODO: change this? for parent, blocks in pblocks: for (a1, a2, b1, b2), t in blocks: # Changed blocks ('!') or blocks made only of blank lines ('~') # belong to the child. if t == '=': child[0][b1:b2] = parent[0][a1:a2] if skipchild: # Now try and match up anything that couldn't be matched, # Reversing pblocks maintains bias towards p2, matching above # behavior. pblocks.reverse() # The heuristics are: # * Work on blocks of changed lines (effectively diff hunks with -U0). # This could potentially be smarter but works well enough. # * For a non-matching section, do a best-effort fit. Match lines in # diff hunks 1:1, dropping lines as necessary. # * Repeat the last line as a last resort. # First, replace as much as possible without repeating the last line. remaining = [(parent, []) for parent, _blocks in pblocks] for idx, (parent, blocks) in enumerate(pblocks): for (a1, a2, b1, b2), _t in blocks: if a2 - a1 >= b2 - b1: for bk in xrange(b1, b2): if child[0][bk].fctx == childfctx: ak = min(a1 + (bk - b1), a2 - 1) child[0][bk] = attr.evolve(parent[0][ak], skip=True) else: remaining[idx][1].append((a1, a2, b1, b2)) # Then, look at anything left, which might involve repeating the last # line. for parent, blocks in remaining: for a1, a2, b1, b2 in blocks: for bk in xrange(b1, b2): if child[0][bk].fctx == childfctx: ak = min(a1 + (bk - b1), a2 - 1) child[0][bk] = attr.evolve(parent[0][ak], skip=True) return child class filectx(basefilectx): """A filecontext object makes access to data related to a particular filerevision convenient.""" def __init__(self, repo, path, changeid=None, fileid=None, filelog=None, changectx=None): """changeid can be a changeset revision, node, or tag. fileid can be a file revision or node.""" self._repo = repo self._path = path assert (changeid is not None or fileid is not None or changectx is not None), \ ("bad args: changeid=%r, fileid=%r, changectx=%r" % (changeid, fileid, changectx)) if filelog is not None: self._filelog = filelog if changeid is not None: self._changeid = changeid if changectx is not None: self._changectx = changectx if fileid is not None: self._fileid = fileid @propertycache def _changectx(self): try: return changectx(self._repo, self._changeid) except error.FilteredRepoLookupError: # Linkrev may point to any revision in the repository. When the # repository is filtered this may lead to `filectx` trying to build # `changectx` for filtered revision. In such case we fallback to # creating `changectx` on the unfiltered version of the reposition. # This fallback should not be an issue because `changectx` from # `filectx` are not used in complex operations that care about # filtering. # # This fallback is a cheap and dirty fix that prevent several # crashes. It does not ensure the behavior is correct. However the # behavior was not correct before filtering either and "incorrect # behavior" is seen as better as "crash" # # Linkrevs have several serious troubles with filtering that are # complicated to solve. Proper handling of the issue here should be # considered when solving linkrev issue are on the table. return changectx(self._repo.unfiltered(), self._changeid) def filectx(self, fileid, changeid=None): '''opens an arbitrary revision of the file without opening a new filelog''' return filectx(self._repo, self._path, fileid=fileid, filelog=self._filelog, changeid=changeid) def rawdata(self): return self._filelog.revision(self._filenode, raw=True) def rawflags(self): """low-level revlog flags""" return self._filelog.flags(self._filerev) def data(self): try: return self._filelog.read(self._filenode) except error.CensoredNodeError: if self._repo.ui.config("censor", "policy") == "ignore": return "" raise error.Abort(_("censored node: %s") % short(self._filenode), hint=_("set censor.policy to ignore errors")) def size(self): return self._filelog.size(self._filerev) @propertycache def _copied(self): """check if file was actually renamed in this changeset revision If rename logged in file revision, we report copy for changeset only if file revisions linkrev points back to the changeset in question or both changeset parents contain different file revisions. """ renamed = self._filelog.renamed(self._filenode) if not renamed: return renamed if self.rev() == self.linkrev(): return renamed name = self.path() fnode = self._filenode for p in self._changectx.parents(): try: if fnode == p.filenode(name): return None except error.LookupError: pass return renamed def children(self): # hard for renames c = self._filelog.children(self._filenode) return [filectx(self._repo, self._path, fileid=x, filelog=self._filelog) for x in c] class committablectx(basectx): """A committablectx object provides common functionality for a context that wants the ability to commit, e.g. workingctx or memctx.""" def __init__(self, repo, text="", user=None, date=None, extra=None, changes=None): self._repo = repo self._rev = None self._node = None self._text = text if date: self._date = util.parsedate(date) if user: self._user = user if changes: self._status = changes self._extra = {} if extra: self._extra = extra.copy() if 'branch' not in self._extra: try: branch = encoding.fromlocal(self._repo.dirstate.branch()) except UnicodeDecodeError: raise error.Abort(_('branch name not in UTF-8!')) self._extra['branch'] = branch if self._extra['branch'] == '': self._extra['branch'] = 'default' def __bytes__(self): return bytes(self._parents[0]) + "+" __str__ = encoding.strmethod(__bytes__) def __nonzero__(self): return True __bool__ = __nonzero__ def _buildflagfunc(self): # Create a fallback function for getting file flags when the # filesystem doesn't support them copiesget = self._repo.dirstate.copies().get parents = self.parents() if len(parents) < 2: # when we have one parent, it's easy: copy from parent man = parents[0].manifest() def func(f): f = copiesget(f, f) return man.flags(f) else: # merges are tricky: we try to reconstruct the unstored # result from the merge (issue1802) p1, p2 = parents pa = p1.ancestor(p2) m1, m2, ma = p1.manifest(), p2.manifest(), pa.manifest() def func(f): f = copiesget(f, f) # may be wrong for merges with copies fl1, fl2, fla = m1.flags(f), m2.flags(f), ma.flags(f) if fl1 == fl2: return fl1 if fl1 == fla: return fl2 if fl2 == fla: return fl1 return '' # punt for conflicts return func @propertycache def _flagfunc(self): return self._repo.dirstate.flagfunc(self._buildflagfunc) @propertycache def _status(self): return self._repo.status() @propertycache def _user(self): return self._repo.ui.username() @propertycache def _date(self): ui = self._repo.ui date = ui.configdate('devel', 'default-date') if date is None: date = util.makedate() return date def subrev(self, subpath): return None def manifestnode(self): return None def user(self): return self._user or self._repo.ui.username() def date(self): return self._date def description(self): return self._text def files(self): return sorted(self._status.modified + self._status.added + self._status.removed) def modified(self): return self._status.modified def added(self): return self._status.added def removed(self): return self._status.removed def deleted(self): return self._status.deleted def branch(self): return encoding.tolocal(self._extra['branch']) def closesbranch(self): return 'close' in self._extra def extra(self): return self._extra def isinmemory(self): return False def tags(self): return [] def bookmarks(self): b = [] for p in self.parents(): b.extend(p.bookmarks()) return b def phase(self): phase = phases.draft # default phase to draft for p in self.parents(): phase = max(phase, p.phase()) return phase def hidden(self): return False def children(self): return [] def flags(self, path): if r'_manifest' in self.__dict__: try: return self._manifest.flags(path) except KeyError: return '' try: return self._flagfunc(path) except OSError: return '' def ancestor(self, c2): """return the "best" ancestor context of self and c2""" return self._parents[0].ancestor(c2) # punt on two parents for now def walk(self, match): '''Generates matching file names.''' return sorted(self._repo.dirstate.walk(match, subrepos=sorted(self.substate), unknown=True, ignored=False)) def matches(self, match): return sorted(self._repo.dirstate.matches(match)) def ancestors(self): for p in self._parents: yield p for a in self._repo.changelog.ancestors( [p.rev() for p in self._parents]): yield changectx(self._repo, a) def markcommitted(self, node): """Perform post-commit cleanup necessary after committing this ctx Specifically, this updates backing stores this working context wraps to reflect the fact that the changes reflected by this workingctx have been committed. For example, it marks modified and added files as normal in the dirstate. """ with self._repo.dirstate.parentchange(): for f in self.modified() + self.added(): self._repo.dirstate.normal(f) for f in self.removed(): self._repo.dirstate.drop(f) self._repo.dirstate.setparents(node) # write changes out explicitly, because nesting wlock at # runtime may prevent 'wlock.release()' in 'repo.commit()' # from immediately doing so for subsequent changing files self._repo.dirstate.write(self._repo.currenttransaction()) def dirty(self, missing=False, merge=True, branch=True): return False class workingctx(committablectx): """A workingctx object makes access to data related to the current working directory convenient. date - any valid date string or (unixtime, offset), or None. user - username string, or None. extra - a dictionary of extra values, or None. changes - a list of file lists as returned by localrepo.status() or None to use the repository status. """ def __init__(self, repo, text="", user=None, date=None, extra=None, changes=None): super(workingctx, self).__init__(repo, text, user, date, extra, changes) def __iter__(self): d = self._repo.dirstate for f in d: if d[f] != 'r': yield f def __contains__(self, key): return self._repo.dirstate[key] not in "?r" def hex(self): return hex(wdirid) @propertycache def _parents(self): p = self._repo.dirstate.parents() if p[1] == nullid: p = p[:-1] return [changectx(self._repo, x) for x in p] def filectx(self, path, filelog=None): """get a file context from the working directory""" return workingfilectx(self._repo, path, workingctx=self, filelog=filelog) def dirty(self, missing=False, merge=True, branch=True): "check whether a working directory is modified" # check subrepos first for s in sorted(self.substate): if self.sub(s).dirty(missing=missing): return True # check current working dir return ((merge and self.p2()) or (branch and self.branch() != self.p1().branch()) or self.modified() or self.added() or self.removed() or (missing and self.deleted())) def add(self, list, prefix=""): with self._repo.wlock(): ui, ds = self._repo.ui, self._repo.dirstate uipath = lambda f: ds.pathto(pathutil.join(prefix, f)) rejected = [] lstat = self._repo.wvfs.lstat for f in list: # ds.pathto() returns an absolute file when this is invoked from # the keyword extension. That gets flagged as non-portable on # Windows, since it contains the drive letter and colon. scmutil.checkportable(ui, os.path.join(prefix, f)) try: st = lstat(f) except OSError: ui.warn(_("%s does not exist!\n") % uipath(f)) rejected.append(f) continue if st.st_size > 10000000: ui.warn(_("%s: up to %d MB of RAM may be required " "to manage this file\n" "(use 'hg revert %s' to cancel the " "pending addition)\n") % (f, 3 * st.st_size // 1000000, uipath(f))) if not (stat.S_ISREG(st.st_mode) or stat.S_ISLNK(st.st_mode)): ui.warn(_("%s not added: only files and symlinks " "supported currently\n") % uipath(f)) rejected.append(f) elif ds[f] in 'amn': ui.warn(_("%s already tracked!\n") % uipath(f)) elif ds[f] == 'r': ds.normallookup(f) else: ds.add(f) return rejected def forget(self, files, prefix=""): with self._repo.wlock(): ds = self._repo.dirstate uipath = lambda f: ds.pathto(pathutil.join(prefix, f)) rejected = [] for f in files: if f not in self._repo.dirstate: self._repo.ui.warn(_("%s not tracked!\n") % uipath(f)) rejected.append(f) elif self._repo.dirstate[f] != 'a': self._repo.dirstate.remove(f) else: self._repo.dirstate.drop(f) return rejected def undelete(self, list): pctxs = self.parents() with self._repo.wlock(): ds = self._repo.dirstate for f in list: if self._repo.dirstate[f] != 'r': self._repo.ui.warn(_("%s not removed!\n") % ds.pathto(f)) else: fctx = f in pctxs[0] and pctxs[0][f] or pctxs[1][f] t = fctx.data() self._repo.wwrite(f, t, fctx.flags()) self._repo.dirstate.normal(f) def copy(self, source, dest): try: st = self._repo.wvfs.lstat(dest) except OSError as err: if err.errno != errno.ENOENT: raise self._repo.ui.warn(_("%s does not exist!\n") % self._repo.dirstate.pathto(dest)) return if not (stat.S_ISREG(st.st_mode) or stat.S_ISLNK(st.st_mode)): self._repo.ui.warn(_("copy failed: %s is not a file or a " "symbolic link\n") % self._repo.dirstate.pathto(dest)) else: with self._repo.wlock(): if self._repo.dirstate[dest] in '?': self._repo.dirstate.add(dest) elif self._repo.dirstate[dest] in 'r': self._repo.dirstate.normallookup(dest) self._repo.dirstate.copy(source, dest) def match(self, pats=None, include=None, exclude=None, default='glob', listsubrepos=False, badfn=None): r = self._repo # Only a case insensitive filesystem needs magic to translate user input # to actual case in the filesystem. icasefs = not util.fscasesensitive(r.root) return matchmod.match(r.root, r.getcwd(), pats, include, exclude, default, auditor=r.auditor, ctx=self, listsubrepos=listsubrepos, badfn=badfn, icasefs=icasefs) def flushall(self): pass # For overlayworkingfilectx compatibility. def _filtersuspectsymlink(self, files): if not files or self._repo.dirstate._checklink: return files # Symlink placeholders may get non-symlink-like contents # via user error or dereferencing by NFS or Samba servers, # so we filter out any placeholders that don't look like a # symlink sane = [] for f in files: if self.flags(f) == 'l': d = self[f].data() if d == '' or len(d) >= 1024 or '\n' in d or util.binary(d): self._repo.ui.debug('ignoring suspect symlink placeholder' ' "%s"\n' % f) continue sane.append(f) return sane def _checklookup(self, files): # check for any possibly clean files if not files: return [], [], [] modified = [] deleted = [] fixup = [] pctx = self._parents[0] # do a full compare of any files that might have changed for f in sorted(files): try: # This will return True for a file that got replaced by a # directory in the interim, but fixing that is pretty hard. if (f not in pctx or self.flags(f) != pctx.flags(f) or pctx[f].cmp(self[f])): modified.append(f) else: fixup.append(f) except (IOError, OSError): # A file become inaccessible in between? Mark it as deleted, # matching dirstate behavior (issue5584). # The dirstate has more complex behavior around whether a # missing file matches a directory, etc, but we don't need to # bother with that: if f has made it to this point, we're sure # it's in the dirstate. deleted.append(f) return modified, deleted, fixup def _poststatusfixup(self, status, fixup): """update dirstate for files that are actually clean""" poststatus = self._repo.postdsstatus() if fixup or poststatus: try: oldid = self._repo.dirstate.identity() # updating the dirstate is optional # so we don't wait on the lock # wlock can invalidate the dirstate, so cache normal _after_ # taking the lock with self._repo.wlock(False): if self._repo.dirstate.identity() == oldid: if fixup: normal = self._repo.dirstate.normal for f in fixup: normal(f) # write changes out explicitly, because nesting # wlock at runtime may prevent 'wlock.release()' # after this block from doing so for subsequent # changing files tr = self._repo.currenttransaction() self._repo.dirstate.write(tr) if poststatus: for ps in poststatus: ps(self, status) else: # in this case, writing changes out breaks # consistency, because .hg/dirstate was # already changed simultaneously after last # caching (see also issue5584 for detail) self._repo.ui.debug('skip updating dirstate: ' 'identity mismatch\n') except error.LockError: pass finally: # Even if the wlock couldn't be grabbed, clear out the list. self._repo.clearpostdsstatus() def _dirstatestatus(self, match, ignored=False, clean=False, unknown=False): '''Gets the status from the dirstate -- internal use only.''' subrepos = [] if '.hgsub' in self: subrepos = sorted(self.substate) cmp, s = self._repo.dirstate.status(match, subrepos, ignored=ignored, clean=clean, unknown=unknown) # check for any possibly clean files fixup = [] if cmp: modified2, deleted2, fixup = self._checklookup(cmp) s.modified.extend(modified2) s.deleted.extend(deleted2) if fixup and clean: s.clean.extend(fixup) self._poststatusfixup(s, fixup) if match.always(): # cache for performance if s.unknown or s.ignored or s.clean: # "_status" is cached with list*=False in the normal route self._status = scmutil.status(s.modified, s.added, s.removed, s.deleted, [], [], []) else: self._status = s return s @propertycache def _manifest(self): """generate a manifest corresponding to the values in self._status This reuse the file nodeid from parent, but we use special node identifiers for added and modified files. This is used by manifests merge to see that files are different and by update logic to avoid deleting newly added files. """ return self._buildstatusmanifest(self._status) def _buildstatusmanifest(self, status): """Builds a manifest that includes the given status results.""" parents = self.parents() man = parents[0].manifest().copy() ff = self._flagfunc for i, l in ((addednodeid, status.added), (modifiednodeid, status.modified)): for f in l: man[f] = i try: man.setflag(f, ff(f)) except OSError: pass for f in status.deleted + status.removed: if f in man: del man[f] return man def _buildstatus(self, other, s, match, listignored, listclean, listunknown): """build a status with respect to another context This includes logic for maintaining the fast path of status when comparing the working directory against its parent, which is to skip building a new manifest if self (working directory) is not comparing against its parent (repo['.']). """ s = self._dirstatestatus(match, listignored, listclean, listunknown) # Filter out symlinks that, in the case of FAT32 and NTFS filesystems, # might have accidentally ended up with the entire contents of the file # they are supposed to be linking to. s.modified[:] = self._filtersuspectsymlink(s.modified) if other != self._repo['.']: s = super(workingctx, self)._buildstatus(other, s, match, listignored, listclean, listunknown) return s def _matchstatus(self, other, match): """override the match method with a filter for directory patterns We use inheritance to customize the match.bad method only in cases of workingctx since it belongs only to the working directory when comparing against the parent changeset. If we aren't comparing against the working directory's parent, then we just use the default match object sent to us. """ if other != self._repo['.']: def bad(f, msg): # 'f' may be a directory pattern from 'match.files()', # so 'f not in ctx1' is not enough if f not in other and not other.hasdir(f): self._repo.ui.warn('%s: %s\n' % (self._repo.dirstate.pathto(f), msg)) match.bad = bad return match def markcommitted(self, node): super(workingctx, self).markcommitted(node) sparse.aftercommit(self._repo, node) class committablefilectx(basefilectx): """A committablefilectx provides common functionality for a file context that wants the ability to commit, e.g. workingfilectx or memfilectx.""" def __init__(self, repo, path, filelog=None, ctx=None): self._repo = repo self._path = path self._changeid = None self._filerev = self._filenode = None if filelog is not None: self._filelog = filelog if ctx: self._changectx = ctx def __nonzero__(self): return True __bool__ = __nonzero__ def linkrev(self): # linked to self._changectx no matter if file is modified or not return self.rev() def parents(self): '''return parent filectxs, following copies if necessary''' def filenode(ctx, path): return ctx._manifest.get(path, nullid) path = self._path fl = self._filelog pcl = self._changectx._parents renamed = self.renamed() if renamed: pl = [renamed + (None,)] else: pl = [(path, filenode(pcl[0], path), fl)] for pc in pcl[1:]: pl.append((path, filenode(pc, path), fl)) return [self._parentfilectx(p, fileid=n, filelog=l) for p, n, l in pl if n != nullid] def children(self): return [] class workingfilectx(committablefilectx): """A workingfilectx object makes access to data related to a particular file in the working directory convenient.""" def __init__(self, repo, path, filelog=None, workingctx=None): super(workingfilectx, self).__init__(repo, path, filelog, workingctx) @propertycache def _changectx(self): return workingctx(self._repo) def data(self): return self._repo.wread(self._path) def renamed(self): rp = self._repo.dirstate.copied(self._path) if not rp: return None return rp, self._changectx._parents[0]._manifest.get(rp, nullid) def size(self): return self._repo.wvfs.lstat(self._path).st_size def date(self): t, tz = self._changectx.date() try: return (self._repo.wvfs.lstat(self._path).st_mtime, tz) except OSError as err: if err.errno != errno.ENOENT: raise return (t, tz) def exists(self): return self._repo.wvfs.exists(self._path) def lexists(self): return self._repo.wvfs.lexists(self._path) def audit(self): return self._repo.wvfs.audit(self._path) def cmp(self, fctx): """compare with other file context returns True if different than fctx. """ # fctx should be a filectx (not a workingfilectx) # invert comparison to reuse the same code path return fctx.cmp(self) def remove(self, ignoremissing=False): """wraps unlink for a repo's working directory""" self._repo.wvfs.unlinkpath(self._path, ignoremissing=ignoremissing) def write(self, data, flags, backgroundclose=False): """wraps repo.wwrite""" self._repo.wwrite(self._path, data, flags, backgroundclose=backgroundclose) def markcopied(self, src): """marks this file a copy of `src`""" if self._repo.dirstate[self._path] in "nma": self._repo.dirstate.copy(src, self._path) def clearunknown(self): """Removes conflicting items in the working directory so that ``write()`` can be called successfully. """ wvfs = self._repo.wvfs f = self._path wvfs.audit(f) if wvfs.isdir(f) and not wvfs.islink(f): wvfs.rmtree(f, forcibly=True) for p in reversed(list(util.finddirs(f))): if wvfs.isfileorlink(p): wvfs.unlink(p) break def setflags(self, l, x): self._repo.wvfs.setflags(self._path, l, x) class overlayworkingctx(workingctx): """Wraps another mutable context with a write-back cache that can be flushed at a later time. self._cache[path] maps to a dict with keys: { 'exists': bool? 'date': date? 'data': str? 'flags': str? } If `exists` is True, `flags` must be non-None and 'date' is non-None. If it is `False`, the file was deleted. """ def __init__(self, repo, wrappedctx): super(overlayworkingctx, self).__init__(repo) self._repo = repo self._wrappedctx = wrappedctx self._clean() def data(self, path): if self.isdirty(path): if self._cache[path]['exists']: if self._cache[path]['data']: return self._cache[path]['data'] else: # Must fallback here, too, because we only set flags. return self._wrappedctx[path].data() else: raise error.ProgrammingError("No such file or directory: %s" % self._path) else: return self._wrappedctx[path].data() def isinmemory(self): return True def filedate(self, path): if self.isdirty(path): return self._cache[path]['date'] else: return self._wrappedctx[path].date() def flags(self, path): if self.isdirty(path): if self._cache[path]['exists']: return self._cache[path]['flags'] else: raise error.ProgrammingError("No such file or directory: %s" % self._path) else: return self._wrappedctx[path].flags() def write(self, path, data, flags=''): if data is None: raise error.ProgrammingError("data must be non-None") self._markdirty(path, exists=True, data=data, date=util.makedate(), flags=flags) def setflags(self, path, l, x): self._markdirty(path, exists=True, date=util.makedate(), flags=(l and 'l' or '') + (x and 'x' or '')) def remove(self, path): self._markdirty(path, exists=False) def exists(self, path): """exists behaves like `lexists`, but needs to follow symlinks and return False if they are broken. """ if self.isdirty(path): # If this path exists and is a symlink, "follow" it by calling # exists on the destination path. if (self._cache[path]['exists'] and 'l' in self._cache[path]['flags']): return self.exists(self._cache[path]['data'].strip()) else: return self._cache[path]['exists'] return self._wrappedctx[path].exists() def lexists(self, path): """lexists returns True if the path exists""" if self.isdirty(path): return self._cache[path]['exists'] return self._wrappedctx[path].lexists() def size(self, path): if self.isdirty(path): if self._cache[path]['exists']: return len(self._cache[path]['data']) else: raise error.ProgrammingError("No such file or directory: %s" % self._path) return self._wrappedctx[path].size() def flushall(self): for path in self._writeorder: entry = self._cache[path] if entry['exists']: self._wrappedctx[path].clearunknown() if entry['data'] is not None: if entry['flags'] is None: raise error.ProgrammingError('data set but not flags') self._wrappedctx[path].write( entry['data'], entry['flags']) else: self._wrappedctx[path].setflags( 'l' in entry['flags'], 'x' in entry['flags']) else: self._wrappedctx[path].remove(path) self._clean() def isdirty(self, path): return path in self._cache def _clean(self): self._cache = {} self._writeorder = [] def _markdirty(self, path, exists, data=None, date=None, flags=''): if path not in self._cache: self._writeorder.append(path) self._cache[path] = { 'exists': exists, 'data': data, 'date': date, 'flags': flags, } def filectx(self, path, filelog=None): return overlayworkingfilectx(self._repo, path, parent=self, filelog=filelog) class overlayworkingfilectx(workingfilectx): """Wrap a ``workingfilectx`` but intercepts all writes into an in-memory cache, which can be flushed through later by calling ``flush()``.""" def __init__(self, repo, path, filelog=None, parent=None): super(overlayworkingfilectx, self).__init__(repo, path, filelog, parent) self._repo = repo self._parent = parent self._path = path def cmp(self, fctx): return self.data() != fctx.data() def ctx(self): return self._parent def data(self): return self._parent.data(self._path) def date(self): return self._parent.filedate(self._path) def exists(self): return self.lexists() def lexists(self): return self._parent.exists(self._path) def renamed(self): # Copies are currently tracked in the dirstate as before. Straight copy # from workingfilectx. rp = self._repo.dirstate.copied(self._path) if not rp: return None return rp, self._changectx._parents[0]._manifest.get(rp, nullid) def size(self): return self._parent.size(self._path) def audit(self): pass def flags(self): return self._parent.flags(self._path) def setflags(self, islink, isexec): return self._parent.setflags(self._path, islink, isexec) def write(self, data, flags, backgroundclose=False): return self._parent.write(self._path, data, flags) def remove(self, ignoremissing=False): return self._parent.remove(self._path) class workingcommitctx(workingctx): """A workingcommitctx object makes access to data related to the revision being committed convenient. This hides changes in the working directory, if they aren't committed in this context. """ def __init__(self, repo, changes, text="", user=None, date=None, extra=None): super(workingctx, self).__init__(repo, text, user, date, extra, changes) def _dirstatestatus(self, match, ignored=False, clean=False, unknown=False): """Return matched files only in ``self._status`` Uncommitted files appear "clean" via this context, even if they aren't actually so in the working directory. """ if clean: clean = [f for f in self._manifest if f not in self._changedset] else: clean = [] return scmutil.status([f for f in self._status.modified if match(f)], [f for f in self._status.added if match(f)], [f for f in self._status.removed if match(f)], [], [], [], clean) @propertycache def _changedset(self): """Return the set of files changed in this context """ changed = set(self._status.modified) changed.update(self._status.added) changed.update(self._status.removed) return changed def makecachingfilectxfn(func): """Create a filectxfn that caches based on the path. We can't use util.cachefunc because it uses all arguments as the cache key and this creates a cycle since the arguments include the repo and memctx. """ cache = {} def getfilectx(repo, memctx, path): if path not in cache: cache[path] = func(repo, memctx, path) return cache[path] return getfilectx def memfilefromctx(ctx): """Given a context return a memfilectx for ctx[path] This is a convenience method for building a memctx based on another context. """ def getfilectx(repo, memctx, path): fctx = ctx[path] # this is weird but apparently we only keep track of one parent # (why not only store that instead of a tuple?) copied = fctx.renamed() if copied: copied = copied[0] return memfilectx(repo, path, fctx.data(), islink=fctx.islink(), isexec=fctx.isexec(), copied=copied, memctx=memctx) return getfilectx def memfilefrompatch(patchstore): """Given a patch (e.g. patchstore object) return a memfilectx This is a convenience method for building a memctx based on a patchstore. """ def getfilectx(repo, memctx, path): data, mode, copied = patchstore.getfile(path) if data is None: return None islink, isexec = mode return memfilectx(repo, path, data, islink=islink, isexec=isexec, copied=copied, memctx=memctx) return getfilectx class memctx(committablectx): """Use memctx to perform in-memory commits via localrepo.commitctx(). Revision information is supplied at initialization time while related files data and is made available through a callback mechanism. 'repo' is the current localrepo, 'parents' is a sequence of two parent revisions identifiers (pass None for every missing parent), 'text' is the commit message and 'files' lists names of files touched by the revision (normalized and relative to repository root). filectxfn(repo, memctx, path) is a callable receiving the repository, the current memctx object and the normalized path of requested file, relative to repository root. It is fired by the commit function for every file in 'files', but calls order is undefined. If the file is available in the revision being committed (updated or added), filectxfn returns a memfilectx object. If the file was removed, filectxfn return None for recent Mercurial. Moved files are represented by marking the source file removed and the new file added with copy information (see memfilectx). user receives the committer name and defaults to current repository username, date is the commit date in any format supported by util.parsedate() and defaults to current date, extra is a dictionary of metadata or is left empty. """ # Mercurial <= 3.1 expects the filectxfn to raise IOError for missing files. # Extensions that need to retain compatibility across Mercurial 3.1 can use # this field to determine what to do in filectxfn. _returnnoneformissingfiles = True def __init__(self, repo, parents, text, files, filectxfn, user=None, date=None, extra=None, branch=None, editor=False): super(memctx, self).__init__(repo, text, user, date, extra) self._rev = None self._node = None parents = [(p or nullid) for p in parents] p1, p2 = parents self._parents = [changectx(self._repo, p) for p in (p1, p2)] files = sorted(set(files)) self._files = files if branch is not None: self._extra['branch'] = encoding.fromlocal(branch) self.substate = {} if isinstance(filectxfn, patch.filestore): filectxfn = memfilefrompatch(filectxfn) elif not callable(filectxfn): # if store is not callable, wrap it in a function filectxfn = memfilefromctx(filectxfn) # memoizing increases performance for e.g. vcs convert scenarios. self._filectxfn = makecachingfilectxfn(filectxfn) if editor: self._text = editor(self._repo, self, []) self._repo.savecommitmessage(self._text) def filectx(self, path, filelog=None): """get a file context from the working directory Returns None if file doesn't exist and should be removed.""" return self._filectxfn(self._repo, self, path) def commit(self): """commit context to the repo""" return self._repo.commitctx(self) @propertycache def _manifest(self): """generate a manifest based on the return values of filectxfn""" # keep this simple for now; just worry about p1 pctx = self._parents[0] man = pctx.manifest().copy() for f in self._status.modified: p1node = nullid p2node = nullid p = pctx[f].parents() # if file isn't in pctx, check p2? if len(p) > 0: p1node = p[0].filenode() if len(p) > 1: p2node = p[1].filenode() man[f] = revlog.hash(self[f].data(), p1node, p2node) for f in self._status.added: man[f] = revlog.hash(self[f].data(), nullid, nullid) for f in self._status.removed: if f in man: del man[f] return man @propertycache def _status(self): """Calculate exact status from ``files`` specified at construction """ man1 = self.p1().manifest() p2 = self._parents[1] # "1 < len(self._parents)" can't be used for checking # existence of the 2nd parent, because "memctx._parents" is # explicitly initialized by the list, of which length is 2. if p2.node() != nullid: man2 = p2.manifest() managing = lambda f: f in man1 or f in man2 else: managing = lambda f: f in man1 modified, added, removed = [], [], [] for f in self._files: if not managing(f): added.append(f) elif self[f]: modified.append(f) else: removed.append(f) return scmutil.status(modified, added, removed, [], [], [], []) class memfilectx(committablefilectx): """memfilectx represents an in-memory file to commit. See memctx and committablefilectx for more details. """ def __init__(self, repo, path, data, islink=False, isexec=False, copied=None, memctx=None): """ path is the normalized file path relative to repository root. data is the file content as a string. islink is True if the file is a symbolic link. isexec is True if the file is executable. copied is the source file path if current file was copied in the revision being committed, or None.""" super(memfilectx, self).__init__(repo, path, None, memctx) self._data = data self._flags = (islink and 'l' or '') + (isexec and 'x' or '') self._copied = None if copied: self._copied = (copied, nullid) def data(self): return self._data def remove(self, ignoremissing=False): """wraps unlink for a repo's working directory""" # need to figure out what to do here del self._changectx[self._path] def write(self, data, flags): """wraps repo.wwrite""" self._data = data class overlayfilectx(committablefilectx): """Like memfilectx but take an original filectx and optional parameters to override parts of it. This is useful when fctx.data() is expensive (i.e. flag processor is expensive) and raw data, flags, and filenode could be reused (ex. rebase or mode-only amend a REVIDX_EXTSTORED file). """ def __init__(self, originalfctx, datafunc=None, path=None, flags=None, copied=None, ctx=None): """originalfctx: filecontext to duplicate datafunc: None or a function to override data (file content). It is a function to be lazy. path, flags, copied, ctx: None or overridden value copied could be (path, rev), or False. copied could also be just path, and will be converted to (path, nullid). This simplifies some callers. """ if path is None: path = originalfctx.path() if ctx is None: ctx = originalfctx.changectx() ctxmatch = lambda: True else: ctxmatch = lambda: ctx == originalfctx.changectx() repo = originalfctx.repo() flog = originalfctx.filelog() super(overlayfilectx, self).__init__(repo, path, flog, ctx) if copied is None: copied = originalfctx.renamed() copiedmatch = lambda: True else: if copied and not isinstance(copied, tuple): # repo._filecommit will recalculate copyrev so nullid is okay copied = (copied, nullid) copiedmatch = lambda: copied == originalfctx.renamed() # When data, copied (could affect data), ctx (could affect filelog # parents) are not overridden, rawdata, rawflags, and filenode may be # reused (repo._filecommit should double check filelog parents). # # path, flags are not hashed in filelog (but in manifestlog) so they do # not affect reusable here. # # If ctx or copied is overridden to a same value with originalfctx, # still consider it's reusable. originalfctx.renamed() may be a bit # expensive so it's not called unless necessary. Assuming datafunc is # always expensive, do not call it for this "reusable" test. reusable = datafunc is None and ctxmatch() and copiedmatch() if datafunc is None: datafunc = originalfctx.data if flags is None: flags = originalfctx.flags() self._datafunc = datafunc self._flags = flags self._copied = copied if reusable: # copy extra fields from originalfctx attrs = ['rawdata', 'rawflags', '_filenode', '_filerev'] for attr_ in attrs: if util.safehasattr(originalfctx, attr_): setattr(self, attr_, getattr(originalfctx, attr_)) def data(self): return self._datafunc() class metadataonlyctx(committablectx): """Like memctx but it's reusing the manifest of different commit. Intended to be used by lightweight operations that are creating metadata-only changes. Revision information is supplied at initialization time. 'repo' is the current localrepo, 'ctx' is original revision which manifest we're reuisng 'parents' is a sequence of two parent revisions identifiers (pass None for every missing parent), 'text' is the commit. user receives the committer name and defaults to current repository username, date is the commit date in any format supported by util.parsedate() and defaults to current date, extra is a dictionary of metadata or is left empty. """ def __new__(cls, repo, originalctx, *args, **kwargs): return super(metadataonlyctx, cls).__new__(cls, repo) def __init__(self, repo, originalctx, parents=None, text=None, user=None, date=None, extra=None, editor=False): if text is None: text = originalctx.description() super(metadataonlyctx, self).__init__(repo, text, user, date, extra) self._rev = None self._node = None self._originalctx = originalctx self._manifestnode = originalctx.manifestnode() if parents is None: parents = originalctx.parents() else: parents = [repo[p] for p in parents if p is not None] parents = parents[:] while len(parents) < 2: parents.append(repo[nullid]) p1, p2 = self._parents = parents # sanity check to ensure that the reused manifest parents are # manifests of our commit parents mp1, mp2 = self.manifestctx().parents if p1 != nullid and p1.manifestnode() != mp1: raise RuntimeError('can\'t reuse the manifest: ' 'its p1 doesn\'t match the new ctx p1') if p2 != nullid and p2.manifestnode() != mp2: raise RuntimeError('can\'t reuse the manifest: ' 'its p2 doesn\'t match the new ctx p2') self._files = originalctx.files() self.substate = {} if editor: self._text = editor(self._repo, self, []) self._repo.savecommitmessage(self._text) def manifestnode(self): return self._manifestnode @property def _manifestctx(self): return self._repo.manifestlog[self._manifestnode] def filectx(self, path, filelog=None): return self._originalctx.filectx(path, filelog=filelog) def commit(self): """commit context to the repo""" return self._repo.commitctx(self) @property def _manifest(self): return self._originalctx.manifest() @propertycache def _status(self): """Calculate exact status from ``files`` specified in the ``origctx`` and parents manifests. """ man1 = self.p1().manifest() p2 = self._parents[1] # "1 < len(self._parents)" can't be used for checking # existence of the 2nd parent, because "metadataonlyctx._parents" is # explicitly initialized by the list, of which length is 2. if p2.node() != nullid: man2 = p2.manifest() managing = lambda f: f in man1 or f in man2 else: managing = lambda f: f in man1 modified, added, removed = [], [], [] for f in self._files: if not managing(f): added.append(f) elif f in self: modified.append(f) else: removed.append(f) return scmutil.status(modified, added, removed, [], [], [], []) class arbitraryfilectx(object): """Allows you to use filectx-like functions on a file in an arbitrary location on disk, possibly not in the working directory. """ def __init__(self, path, repo=None): # Repo is optional because contrib/simplemerge uses this class. self._repo = repo self._path = path def cmp(self, fctx): # filecmp follows symlinks whereas `cmp` should not, so skip the fast # path if either side is a symlink. symlinks = ('l' in self.flags() or 'l' in fctx.flags()) if not symlinks and isinstance(fctx, workingfilectx) and self._repo: # Add a fast-path for merge if both sides are disk-backed. # Note that filecmp uses the opposite return values (True if same) # from our cmp functions (True if different). return not filecmp.cmp(self.path(), self._repo.wjoin(fctx.path())) return self.data() != fctx.data() def path(self): return self._path def flags(self): return '' def data(self): return util.readfile(self._path) def decodeddata(self): with open(self._path, "rb") as f: return f.read() def remove(self): util.unlink(self._path) def write(self, data, flags): assert not flags with open(self._path, "w") as f: f.write(data)