Mercurial > hg
view hgext/remotenames.py @ 47866:4162f6b40f2c stable
windows: degrade to py2 behavior when reading a non-symlink as a symlink
While waiting for the push to hg-committed in WSL to complete, I ran a
`phabimport` from Windows and got this traceback:
$ hg phabimport 11313
** Unknown exception encountered with possibly-broken third-party extension "mercurial_keyring" (version N/A)
** which supports versions unknown of Mercurial.
** Please disable "mercurial_keyring" and try your action again.
** If that fixes the bug please report it to https://foss.heptapod.net/mercurial/mercurial_keyring/issues
** Python 3.9.5 (default, May 6 2021, 17:29:31) [MSC v.1928 64 bit (AMD64)]
** Mercurial Distributed SCM (version 5.9rc1+hg32.0e2f5733563d)
** Extensions loaded: absorb, blackbox, evolve 10.3.3, extdiff, fastannotate, fix, mercurial_keyring, mq, phabblocker 20210126, phabricator, rebase, show, strip, topic 0.22.3
Traceback (most recent call last):
File "mercurial.lock", line 279, in _trylock
File "mercurial.vfs", line 202, in makelock
File "mercurial.util", line 2147, in makelock
FileExistsError: [WinError 183] Cannot create a file when that file already exists: b'hp-omen:78348' -> b'C:\\Users\\Matt\\hg/.hg/store/lock'
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "<string>", line 24, in <module>
File "mercurial.dispatch", line 144, in run
File "mercurial.dispatch", line 250, in dispatch
File "mercurial.dispatch", line 294, in _rundispatch
File "mercurial.dispatch", line 470, in _runcatch
File "mercurial.dispatch", line 480, in _callcatch
File "mercurial.scmutil", line 153, in callcatch
File "mercurial.dispatch", line 460, in _runcatchfunc
File "mercurial.dispatch", line 1273, in _dispatch
File "mercurial.dispatch", line 918, in runcommand
File "mercurial.dispatch", line 1285, in _runcommand
File "mercurial.dispatch", line 1271, in <lambda>
File "mercurial.util", line 1886, in check
File "mercurial.util", line 1886, in check
File "hgext.mq", line 4239, in mqcommand
File "mercurial.util", line 1886, in check
File "mercurial.util", line 1886, in check
File "hgext.phabricator", line 314, in inner
File "hgext.phabricator", line 2222, in phabimport
File "hgext.phabricator", line 2123, in readpatch
File "hgext.phabricator", line 2199, in _write
File "mercurial.localrepo", line 2956, in lock
File "mercurial.localrepo", line 2918, in _lock
File "mercurial.lock", line 152, in trylock
File "mercurial.lock", line 283, in _trylock
File "mercurial.lock", line 314, in _readlock
File "mercurial.vfs", line 221, in readlock
File "mercurial.util", line 2163, in readlock
File "mercurial.windows", line 619, in readlink
ValueError: not a symbolic link
Both exceptions look accurate (the file exists, and the Windows side can't read
WSL side symlinks). I didn't try to reproduce this entirely within the Windows
side, but we can do better than a cryptic stacktrace. With this change, the
same scenario results in this abort:
abort: C:\Users\Matt\hg/.hg/store/lock: The file cannot be accessed by the system
When both the `push` and `phabimport` are done on the Windows side, it prints a
message about waiting for the lock, and successfully applies the patch after the
push completes.
I'm not sure if there's enough info to be able to convert the abort into the
wait scenario. As it stands now, we don't support symlinks on Windows, which
requires either a UAC Administrator level process or an opt-in in developer
mode, and there are several places where the new symlink on Windows support in
py3 was explicitly disabled in order to get tests to pass quicker.
Differential Revision: https://phab.mercurial-scm.org/D11333
author | Matt Harbison <matt_harbison@yahoo.com> |
---|---|
date | Sun, 22 Aug 2021 17:59:21 -0400 |
parents | f38bf44e077f |
children | 6000f5b25c9b |
line wrap: on
line source
# remotenames.py - extension to display remotenames # # Copyright 2017 Augie Fackler <raf@durin42.com> # Copyright 2017 Sean Farley <sean@farley.io> # # This software may be used and distributed according to the terms of the # GNU General Public License version 2 or any later version. """ showing remotebookmarks and remotebranches in UI (EXPERIMENTAL) By default both remotebookmarks and remotebranches are turned on. Config knob to control the individually are as follows. Config options to tweak the default behaviour: remotenames.bookmarks Boolean value to enable or disable showing of remotebookmarks (default: True) remotenames.branches Boolean value to enable or disable showing of remotebranches (default: True) remotenames.hoistedpeer Name of the peer whose remotebookmarks should be hoisted into the top-level namespace (default: 'default') """ from __future__ import absolute_import from mercurial.i18n import _ from mercurial.node import bin from mercurial import ( bookmarks, error, extensions, logexchange, namespaces, pycompat, registrar, revsetlang, smartset, templateutil, util, ) from mercurial.utils import stringutil if pycompat.ispy3: import collections.abc mutablemapping = collections.abc.MutableMapping else: import collections mutablemapping = collections.MutableMapping # Note for extension authors: ONLY specify testedwith = 'ships-with-hg-core' for # extensions which SHIP WITH MERCURIAL. Non-mainline extensions should # be specifying the version(s) of Mercurial they are tested with, or # leave the attribute unspecified. testedwith = b'ships-with-hg-core' configtable = {} configitem = registrar.configitem(configtable) templatekeyword = registrar.templatekeyword() revsetpredicate = registrar.revsetpredicate() configitem( b'remotenames', b'bookmarks', default=True, ) configitem( b'remotenames', b'branches', default=True, ) configitem( b'remotenames', b'hoistedpeer', default=b'default', ) class lazyremotenamedict(mutablemapping): """ Read-only dict-like Class to lazily resolve remotename entries We are doing that because remotenames startup was slow. We lazily read the remotenames file once to figure out the potential entries and store them in self.potentialentries. Then when asked to resolve an entry, if it is not in self.potentialentries, then it isn't there, if it is in self.potentialentries we resolve it and store the result in self.cache. We cannot be lazy is when asked all the entries (keys). """ def __init__(self, kind, repo): self.cache = {} self.potentialentries = {} self._kind = kind # bookmarks or branches self._repo = repo self.loaded = False def _load(self): """Read the remotenames file, store entries matching selected kind""" self.loaded = True repo = self._repo for node, rpath, rname in logexchange.readremotenamefile( repo, self._kind ): name = rpath + b'/' + rname self.potentialentries[name] = (node, rpath, name) def _resolvedata(self, potentialentry): """Check that the node for potentialentry exists and return it""" if not potentialentry in self.potentialentries: return None node, remote, name = self.potentialentries[potentialentry] repo = self._repo binnode = bin(node) # if the node doesn't exist, skip it try: repo.changelog.rev(binnode) except LookupError: return None # Skip closed branches if self._kind == b'branches' and repo[binnode].closesbranch(): return None return [binnode] def __getitem__(self, key): if not self.loaded: self._load() val = self._fetchandcache(key) if val is not None: return val else: raise KeyError() def __iter__(self): return iter(self.potentialentries) def __len__(self): return len(self.potentialentries) def __setitem__(self): raise NotImplementedError def __delitem__(self): raise NotImplementedError def _fetchandcache(self, key): if key in self.cache: return self.cache[key] val = self._resolvedata(key) if val is not None: self.cache[key] = val return val else: return None def keys(self): """Get a list of bookmark or branch names""" if not self.loaded: self._load() return self.potentialentries.keys() def iteritems(self): """Iterate over (name, node) tuples""" if not self.loaded: self._load() for k, vtup in pycompat.iteritems(self.potentialentries): yield (k, [bin(vtup[0])]) items = iteritems class remotenames(object): """ This class encapsulates all the remotenames state. It also contains methods to access that state in convenient ways. Remotenames are lazy loaded. Whenever client code needs to ensure the freshest copy of remotenames, use the `clearnames` method to force an eventual load. """ def __init__(self, repo, *args): self._repo = repo self.clearnames() def clearnames(self): """Clear all remote names state""" self.bookmarks = lazyremotenamedict(b"bookmarks", self._repo) self.branches = lazyremotenamedict(b"branches", self._repo) self._invalidatecache() def _invalidatecache(self): self._nodetobmarks = None self._nodetobranch = None self._hoisttonodes = None self._nodetohoists = None def bmarktonodes(self): return self.bookmarks def nodetobmarks(self): if not self._nodetobmarks: bmarktonodes = self.bmarktonodes() self._nodetobmarks = {} for name, node in pycompat.iteritems(bmarktonodes): self._nodetobmarks.setdefault(node[0], []).append(name) return self._nodetobmarks def branchtonodes(self): return self.branches def nodetobranch(self): if not self._nodetobranch: branchtonodes = self.branchtonodes() self._nodetobranch = {} for name, nodes in pycompat.iteritems(branchtonodes): for node in nodes: self._nodetobranch.setdefault(node, []).append(name) return self._nodetobranch def hoisttonodes(self, hoist): if not self._hoisttonodes: marktonodes = self.bmarktonodes() self._hoisttonodes = {} hoist += b'/' for name, node in pycompat.iteritems(marktonodes): if name.startswith(hoist): name = name[len(hoist) :] self._hoisttonodes[name] = node return self._hoisttonodes def nodetohoists(self, hoist): if not self._nodetohoists: marktonodes = self.bmarktonodes() self._nodetohoists = {} hoist += b'/' for name, node in pycompat.iteritems(marktonodes): if name.startswith(hoist): name = name[len(hoist) :] self._nodetohoists.setdefault(node[0], []).append(name) return self._nodetohoists def wrapprintbookmarks(orig, ui, repo, fm, bmarks): if b'remotebookmarks' not in repo.names: return ns = repo.names[b'remotebookmarks'] for name in ns.listnames(repo): nodes = ns.nodes(repo, name) if not nodes: continue node = nodes[0] bmarks[name] = (node, b' ', b'') return orig(ui, repo, fm, bmarks) def extsetup(ui): extensions.wrapfunction(bookmarks, b'_printbookmarks', wrapprintbookmarks) def reposetup(ui, repo): # set the config option to store remotenames repo.ui.setconfig(b'experimental', b'remotenames', True, b'remotenames-ext') if not repo.local(): return repo._remotenames = remotenames(repo) ns = namespaces.namespace if ui.configbool(b'remotenames', b'bookmarks'): remotebookmarkns = ns( b'remotebookmarks', templatename=b'remotebookmarks', colorname=b'remotebookmark', logfmt=b'remote bookmark: %s\n', listnames=lambda repo: repo._remotenames.bmarktonodes().keys(), namemap=lambda repo, name: repo._remotenames.bmarktonodes().get( name, [] ), nodemap=lambda repo, node: repo._remotenames.nodetobmarks().get( node, [] ), ) repo.names.addnamespace(remotebookmarkns) # hoisting only works if there are remote bookmarks hoist = ui.config(b'remotenames', b'hoistedpeer') if hoist: hoistednamens = ns( b'hoistednames', templatename=b'hoistednames', colorname=b'hoistedname', logfmt=b'hoisted name: %s\n', listnames=lambda repo: repo._remotenames.hoisttonodes( hoist ).keys(), namemap=lambda repo, name: repo._remotenames.hoisttonodes( hoist ).get(name, []), nodemap=lambda repo, node: repo._remotenames.nodetohoists( hoist ).get(node, []), ) repo.names.addnamespace(hoistednamens) if ui.configbool(b'remotenames', b'branches'): remotebranchns = ns( b'remotebranches', templatename=b'remotebranches', colorname=b'remotebranch', logfmt=b'remote branch: %s\n', listnames=lambda repo: repo._remotenames.branchtonodes().keys(), namemap=lambda repo, name: repo._remotenames.branchtonodes().get( name, [] ), nodemap=lambda repo, node: repo._remotenames.nodetobranch().get( node, [] ), ) repo.names.addnamespace(remotebranchns) @templatekeyword(b'remotenames', requires={b'repo', b'ctx'}) def remotenameskw(context, mapping): """List of strings. Remote names associated with the changeset.""" repo = context.resource(mapping, b'repo') ctx = context.resource(mapping, b'ctx') remotenames = [] if b'remotebookmarks' in repo.names: remotenames = repo.names[b'remotebookmarks'].names(repo, ctx.node()) if b'remotebranches' in repo.names: remotenames += repo.names[b'remotebranches'].names(repo, ctx.node()) return templateutil.compatlist( context, mapping, b'remotename', remotenames, plural=b'remotenames' ) @templatekeyword(b'remotebookmarks', requires={b'repo', b'ctx'}) def remotebookmarkskw(context, mapping): """List of strings. Remote bookmarks associated with the changeset.""" repo = context.resource(mapping, b'repo') ctx = context.resource(mapping, b'ctx') remotebmarks = [] if b'remotebookmarks' in repo.names: remotebmarks = repo.names[b'remotebookmarks'].names(repo, ctx.node()) return templateutil.compatlist( context, mapping, b'remotebookmark', remotebmarks, plural=b'remotebookmarks', ) @templatekeyword(b'remotebranches', requires={b'repo', b'ctx'}) def remotebrancheskw(context, mapping): """List of strings. Remote branches associated with the changeset.""" repo = context.resource(mapping, b'repo') ctx = context.resource(mapping, b'ctx') remotebranches = [] if b'remotebranches' in repo.names: remotebranches = repo.names[b'remotebranches'].names(repo, ctx.node()) return templateutil.compatlist( context, mapping, b'remotebranch', remotebranches, plural=b'remotebranches', ) def _revsetutil(repo, subset, x, rtypes): """utility function to return a set of revs based on the rtypes""" args = revsetlang.getargs(x, 0, 1, _(b'only one argument accepted')) if args: kind, pattern, matcher = stringutil.stringmatcher( revsetlang.getstring(args[0], _(b'argument must be a string')) ) else: kind = pattern = None matcher = util.always nodes = set() cl = repo.changelog for rtype in rtypes: if rtype in repo.names: ns = repo.names[rtype] for name in ns.listnames(repo): if not matcher(name): continue nodes.update(ns.nodes(repo, name)) if kind == b'literal' and not nodes: raise error.RepoLookupError( _(b"remote name '%s' does not exist") % pattern ) revs = (cl.rev(n) for n in nodes if cl.hasnode(n)) return subset & smartset.baseset(revs) @revsetpredicate(b'remotenames([name])') def remotenamesrevset(repo, subset, x): """All changesets which have a remotename on them. If `name` is specified, only remotenames of matching remote paths are considered. Pattern matching is supported for `name`. See :hg:`help revisions.patterns`. """ return _revsetutil(repo, subset, x, (b'remotebookmarks', b'remotebranches')) @revsetpredicate(b'remotebranches([name])') def remotebranchesrevset(repo, subset, x): """All changesets which are branch heads on remotes. If `name` is specified, only remotenames of matching remote paths are considered. Pattern matching is supported for `name`. See :hg:`help revisions.patterns`. """ return _revsetutil(repo, subset, x, (b'remotebranches',)) @revsetpredicate(b'remotebookmarks([name])') def remotebmarksrevset(repo, subset, x): """All changesets which have bookmarks on remotes. If `name` is specified, only remotenames of matching remote paths are considered. Pattern matching is supported for `name`. See :hg:`help revisions.patterns`. """ return _revsetutil(repo, subset, x, (b'remotebookmarks',))