Mercurial > hg
view mercurial/phases.py @ 39506:b66ea3fc3a86
sparse-revlog: set max delta chain length to on thousand
The new snapshot system used in the sparse-revlog case gave us some small size
benefit so far. However its most important property is to gracefully handle
harder limit on delta chainlength.
Long delta chain has a very detrimental impact on read (and write) performance
in revlog. Being able to shorter them provide a great boost. However, shorting
delta used to result significantly lower compression ratio. The intermediate
snapshots effectively suppress most of this effect (even all in some case).
# Effect on the test repository
The repository we use for test is not "realistic" but can still show this in
action using an unreasonably low chain limit. Limiting the chain length show a
sizeable increase but stay under control: +6% for limit=15; +15% for limit=10.
Without the snapshot system the increase is significantly bigger: +45% for
limit=15; +80% for limit=10. Even slightly larger than without delta chain
limit, the resulting size is still smaller than before we started doing
snapshots.
Here is a table for comparison. *Since the repository is not branchy, the
initial sparse-revlog version does not bring much benefit compare to the
non-sparse one):
chain length limit | none | limit=15 | limit=10 |
without sparse-revlog | 62 818 987 | 112 664 615 | 131 222 574 |
without snapshot | 74 365 490 | 108 211 410 | 133 857 764 |
with snapshot | 59 230 936 | 63 002 924 | 68 415 329 |
# Effect On Real Life Repositories
The series provides significant benefits on all kind of repositories.
Using `hg debugupgraderepo -o redeltaparent --run`, we recomputed delta chain
for various repositories with different settings:
- delta chain length: unlimited or 1000 limit
- sparse-revlog: enabled or disabled
- this series: applied or not applied
We can observe multiple types of effect:
- On very branchy repositories:
* The delta chain limit as low impact on the repo size.
* Intermediate snapshot greatly reduces manifest size:
- pypy: -80%
- netbeans: -95%
* The delta chain limit is effective, without a size impact:
- netbeans average: 613 -> 282
- private #1 average: 1 068 -> 307
- On more linear repository:
* Intermediate snapshot limit the impact of delta chain limit:
- mozilla:
without the series: +360%
with the series: +25%
* The delta chain limit provides large improvement:
- mozilla's average chain length:
unlimited: 15 338
limited: 469
* Despite the chain length limit, the manifest size is reduced:
- mercurial: -25%
- mozilla: -30%
It is clear that the use of chains of intermediate snapshots provide large
benefits both in storage size and delta chains quality. We should now switch our
effort toward making sure the write performance are acceptable. Then,
`sparse-revlog` will be a suitable format for all new repository.
# Raw Statistic
* no-sparse: general delta repository not using sparse-revlog
* no-snapshot: sparse-revlog repository not using this series
* snapshot: sparse-revlog repository using this series
mercurial
Manifest Size:
limit | none | 1000
------------|-------------|------------
no-sparse | 8 021 373 | 8 199 366
no-snapshot | 8 103 561 | 8 259 719
snapshot | 6 137 116 | 6 126 433
Manifest Chain length data
limit || none || 1000 ||
value || average | max || average | max ||
------------||---------|---------||---------|---------||
no-sparse || 307 | 1456 || 279 | 1000 ||
no-snapshot || 312 | 1456 || 283 | 1000 ||
snapshot || 248 | 1208 || 241 | 1000 ||
Full Store Size
limit | none | 1000
------------|-------------|------------
no-sparse | 51 013 198 | 51 201 574
no-snapshot | 50 930 795 | 51 141 006
snapshot | 48 072 037 | 48 093 572
pypy
Manifest Size:
limit | none | 1000
------------|-------------|------------
no-sparse | 193 987 784 | 193 987 784
no-snapshot | 163 171 745 | 163 312 229
snapshot | 34 605 900 | 34 600 750
Manifest Chain length data
limit || none || 1000 ||
value || average | max || average | max ||
------------||---------|---------||---------|---------||
no-sparse || 101 | 692 || 101 | 692 ||
no-snapshot || 151 | 1307 || 148 | 1000 ||
snapshot || 128 | 1309 || 125 | 1000 ||
Full Store Size
limit | none | 1000
------------|-------------|------------
no-sparse | 495 931 473 | 495 931 473
no-snapshot | 465 441 017 | 465 581 501
snapshot | 355 467 301 | 355 472 451
Mozilla
Manifest Size:
limit | none | 1000
------------|----------------|---------------
no-sparse | 416 757 148 | 1 869 009 668
no-snapshot | 401 592 370 | 1 843 493 795
snapshot | 224 359 521 | 284 615 500
Manifest Chain length data
limit || none || 1000 ||
value || average | max || average | max ||
------------||---------|---------||---------|---------||
no-sparse || 15 333 | 58 980 || 468 | 1 000 ||
no-snapshot || 15 336 | 58 980 || 469 | 1 000 ||
snapshot || 15 338 | 58 983 || 469 | 1 000 ||
Full Store Size
limit | none | 1000
------------|----------------|---------------
no-sparse | 2 712 477 887 | 4 164 995 451
no-snapshot | 2 698 887 835 | 4 141 054 304
snapshot | 2 518 130 385 | 2 578 587 596
Netbeans
Manifest Size:
limit | none | 1000
------------|----------------|---------------
no-sparse | 4 766 794 101 | 4 870 642 687
no-snapshot | 4 334 806 082 | 4 428 681 309
snapshot | 232 659 666 | 240 330 665
Manifest Chain length data
limit || none || 1000 ||
value || average | max || average | max ||
------------||---------|---------||---------|---------||
no-sparse || 597 | 6802 || 254 | 1 000 ||
no-snapshot || 648 | 6 802 || 305 | 1 000 ||
snapshot || 613 | 6 804 || 282 | 1 000 ||
Full Store Size
limit | none | 1000
------------|----------------|---------------
no-sparse | 5 807 347 998 | 5 911 196 584
no-snapshot | 5 375 398 602 | 5 469 273 829
snapshot | 1 282 519 928 | 1 290 190 927
Private repo #1
Manifest Size:
limit | none | 1000
------------|-----------------|---------------
no-sparse | 41 389 010 840 | 41 398 162 091
no-snapshot | 9 737 319 435 | 10 223 773 150
snapshot | 744 215 807 | 747 961 822
Manifest Chain length data
limit || none || 1000 ||
value || average | max || average | max ||
------------||---------|---------||---------|---------||
no-sparse || 245 | 8 885 || 81 | 1 000 ||
no-snapshot || 1 225 | 8 885 || 336 | 1 000 ||
snapshot || 1 068 | 7 909 || 307 | 1 000 ||
Full Store Size
limit | none | 1000
------------|----------------|---------------
no-sparse | 49 646 065 126 | 49 655 216 377
no-snapshot | 17 924 862 856 | 18 411 316 571
snapshot | 9 009 024 710 | 9 012 770 725
Private repo #2
We currently have less data available for that repository.
* Before is a sparse-revlog repository without this series
* After is a sparse-revlog repository with this series + 1000 chain limit
Manifest Size:
Before: 1 531 485 040 bytes
After: 1 091 422 451 bytes
Manifest Chain:
Before: 2 218 avg; 6 575 Max
After: 442 avg; 1 000 Max
Full Store Size
Before: 15 203 955 615
after: 8 207 180 693
author | Boris Feld <boris.feld@octobus.net> |
---|---|
date | Fri, 07 Sep 2018 11:18:45 -0400 |
parents | 7775c1fb8fa0 |
children | 49c7b701fdc2 |
line wrap: on
line source
""" Mercurial phases support code --- Copyright 2011 Pierre-Yves David <pierre-yves.david@ens-lyon.org> Logilab SA <contact@logilab.fr> Augie Fackler <durin42@gmail.com> This software may be used and distributed according to the terms of the GNU General Public License version 2 or any later version. --- This module implements most phase logic in mercurial. Basic Concept ============= A 'changeset phase' is an indicator that tells us how a changeset is manipulated and communicated. The details of each phase is described below, here we describe the properties they have in common. Like bookmarks, phases are not stored in history and thus are not permanent and leave no audit trail. First, no changeset can be in two phases at once. Phases are ordered, so they can be considered from lowest to highest. The default, lowest phase is 'public' - this is the normal phase of existing changesets. A child changeset can not be in a lower phase than its parents. These phases share a hierarchy of traits: immutable shared public: X X draft: X secret: Local commits are draft by default. Phase Movement and Exchange =========================== Phase data is exchanged by pushkey on pull and push. Some servers have a publish option set, we call such a server a "publishing server". Pushing a draft changeset to a publishing server changes the phase to public. A small list of fact/rules define the exchange of phase: * old client never changes server states * pull never changes server states * publish and old server changesets are seen as public by client * any secret changeset seen in another repository is lowered to at least draft Here is the final table summing up the 49 possible use cases of phase exchange: server old publish non-publish N X N D P N D P old client pull N - X/X - X/D X/P - X/D X/P X - X/X - X/D X/P - X/D X/P push X X/X X/X X/P X/P X/P X/D X/D X/P new client pull N - P/X - P/D P/P - D/D P/P D - P/X - P/D P/P - D/D P/P P - P/X - P/D P/P - P/D P/P push D P/X P/X P/P P/P P/P D/D D/D P/P P P/X P/X P/P P/P P/P P/P P/P P/P Legend: A/B = final state on client / state on server * N = new/not present, * P = public, * D = draft, * X = not tracked (i.e., the old client or server has no internal way of recording the phase.) passive = only pushes A cell here can be read like this: "When a new client pushes a draft changeset (D) to a publishing server where it's not present (N), it's marked public on both sides (P/P)." Note: old client behave as a publishing server with draft only content - other people see it as public - content is pushed as draft """ from __future__ import absolute_import import errno import struct from .i18n import _ from .node import ( bin, hex, nullid, nullrev, short, ) from . import ( error, pycompat, smartset, txnutil, util, ) _fphasesentry = struct.Struct('>i20s') INTERNAL_FLAG = 64 # Phases for mercurial internal usage only HIDEABLE_FLAG = 32 # Phases that are hideable # record phase index public, draft, secret = range(3) internal = INTERNAL_FLAG | HIDEABLE_FLAG allphases = range(internal + 1) trackedphases = allphases[1:] # record phase names phasenames = [None] * len(allphases) phasenames[:3] = ['public', 'draft', 'secret'] phasenames[internal] = 'internal' # record phase property mutablephases = tuple(allphases[1:]) remotehiddenphases = tuple(allphases[2:]) localhiddenphases = tuple(p for p in allphases if p & HIDEABLE_FLAG) def supportinternal(repo): """True if the internal phase can be used on a repository""" return 'internal-phase' in repo.requirements def _readroots(repo, phasedefaults=None): """Read phase roots from disk phasedefaults is a list of fn(repo, roots) callable, which are executed if the phase roots file does not exist. When phases are being initialized on an existing repository, this could be used to set selected changesets phase to something else than public. Return (roots, dirty) where dirty is true if roots differ from what is being stored. """ repo = repo.unfiltered() dirty = False roots = [set() for i in allphases] try: f, pending = txnutil.trypending(repo.root, repo.svfs, 'phaseroots') try: for line in f: phase, nh = line.split() roots[int(phase)].add(bin(nh)) finally: f.close() except IOError as inst: if inst.errno != errno.ENOENT: raise if phasedefaults: for f in phasedefaults: roots = f(repo, roots) dirty = True return roots, dirty def binaryencode(phasemapping): """encode a 'phase -> nodes' mapping into a binary stream Since phases are integer the mapping is actually a python list: [[PUBLIC_HEADS], [DRAFTS_HEADS], [SECRET_HEADS]] """ binarydata = [] for phase, nodes in enumerate(phasemapping): for head in nodes: binarydata.append(_fphasesentry.pack(phase, head)) return ''.join(binarydata) def binarydecode(stream): """decode a binary stream into a 'phase -> nodes' mapping Since phases are integer the mapping is actually a python list.""" headsbyphase = [[] for i in allphases] entrysize = _fphasesentry.size while True: entry = stream.read(entrysize) if len(entry) < entrysize: if entry: raise error.Abort(_('bad phase-heads stream')) break phase, node = _fphasesentry.unpack(entry) headsbyphase[phase].append(node) return headsbyphase def _trackphasechange(data, rev, old, new): """add a phase move the <data> dictionnary If data is None, nothing happens. """ if data is None: return existing = data.get(rev) if existing is not None: old = existing[0] data[rev] = (old, new) class phasecache(object): def __init__(self, repo, phasedefaults, _load=True): if _load: # Cheap trick to allow shallow-copy without copy module self.phaseroots, self.dirty = _readroots(repo, phasedefaults) self._loadedrevslen = 0 self._phasesets = None self.filterunknown(repo) self.opener = repo.svfs def getrevset(self, repo, phases, subset=None): """return a smartset for the given phases""" self.loadphaserevs(repo) # ensure phase's sets are loaded phases = set(phases) if public not in phases: # fast path: _phasesets contains the interesting sets, # might only need a union and post-filtering. if len(phases) == 1: [p] = phases revs = self._phasesets[p] else: revs = set.union(*[self._phasesets[p] for p in phases]) if repo.changelog.filteredrevs: revs = revs - repo.changelog.filteredrevs if subset is None: return smartset.baseset(revs) else: return subset & smartset.baseset(revs) else: phases = set(allphases).difference(phases) if not phases: return smartset.fullreposet(repo) if len(phases) == 1: [p] = phases revs = self._phasesets[p] else: revs = set.union(*[self._phasesets[p] for p in phases]) if subset is None: subset = smartset.fullreposet(repo) if not revs: return subset return subset.filter(lambda r: r not in revs) def copy(self): # Shallow copy meant to ensure isolation in # advance/retractboundary(), nothing more. ph = self.__class__(None, None, _load=False) ph.phaseroots = self.phaseroots[:] ph.dirty = self.dirty ph.opener = self.opener ph._loadedrevslen = self._loadedrevslen ph._phasesets = self._phasesets return ph def replace(self, phcache): """replace all values in 'self' with content of phcache""" for a in ('phaseroots', 'dirty', 'opener', '_loadedrevslen', '_phasesets'): setattr(self, a, getattr(phcache, a)) def _getphaserevsnative(self, repo): repo = repo.unfiltered() nativeroots = [] for phase in trackedphases: nativeroots.append(pycompat.maplist(repo.changelog.rev, self.phaseroots[phase])) return repo.changelog.computephases(nativeroots) def _computephaserevspure(self, repo): repo = repo.unfiltered() cl = repo.changelog self._phasesets = [set() for phase in allphases] lowerroots = set() for phase in reversed(trackedphases): roots = pycompat.maplist(cl.rev, self.phaseroots[phase]) if roots: ps = set(cl.descendants(roots)) for root in roots: ps.add(root) ps.difference_update(lowerroots) lowerroots.update(ps) self._phasesets[phase] = ps self._loadedrevslen = len(cl) def loadphaserevs(self, repo): """ensure phase information is loaded in the object""" if self._phasesets is None: try: res = self._getphaserevsnative(repo) self._loadedrevslen, self._phasesets = res except AttributeError: self._computephaserevspure(repo) def invalidate(self): self._loadedrevslen = 0 self._phasesets = None def phase(self, repo, rev): # We need a repo argument here to be able to build _phasesets # if necessary. The repository instance is not stored in # phasecache to avoid reference cycles. The changelog instance # is not stored because it is a filecache() property and can # be replaced without us being notified. if rev == nullrev: return public if rev < nullrev: raise ValueError(_('cannot lookup negative revision')) if rev >= self._loadedrevslen: self.invalidate() self.loadphaserevs(repo) for phase in trackedphases: if rev in self._phasesets[phase]: return phase return public def write(self): if not self.dirty: return f = self.opener('phaseroots', 'w', atomictemp=True, checkambig=True) try: self._write(f) finally: f.close() def _write(self, fp): for phase, roots in enumerate(self.phaseroots): for h in sorted(roots): fp.write('%i %s\n' % (phase, hex(h))) self.dirty = False def _updateroots(self, phase, newroots, tr): self.phaseroots[phase] = newroots self.invalidate() self.dirty = True tr.addfilegenerator('phase', ('phaseroots',), self._write) tr.hookargs['phases_moved'] = '1' def registernew(self, repo, tr, targetphase, nodes): repo = repo.unfiltered() self._retractboundary(repo, tr, targetphase, nodes) if tr is not None and 'phases' in tr.changes: phasetracking = tr.changes['phases'] torev = repo.changelog.rev phase = self.phase for n in nodes: rev = torev(n) revphase = phase(repo, rev) _trackphasechange(phasetracking, rev, None, revphase) repo.invalidatevolatilesets() def advanceboundary(self, repo, tr, targetphase, nodes, dryrun=None): """Set all 'nodes' to phase 'targetphase' Nodes with a phase lower than 'targetphase' are not affected. If dryrun is True, no actions will be performed Returns a set of revs whose phase is changed or should be changed """ # Be careful to preserve shallow-copied values: do not update # phaseroots values, replace them. if tr is None: phasetracking = None else: phasetracking = tr.changes.get('phases') repo = repo.unfiltered() changes = set() # set of revisions to be changed delroots = [] # set of root deleted by this path for phase in pycompat.xrange(targetphase + 1, len(allphases)): # filter nodes that are not in a compatible phase already nodes = [n for n in nodes if self.phase(repo, repo[n].rev()) >= phase] if not nodes: break # no roots to move anymore olds = self.phaseroots[phase] affected = repo.revs('%ln::%ln', olds, nodes) changes.update(affected) if dryrun: continue for r in affected: _trackphasechange(phasetracking, r, self.phase(repo, r), targetphase) roots = set(ctx.node() for ctx in repo.set( 'roots((%ln::) - %ld)', olds, affected)) if olds != roots: self._updateroots(phase, roots, tr) # some roots may need to be declared for lower phases delroots.extend(olds - roots) if not dryrun: # declare deleted root in the target phase if targetphase != 0: self._retractboundary(repo, tr, targetphase, delroots) repo.invalidatevolatilesets() return changes def retractboundary(self, repo, tr, targetphase, nodes): oldroots = self.phaseroots[:targetphase + 1] if tr is None: phasetracking = None else: phasetracking = tr.changes.get('phases') repo = repo.unfiltered() if (self._retractboundary(repo, tr, targetphase, nodes) and phasetracking is not None): # find the affected revisions new = self.phaseroots[targetphase] old = oldroots[targetphase] affected = set(repo.revs('(%ln::) - (%ln::)', new, old)) # find the phase of the affected revision for phase in pycompat.xrange(targetphase, -1, -1): if phase: roots = oldroots[phase] revs = set(repo.revs('%ln::%ld', roots, affected)) affected -= revs else: # public phase revs = affected for r in revs: _trackphasechange(phasetracking, r, phase, targetphase) repo.invalidatevolatilesets() def _retractboundary(self, repo, tr, targetphase, nodes): # Be careful to preserve shallow-copied values: do not update # phaseroots values, replace them. if targetphase == internal and not supportinternal(repo): msg = 'this repository does not support the internal phase' raise error.ProgrammingError(msg) repo = repo.unfiltered() currentroots = self.phaseroots[targetphase] finalroots = oldroots = set(currentroots) newroots = [n for n in nodes if self.phase(repo, repo[n].rev()) < targetphase] if newroots: if nullid in newroots: raise error.Abort(_('cannot change null revision phase')) currentroots = currentroots.copy() currentroots.update(newroots) # Only compute new roots for revs above the roots that are being # retracted. minnewroot = min(repo[n].rev() for n in newroots) aboveroots = [n for n in currentroots if repo[n].rev() >= minnewroot] updatedroots = repo.set('roots(%ln::)', aboveroots) finalroots = set(n for n in currentroots if repo[n].rev() < minnewroot) finalroots.update(ctx.node() for ctx in updatedroots) if finalroots != oldroots: self._updateroots(targetphase, finalroots, tr) return True return False def filterunknown(self, repo): """remove unknown nodes from the phase boundary Nothing is lost as unknown nodes only hold data for their descendants. """ filtered = False nodemap = repo.changelog.nodemap # to filter unknown nodes for phase, nodes in enumerate(self.phaseroots): missing = sorted(node for node in nodes if node not in nodemap) if missing: for mnode in missing: repo.ui.debug( 'removing unknown node %s from %i-phase boundary\n' % (short(mnode), phase)) nodes.symmetric_difference_update(missing) filtered = True if filtered: self.dirty = True # filterunknown is called by repo.destroyed, we may have no changes in # root but _phasesets contents is certainly invalid (or at least we # have not proper way to check that). related to issue 3858. # # The other caller is __init__ that have no _phasesets initialized # anyway. If this change we should consider adding a dedicated # "destroyed" function to phasecache or a proper cache key mechanism # (see branchmap one) self.invalidate() def advanceboundary(repo, tr, targetphase, nodes, dryrun=None): """Add nodes to a phase changing other nodes phases if necessary. This function move boundary *forward* this means that all nodes are set in the target phase or kept in a *lower* phase. Simplify boundary to contains phase roots only. If dryrun is True, no actions will be performed Returns a set of revs whose phase is changed or should be changed """ phcache = repo._phasecache.copy() changes = phcache.advanceboundary(repo, tr, targetphase, nodes, dryrun=dryrun) if not dryrun: repo._phasecache.replace(phcache) return changes def retractboundary(repo, tr, targetphase, nodes): """Set nodes back to a phase changing other nodes phases if necessary. This function move boundary *backward* this means that all nodes are set in the target phase or kept in a *higher* phase. Simplify boundary to contains phase roots only.""" phcache = repo._phasecache.copy() phcache.retractboundary(repo, tr, targetphase, nodes) repo._phasecache.replace(phcache) def registernew(repo, tr, targetphase, nodes): """register a new revision and its phase Code adding revisions to the repository should use this function to set new changeset in their target phase (or higher). """ phcache = repo._phasecache.copy() phcache.registernew(repo, tr, targetphase, nodes) repo._phasecache.replace(phcache) def listphases(repo): """List phases root for serialization over pushkey""" # Use ordered dictionary so behavior is deterministic. keys = util.sortdict() value = '%i' % draft cl = repo.unfiltered().changelog for root in repo._phasecache.phaseroots[draft]: if repo._phasecache.phase(repo, cl.rev(root)) <= draft: keys[hex(root)] = value if repo.publishing(): # Add an extra data to let remote know we are a publishing # repo. Publishing repo can't just pretend they are old repo. # When pushing to a publishing repo, the client still need to # push phase boundary # # Push do not only push changeset. It also push phase data. # New phase data may apply to common changeset which won't be # push (as they are common). Here is a very simple example: # # 1) repo A push changeset X as draft to repo B # 2) repo B make changeset X public # 3) repo B push to repo A. X is not pushed but the data that # X as now public should # # The server can't handle it on it's own as it has no idea of # client phase data. keys['publishing'] = 'True' return keys def pushphase(repo, nhex, oldphasestr, newphasestr): """List phases root for serialization over pushkey""" repo = repo.unfiltered() with repo.lock(): currentphase = repo[nhex].phase() newphase = abs(int(newphasestr)) # let's avoid negative index surprise oldphase = abs(int(oldphasestr)) # let's avoid negative index surprise if currentphase == oldphase and newphase < oldphase: with repo.transaction('pushkey-phase') as tr: advanceboundary(repo, tr, newphase, [bin(nhex)]) return True elif currentphase == newphase: # raced, but got correct result return True else: return False def subsetphaseheads(repo, subset): """Finds the phase heads for a subset of a history Returns a list indexed by phase number where each item is a list of phase head nodes. """ cl = repo.changelog headsbyphase = [[] for i in allphases] # No need to keep track of secret phase; any heads in the subset that # are not mentioned are implicitly secret. for phase in allphases[:secret]: revset = "heads(%%ln & %s())" % phasenames[phase] headsbyphase[phase] = [cl.node(r) for r in repo.revs(revset, subset)] return headsbyphase def updatephases(repo, trgetter, headsbyphase): """Updates the repo with the given phase heads""" # Now advance phase boundaries of all but secret phase # # run the update (and fetch transaction) only if there are actually things # to update. This avoid creating empty transaction during no-op operation. for phase in allphases[:-1]: revset = '%ln - _phase(%s)' heads = [c.node() for c in repo.set(revset, headsbyphase[phase], phase)] if heads: advanceboundary(repo, trgetter(), phase, heads) def analyzeremotephases(repo, subset, roots): """Compute phases heads and root in a subset of node from root dict * subset is heads of the subset * roots is {<nodeid> => phase} mapping. key and value are string. Accept unknown element input """ repo = repo.unfiltered() # build list from dictionary draftroots = [] nodemap = repo.changelog.nodemap # to filter unknown nodes for nhex, phase in roots.iteritems(): if nhex == 'publishing': # ignore data related to publish option continue node = bin(nhex) phase = int(phase) if phase == public: if node != nullid: repo.ui.warn(_('ignoring inconsistent public root' ' from remote: %s\n') % nhex) elif phase == draft: if node in nodemap: draftroots.append(node) else: repo.ui.warn(_('ignoring unexpected root from remote: %i %s\n') % (phase, nhex)) # compute heads publicheads = newheads(repo, subset, draftroots) return publicheads, draftroots class remotephasessummary(object): """summarize phase information on the remote side :publishing: True is the remote is publishing :publicheads: list of remote public phase heads (nodes) :draftheads: list of remote draft phase heads (nodes) :draftroots: list of remote draft phase root (nodes) """ def __init__(self, repo, remotesubset, remoteroots): unfi = repo.unfiltered() self._allremoteroots = remoteroots self.publishing = remoteroots.get('publishing', False) ana = analyzeremotephases(repo, remotesubset, remoteroots) self.publicheads, self.draftroots = ana # Get the list of all "heads" revs draft on remote dheads = unfi.set('heads(%ln::%ln)', self.draftroots, remotesubset) self.draftheads = [c.node() for c in dheads] def newheads(repo, heads, roots): """compute new head of a subset minus another * `heads`: define the first subset * `roots`: define the second we subtract from the first""" # prevent an import cycle # phases > dagop > patch > copies > scmutil > obsolete > obsutil > phases from . import dagop repo = repo.unfiltered() cl = repo.changelog rev = cl.nodemap.get if not roots: return heads if not heads or heads == [nullid]: return [] # The logic operated on revisions, convert arguments early for convenience new_heads = set(rev(n) for n in heads if n != nullid) roots = [rev(n) for n in roots] # compute the area we need to remove affected_zone = repo.revs("(%ld::%ld)", roots, new_heads) # heads in the area are no longer heads new_heads.difference_update(affected_zone) # revisions in the area have children outside of it, # They might be new heads candidates = repo.revs("parents(%ld + (%ld and merge())) and not null", roots, affected_zone) candidates -= affected_zone if new_heads or candidates: # remove candidate that are ancestors of other heads new_heads.update(candidates) prunestart = repo.revs("parents(%ld) and not null", new_heads) pruned = dagop.reachableroots(repo, candidates, prunestart) new_heads.difference_update(pruned) return pycompat.maplist(cl.node, sorted(new_heads)) def newcommitphase(ui): """helper to get the target phase of new commit Handle all possible values for the phases.new-commit options. """ v = ui.config('phases', 'new-commit') try: return phasenames.index(v) except ValueError: try: return int(v) except ValueError: msg = _("phases.new-commit: not a valid phase name ('%s')") raise error.ConfigError(msg % v) def hassecret(repo): """utility function that check if a repo have any secret changeset.""" return bool(repo._phasecache.phaseroots[2]) def preparehookargs(node, old, new): if old is None: old = '' else: old = phasenames[old] return {'node': node, 'oldphase': old, 'phase': phasenames[new]}