Mercurial > hg
view hgext/rebase.py @ 48792:8bbb1abb9d19
tracked-file: rename the format option to use `use-`
This is more consistent with the other options.
Differential Revision: https://phab.mercurial-scm.org/D12200
author | Pierre-Yves David <pierre-yves.david@octobus.net> |
---|---|
date | Thu, 17 Feb 2022 06:41:54 +0100 |
parents | 402a6b6173e9 |
children | 6000f5b25c9b |
line wrap: on
line source
# rebase.py - rebasing feature for mercurial # # Copyright 2008 Stefano Tortarolo <stefano.tortarolo at gmail dot com> # # This software may be used and distributed according to the terms of the # GNU General Public License version 2 or any later version. '''command to move sets of revisions to a different ancestor This extension lets you rebase changesets in an existing Mercurial repository. For more information: https://mercurial-scm.org/wiki/RebaseExtension ''' from __future__ import absolute_import import errno import os from mercurial.i18n import _ from mercurial.node import ( nullrev, short, wdirrev, ) from mercurial.pycompat import open from mercurial import ( bookmarks, cmdutil, commands, copies, destutil, dirstateguard, error, extensions, logcmdutil, merge as mergemod, mergestate as mergestatemod, mergeutil, obsolete, obsutil, patch, phases, pycompat, registrar, repair, revset, revsetlang, rewriteutil, scmutil, smartset, state as statemod, util, ) # The following constants are used throughout the rebase module. The ordering of # their values must be maintained. # Indicates that a revision needs to be rebased revtodo = -1 revtodostr = b'-1' # legacy revstates no longer needed in current code # -2: nullmerge, -3: revignored, -4: revprecursor, -5: revpruned legacystates = {b'-2', b'-3', b'-4', b'-5'} cmdtable = {} command = registrar.command(cmdtable) configtable = {} configitem = registrar.configitem(configtable) configitem( b'devel', b'rebase.force-in-memory-merge', default=False, ) # Note for extension authors: ONLY specify testedwith = 'ships-with-hg-core' for # extensions which SHIP WITH MERCURIAL. Non-mainline extensions should # be specifying the version(s) of Mercurial they are tested with, or # leave the attribute unspecified. testedwith = b'ships-with-hg-core' def _nothingtorebase(): return 1 def _savegraft(ctx, extra): s = ctx.extra().get(b'source', None) if s is not None: extra[b'source'] = s s = ctx.extra().get(b'intermediate-source', None) if s is not None: extra[b'intermediate-source'] = s def _savebranch(ctx, extra): extra[b'branch'] = ctx.branch() def _destrebase(repo, sourceset, destspace=None): """small wrapper around destmerge to pass the right extra args Please wrap destutil.destmerge instead.""" return destutil.destmerge( repo, action=b'rebase', sourceset=sourceset, onheadcheck=False, destspace=destspace, ) revsetpredicate = registrar.revsetpredicate() @revsetpredicate(b'_destrebase') def _revsetdestrebase(repo, subset, x): # ``_rebasedefaultdest()`` # default destination for rebase. # # XXX: Currently private because I expect the signature to change. # # XXX: - bailing out in case of ambiguity vs returning all data. # i18n: "_rebasedefaultdest" is a keyword sourceset = None if x is not None: sourceset = revset.getset(repo, smartset.fullreposet(repo), x) return subset & smartset.baseset([_destrebase(repo, sourceset)]) @revsetpredicate(b'_destautoorphanrebase') def _revsetdestautoorphanrebase(repo, subset, x): # ``_destautoorphanrebase()`` # automatic rebase destination for a single orphan revision. unfi = repo.unfiltered() obsoleted = unfi.revs(b'obsolete()') src = revset.getset(repo, subset, x).first() # Empty src or already obsoleted - Do not return a destination if not src or src in obsoleted: return smartset.baseset() dests = destutil.orphanpossibledestination(repo, src) if len(dests) > 1: raise error.StateError( _(b"ambiguous automatic rebase: %r could end up on any of %r") % (src, dests) ) # We have zero or one destination, so we can just return here. return smartset.baseset(dests) def _ctxdesc(ctx): """short description for a context""" return cmdutil.format_changeset_summary( ctx.repo().ui, ctx, command=b'rebase' ) class rebaseruntime(object): """This class is a container for rebase runtime state""" def __init__(self, repo, ui, inmemory=False, dryrun=False, opts=None): if opts is None: opts = {} # prepared: whether we have rebasestate prepared or not. Currently it # decides whether "self.repo" is unfiltered or not. # The rebasestate has explicit hash to hash instructions not depending # on visibility. If rebasestate exists (in-memory or on-disk), use # unfiltered repo to avoid visibility issues. # Before knowing rebasestate (i.e. when starting a new rebase (not # --continue or --abort)), the original repo should be used so # visibility-dependent revsets are correct. self.prepared = False self.resume = False self._repo = repo self.ui = ui self.opts = opts self.originalwd = None self.external = nullrev # Mapping between the old revision id and either what is the new rebased # revision or what needs to be done with the old revision. The state # dict will be what contains most of the rebase progress state. self.state = {} self.activebookmark = None self.destmap = {} self.skipped = set() self.collapsef = opts.get('collapse', False) self.collapsemsg = cmdutil.logmessage(ui, pycompat.byteskwargs(opts)) self.date = opts.get('date', None) e = opts.get('extrafn') # internal, used by e.g. hgsubversion self.extrafns = [_savegraft] if e: self.extrafns = [e] self.backupf = ui.configbool(b'rewrite', b'backup-bundle') self.keepf = opts.get('keep', False) self.keepbranchesf = opts.get('keepbranches', False) self.skipemptysuccessorf = rewriteutil.skip_empty_successor( repo.ui, b'rebase' ) self.obsolete_with_successor_in_destination = {} self.obsolete_with_successor_in_rebase_set = set() self.inmemory = inmemory self.dryrun = dryrun self.stateobj = statemod.cmdstate(repo, b'rebasestate') @property def repo(self): if self.prepared: return self._repo.unfiltered() else: return self._repo def storestatus(self, tr=None): """Store the current status to allow recovery""" if tr: tr.addfilegenerator( b'rebasestate', (b'rebasestate',), self._writestatus, location=b'plain', ) else: with self.repo.vfs(b"rebasestate", b"w") as f: self._writestatus(f) def _writestatus(self, f): repo = self.repo assert repo.filtername is None f.write(repo[self.originalwd].hex() + b'\n') # was "dest". we now write dest per src root below. f.write(b'\n') f.write(repo[self.external].hex() + b'\n') f.write(b'%d\n' % int(self.collapsef)) f.write(b'%d\n' % int(self.keepf)) f.write(b'%d\n' % int(self.keepbranchesf)) f.write(b'%s\n' % (self.activebookmark or b'')) destmap = self.destmap for d, v in pycompat.iteritems(self.state): oldrev = repo[d].hex() if v >= 0: newrev = repo[v].hex() else: newrev = b"%d" % v destnode = repo[destmap[d]].hex() f.write(b"%s:%s:%s\n" % (oldrev, newrev, destnode)) repo.ui.debug(b'rebase status stored\n') def restorestatus(self): """Restore a previously stored status""" if not self.stateobj.exists(): cmdutil.wrongtooltocontinue(self.repo, _(b'rebase')) data = self._read() self.repo.ui.debug(b'rebase status resumed\n') self.originalwd = data[b'originalwd'] self.destmap = data[b'destmap'] self.state = data[b'state'] self.skipped = data[b'skipped'] self.collapsef = data[b'collapse'] self.keepf = data[b'keep'] self.keepbranchesf = data[b'keepbranches'] self.external = data[b'external'] self.activebookmark = data[b'activebookmark'] def _read(self): self.prepared = True repo = self.repo assert repo.filtername is None data = { b'keepbranches': None, b'collapse': None, b'activebookmark': None, b'external': nullrev, b'keep': None, b'originalwd': None, } legacydest = None state = {} destmap = {} if True: f = repo.vfs(b"rebasestate") for i, l in enumerate(f.read().splitlines()): if i == 0: data[b'originalwd'] = repo[l].rev() elif i == 1: # this line should be empty in newer version. but legacy # clients may still use it if l: legacydest = repo[l].rev() elif i == 2: data[b'external'] = repo[l].rev() elif i == 3: data[b'collapse'] = bool(int(l)) elif i == 4: data[b'keep'] = bool(int(l)) elif i == 5: data[b'keepbranches'] = bool(int(l)) elif i == 6 and not (len(l) == 81 and b':' in l): # line 6 is a recent addition, so for backwards # compatibility check that the line doesn't look like the # oldrev:newrev lines data[b'activebookmark'] = l else: args = l.split(b':') oldrev = repo[args[0]].rev() newrev = args[1] if newrev in legacystates: continue if len(args) > 2: destrev = repo[args[2]].rev() else: destrev = legacydest destmap[oldrev] = destrev if newrev == revtodostr: state[oldrev] = revtodo # Legacy compat special case else: state[oldrev] = repo[newrev].rev() if data[b'keepbranches'] is None: raise error.Abort(_(b'.hg/rebasestate is incomplete')) data[b'destmap'] = destmap data[b'state'] = state skipped = set() # recompute the set of skipped revs if not data[b'collapse']: seen = set(destmap.values()) for old, new in sorted(state.items()): if new != revtodo and new in seen: skipped.add(old) seen.add(new) data[b'skipped'] = skipped repo.ui.debug( b'computed skipped revs: %s\n' % (b' '.join(b'%d' % r for r in sorted(skipped)) or b'') ) return data def _handleskippingobsolete(self): """Compute structures necessary for skipping obsolete revisions""" if self.keepf: return if not self.ui.configbool(b'experimental', b'rebaseskipobsolete'): return obsoleteset = {r for r in self.state if self.repo[r].obsolete()} ( self.obsolete_with_successor_in_destination, self.obsolete_with_successor_in_rebase_set, ) = _compute_obsolete_sets(self.repo, obsoleteset, self.destmap) skippedset = set(self.obsolete_with_successor_in_destination) skippedset.update(self.obsolete_with_successor_in_rebase_set) _checkobsrebase(self.repo, self.ui, obsoleteset, skippedset) if obsolete.isenabled(self.repo, obsolete.allowdivergenceopt): self.obsolete_with_successor_in_rebase_set = set() else: for rev in self.repo.revs( b'descendants(%ld) and not %ld', self.obsolete_with_successor_in_rebase_set, self.obsolete_with_successor_in_rebase_set, ): self.state.pop(rev, None) self.destmap.pop(rev, None) def _prepareabortorcontinue( self, isabort, backup=True, suppwarns=False, dryrun=False, confirm=False ): self.resume = True try: self.restorestatus() # Calculate self.obsolete_* sets self._handleskippingobsolete() self.collapsemsg = restorecollapsemsg(self.repo, isabort) except error.RepoLookupError: if isabort: clearstatus(self.repo) clearcollapsemsg(self.repo) self.repo.ui.warn( _( b'rebase aborted (no revision is removed,' b' only broken state is cleared)\n' ) ) return 0 else: msg = _(b'cannot continue inconsistent rebase') hint = _(b'use "hg rebase --abort" to clear broken state') raise error.Abort(msg, hint=hint) if isabort: backup = backup and self.backupf return self._abort( backup=backup, suppwarns=suppwarns, dryrun=dryrun, confirm=confirm, ) def _preparenewrebase(self, destmap): if not destmap: return _nothingtorebase() result = buildstate(self.repo, destmap, self.collapsef) if not result: # Empty state built, nothing to rebase self.ui.status(_(b'nothing to rebase\n')) return _nothingtorebase() (self.originalwd, self.destmap, self.state) = result if self.collapsef: dests = set(self.destmap.values()) if len(dests) != 1: raise error.InputError( _(b'--collapse does not work with multiple destinations') ) destrev = next(iter(dests)) destancestors = self.repo.changelog.ancestors( [destrev], inclusive=True ) self.external = externalparent(self.repo, self.state, destancestors) for destrev in sorted(set(destmap.values())): dest = self.repo[destrev] if dest.closesbranch() and not self.keepbranchesf: self.ui.status(_(b'reopening closed branch head %s\n') % dest) # Calculate self.obsolete_* sets self._handleskippingobsolete() if not self.keepf: rebaseset = set(destmap.keys()) rebaseset -= set(self.obsolete_with_successor_in_destination) rebaseset -= self.obsolete_with_successor_in_rebase_set # We have our own divergence-checking in the rebase extension overrides = {} if obsolete.isenabled(self.repo, obsolete.createmarkersopt): overrides = { (b'experimental', b'evolution.allowdivergence'): b'true' } try: with self.ui.configoverride(overrides): rewriteutil.precheck(self.repo, rebaseset, action=b'rebase') except error.Abort as e: if e.hint is None: e.hint = _(b'use --keep to keep original changesets') raise e self.prepared = True def _assignworkingcopy(self): if self.inmemory: from mercurial.context import overlayworkingctx self.wctx = overlayworkingctx(self.repo) self.repo.ui.debug(b"rebasing in memory\n") else: self.wctx = self.repo[None] self.repo.ui.debug(b"rebasing on disk\n") self.repo.ui.log( b"rebase", b"using in-memory rebase: %r\n", self.inmemory, rebase_imm_used=self.inmemory, ) def _performrebase(self, tr): self._assignworkingcopy() repo, ui = self.repo, self.ui if self.keepbranchesf: # insert _savebranch at the start of extrafns so if # there's a user-provided extrafn it can clobber branch if # desired self.extrafns.insert(0, _savebranch) if self.collapsef: branches = set() for rev in self.state: branches.add(repo[rev].branch()) if len(branches) > 1: raise error.InputError( _(b'cannot collapse multiple named branches') ) # Keep track of the active bookmarks in order to reset them later self.activebookmark = self.activebookmark or repo._activebookmark if self.activebookmark: bookmarks.deactivate(repo) # Store the state before we begin so users can run 'hg rebase --abort' # if we fail before the transaction closes. self.storestatus() if tr: # When using single transaction, store state when transaction # commits. self.storestatus(tr) cands = [k for k, v in pycompat.iteritems(self.state) if v == revtodo] p = repo.ui.makeprogress( _(b"rebasing"), unit=_(b'changesets'), total=len(cands) ) def progress(ctx): p.increment(item=(b"%d:%s" % (ctx.rev(), ctx))) for subset in sortsource(self.destmap): sortedrevs = self.repo.revs(b'sort(%ld, -topo)', subset) for rev in sortedrevs: self._rebasenode(tr, rev, progress) p.complete() ui.note(_(b'rebase merging completed\n')) def _concludenode(self, rev, editor, commitmsg=None): """Commit the wd changes with parents p1 and p2. Reuse commit info from rev but also store useful information in extra. Return node of committed revision.""" repo = self.repo ctx = repo[rev] if commitmsg is None: commitmsg = ctx.description() # Skip replacement if collapsing, as that degenerates to p1 for all # nodes. if not self.collapsef: cl = repo.changelog commitmsg = rewriteutil.update_hash_refs( repo, commitmsg, { cl.node(oldrev): [cl.node(newrev)] for oldrev, newrev in self.state.items() if newrev != revtodo }, ) date = self.date if date is None: date = ctx.date() extra = {b'rebase_source': ctx.hex()} for c in self.extrafns: c(ctx, extra) destphase = max(ctx.phase(), phases.draft) overrides = { (b'phases', b'new-commit'): destphase, (b'ui', b'allowemptycommit'): not self.skipemptysuccessorf, } with repo.ui.configoverride(overrides, b'rebase'): if self.inmemory: newnode = commitmemorynode( repo, wctx=self.wctx, extra=extra, commitmsg=commitmsg, editor=editor, user=ctx.user(), date=date, ) else: newnode = commitnode( repo, extra=extra, commitmsg=commitmsg, editor=editor, user=ctx.user(), date=date, ) return newnode def _rebasenode(self, tr, rev, progressfn): repo, ui, opts = self.repo, self.ui, self.opts ctx = repo[rev] desc = _ctxdesc(ctx) if self.state[rev] == rev: ui.status(_(b'already rebased %s\n') % desc) elif rev in self.obsolete_with_successor_in_rebase_set: msg = ( _( b'note: not rebasing %s and its descendants as ' b'this would cause divergence\n' ) % desc ) repo.ui.status(msg) self.skipped.add(rev) elif rev in self.obsolete_with_successor_in_destination: succ = self.obsolete_with_successor_in_destination[rev] if succ is None: msg = _(b'note: not rebasing %s, it has no successor\n') % desc else: succdesc = _ctxdesc(repo[succ]) msg = _( b'note: not rebasing %s, already in destination as %s\n' ) % (desc, succdesc) repo.ui.status(msg) # Make clearrebased aware state[rev] is not a true successor self.skipped.add(rev) # Record rev as moved to its desired destination in self.state. # This helps bookmark and working parent movement. dest = max( adjustdest(repo, rev, self.destmap, self.state, self.skipped) ) self.state[rev] = dest elif self.state[rev] == revtodo: ui.status(_(b'rebasing %s\n') % desc) progressfn(ctx) p1, p2, base = defineparents( repo, rev, self.destmap, self.state, self.skipped, self.obsolete_with_successor_in_destination, ) if self.resume and self.wctx.p1().rev() == p1: repo.ui.debug(b'resuming interrupted rebase\n') self.resume = False else: overrides = {(b'ui', b'forcemerge'): opts.get('tool', b'')} with ui.configoverride(overrides, b'rebase'): try: rebasenode( repo, rev, p1, p2, base, self.collapsef, wctx=self.wctx, ) except error.InMemoryMergeConflictsError: if self.dryrun: raise error.ConflictResolutionRequired(b'rebase') if self.collapsef: # TODO: Make the overlayworkingctx reflected # in the working copy here instead of re-raising # so the entire rebase operation is retried. raise ui.status( _( b"hit merge conflicts; rebasing that " b"commit again in the working copy\n" ) ) try: cmdutil.bailifchanged(repo) except error.Abort: clearstatus(repo) clearcollapsemsg(repo) raise self.inmemory = False self._assignworkingcopy() mergemod.update(repo[p1], wc=self.wctx) rebasenode( repo, rev, p1, p2, base, self.collapsef, wctx=self.wctx, ) if not self.collapsef: merging = p2 != nullrev editform = cmdutil.mergeeditform(merging, b'rebase') editor = cmdutil.getcommiteditor(editform=editform, **opts) # We need to set parents again here just in case we're continuing # a rebase started with an old hg version (before 9c9cfecd4600), # because those old versions would have left us with two dirstate # parents, and we don't want to create a merge commit here (unless # we're rebasing a merge commit). self.wctx.setparents(repo[p1].node(), repo[p2].node()) newnode = self._concludenode(rev, editor) else: # Skip commit if we are collapsing newnode = None # Update the state if newnode is not None: self.state[rev] = repo[newnode].rev() ui.debug(b'rebased as %s\n' % short(newnode)) if repo[newnode].isempty(): ui.warn( _( b'note: created empty successor for %s, its ' b'destination already has all its changes\n' ) % desc ) else: if not self.collapsef: ui.warn( _( b'note: not rebasing %s, its destination already ' b'has all its changes\n' ) % desc ) self.skipped.add(rev) self.state[rev] = p1 ui.debug(b'next revision set to %d\n' % p1) else: ui.status( _(b'already rebased %s as %s\n') % (desc, repo[self.state[rev]]) ) if not tr: # When not using single transaction, store state after each # commit is completely done. On InterventionRequired, we thus # won't store the status. Instead, we'll hit the "len(parents) == 2" # case and realize that the commit was in progress. self.storestatus() def _finishrebase(self): repo, ui, opts = self.repo, self.ui, self.opts fm = ui.formatter(b'rebase', pycompat.byteskwargs(opts)) fm.startitem() if self.collapsef: p1, p2, _base = defineparents( repo, min(self.state), self.destmap, self.state, self.skipped, self.obsolete_with_successor_in_destination, ) editopt = opts.get('edit') editform = b'rebase.collapse' if self.collapsemsg: commitmsg = self.collapsemsg else: commitmsg = b'Collapsed revision' for rebased in sorted(self.state): if rebased not in self.skipped: commitmsg += b'\n* %s' % repo[rebased].description() editopt = True editor = cmdutil.getcommiteditor(edit=editopt, editform=editform) revtoreuse = max(self.state) self.wctx.setparents(repo[p1].node(), repo[self.external].node()) newnode = self._concludenode( revtoreuse, editor, commitmsg=commitmsg ) if newnode is not None: newrev = repo[newnode].rev() for oldrev in self.state: self.state[oldrev] = newrev if b'qtip' in repo.tags(): updatemq(repo, self.state, self.skipped, **opts) # restore original working directory # (we do this before stripping) newwd = self.state.get(self.originalwd, self.originalwd) if newwd < 0: # original directory is a parent of rebase set root or ignored newwd = self.originalwd if newwd not in [c.rev() for c in repo[None].parents()]: ui.note(_(b"update back to initial working directory parent\n")) mergemod.update(repo[newwd]) collapsedas = None if self.collapsef and not self.keepf: collapsedas = newnode clearrebased( ui, repo, self.destmap, self.state, self.skipped, collapsedas, self.keepf, fm=fm, backup=self.backupf, ) clearstatus(repo) clearcollapsemsg(repo) ui.note(_(b"rebase completed\n")) util.unlinkpath(repo.sjoin(b'undo'), ignoremissing=True) if self.skipped: skippedlen = len(self.skipped) ui.note(_(b"%d revisions have been skipped\n") % skippedlen) fm.end() if ( self.activebookmark and self.activebookmark in repo._bookmarks and repo[b'.'].node() == repo._bookmarks[self.activebookmark] ): bookmarks.activate(repo, self.activebookmark) def _abort(self, backup=True, suppwarns=False, dryrun=False, confirm=False): '''Restore the repository to its original state.''' repo = self.repo try: # If the first commits in the rebased set get skipped during the # rebase, their values within the state mapping will be the dest # rev id. The rebased list must must not contain the dest rev # (issue4896) rebased = [ s for r, s in self.state.items() if s >= 0 and s != r and s != self.destmap[r] ] immutable = [d for d in rebased if not repo[d].mutable()] cleanup = True if immutable: repo.ui.warn( _(b"warning: can't clean up public changesets %s\n") % b', '.join(bytes(repo[r]) for r in immutable), hint=_(b"see 'hg help phases' for details"), ) cleanup = False descendants = set() if rebased: descendants = set(repo.changelog.descendants(rebased)) if descendants - set(rebased): repo.ui.warn( _( b"warning: new changesets detected on " b"destination branch, can't strip\n" ) ) cleanup = False if cleanup: if rebased: strippoints = [ c.node() for c in repo.set(b'roots(%ld)', rebased) ] updateifonnodes = set(rebased) updateifonnodes.update(self.destmap.values()) if not dryrun and not confirm: updateifonnodes.add(self.originalwd) shouldupdate = repo[b'.'].rev() in updateifonnodes # Update away from the rebase if necessary if shouldupdate: mergemod.clean_update(repo[self.originalwd]) # Strip from the first rebased revision if rebased: repair.strip(repo.ui, repo, strippoints, backup=backup) if self.activebookmark and self.activebookmark in repo._bookmarks: bookmarks.activate(repo, self.activebookmark) finally: clearstatus(repo) clearcollapsemsg(repo) if not suppwarns: repo.ui.warn(_(b'rebase aborted\n')) return 0 @command( b'rebase', [ ( b's', b'source', [], _(b'rebase the specified changesets and their descendants'), _(b'REV'), ), ( b'b', b'base', [], _(b'rebase everything from branching point of specified changeset'), _(b'REV'), ), (b'r', b'rev', [], _(b'rebase these revisions'), _(b'REV')), ( b'd', b'dest', b'', _(b'rebase onto the specified changeset'), _(b'REV'), ), (b'', b'collapse', False, _(b'collapse the rebased changesets')), ( b'm', b'message', b'', _(b'use text as collapse commit message'), _(b'TEXT'), ), (b'e', b'edit', False, _(b'invoke editor on commit messages')), ( b'l', b'logfile', b'', _(b'read collapse commit message from file'), _(b'FILE'), ), (b'k', b'keep', False, _(b'keep original changesets')), (b'', b'keepbranches', False, _(b'keep original branch names')), (b'D', b'detach', False, _(b'(DEPRECATED)')), (b'i', b'interactive', False, _(b'(DEPRECATED)')), (b't', b'tool', b'', _(b'specify merge tool')), (b'', b'stop', False, _(b'stop interrupted rebase')), (b'c', b'continue', False, _(b'continue an interrupted rebase')), (b'a', b'abort', False, _(b'abort an interrupted rebase')), ( b'', b'auto-orphans', b'', _( b'automatically rebase orphan revisions ' b'in the specified revset (EXPERIMENTAL)' ), ), ] + cmdutil.dryrunopts + cmdutil.formatteropts + cmdutil.confirmopts, _(b'[[-s REV]... | [-b REV]... | [-r REV]...] [-d REV] [OPTION]...'), helpcategory=command.CATEGORY_CHANGE_MANAGEMENT, ) def rebase(ui, repo, **opts): """move changeset (and descendants) to a different branch Rebase uses repeated merging to graft changesets from one part of history (the source) onto another (the destination). This can be useful for linearizing *local* changes relative to a master development tree. Published commits cannot be rebased (see :hg:`help phases`). To copy commits, see :hg:`help graft`. If you don't specify a destination changeset (``-d/--dest``), rebase will use the same logic as :hg:`merge` to pick a destination. if the current branch contains exactly one other head, the other head is merged with by default. Otherwise, an explicit revision with which to merge with must be provided. (destination changeset is not modified by rebasing, but new changesets are added as its descendants.) Here are the ways to select changesets: 1. Explicitly select them using ``--rev``. 2. Use ``--source`` to select a root changeset and include all of its descendants. 3. Use ``--base`` to select a changeset; rebase will find ancestors and their descendants which are not also ancestors of the destination. 4. If you do not specify any of ``--rev``, ``--source``, or ``--base``, rebase will use ``--base .`` as above. If ``--source`` or ``--rev`` is used, special names ``SRC`` and ``ALLSRC`` can be used in ``--dest``. Destination would be calculated per source revision with ``SRC`` substituted by that single source revision and ``ALLSRC`` substituted by all source revisions. Rebase will destroy original changesets unless you use ``--keep``. It will also move your bookmarks (even if you do). Some changesets may be dropped if they do not contribute changes (e.g. merges from the destination branch). Unlike ``merge``, rebase will do nothing if you are at the branch tip of a named branch with two heads. You will need to explicitly specify source and/or destination. If you need to use a tool to automate merge/conflict decisions, you can specify one with ``--tool``, see :hg:`help merge-tools`. As a caveat: the tool will not be used to mediate when a file was deleted, there is no hook presently available for this. If a rebase is interrupted to manually resolve a conflict, it can be continued with --continue/-c, aborted with --abort/-a, or stopped with --stop. .. container:: verbose Examples: - move "local changes" (current commit back to branching point) to the current branch tip after a pull:: hg rebase - move a single changeset to the stable branch:: hg rebase -r 5f493448 -d stable - splice a commit and all its descendants onto another part of history:: hg rebase --source c0c3 --dest 4cf9 - rebase everything on a branch marked by a bookmark onto the default branch:: hg rebase --base myfeature --dest default - collapse a sequence of changes into a single commit:: hg rebase --collapse -r 1520:1525 -d . - move a named branch while preserving its name:: hg rebase -r "branch(featureX)" -d 1.3 --keepbranches - stabilize orphaned changesets so history looks linear:: hg rebase -r 'orphan()-obsolete()'\ -d 'first(max((successors(max(roots(ALLSRC) & ::SRC)^)-obsolete())::) +\ max(::((roots(ALLSRC) & ::SRC)^)-obsolete()))' Configuration Options: You can make rebase require a destination if you set the following config option:: [commands] rebase.requiredest = True By default, rebase will close the transaction after each commit. For performance purposes, you can configure rebase to use a single transaction across the entire rebase. WARNING: This setting introduces a significant risk of losing the work you've done in a rebase if the rebase aborts unexpectedly:: [rebase] singletransaction = True By default, rebase writes to the working copy, but you can configure it to run in-memory for better performance. When the rebase is not moving the parent(s) of the working copy (AKA the "currently checked out changesets"), this may also allow it to run even if the working copy is dirty:: [rebase] experimental.inmemory = True Return Values: Returns 0 on success, 1 if nothing to rebase or there are unresolved conflicts. """ inmemory = ui.configbool(b'rebase', b'experimental.inmemory') action = cmdutil.check_at_most_one_arg(opts, 'abort', 'stop', 'continue') if action: cmdutil.check_incompatible_arguments( opts, action, ['confirm', 'dry_run'] ) cmdutil.check_incompatible_arguments( opts, action, ['rev', 'source', 'base', 'dest'] ) cmdutil.check_at_most_one_arg(opts, 'confirm', 'dry_run') cmdutil.check_at_most_one_arg(opts, 'rev', 'source', 'base') if action or repo.currenttransaction() is not None: # in-memory rebase is not compatible with resuming rebases. # (Or if it is run within a transaction, since the restart logic can # fail the entire transaction.) inmemory = False if opts.get('auto_orphans'): disallowed_opts = set(opts) - {'auto_orphans'} cmdutil.check_incompatible_arguments( opts, 'auto_orphans', disallowed_opts ) userrevs = list(repo.revs(opts.get('auto_orphans'))) opts['rev'] = [revsetlang.formatspec(b'%ld and orphan()', userrevs)] opts['dest'] = b'_destautoorphanrebase(SRC)' if opts.get('dry_run') or opts.get('confirm'): return _dryrunrebase(ui, repo, action, opts) elif action == 'stop': rbsrt = rebaseruntime(repo, ui) with repo.wlock(), repo.lock(): rbsrt.restorestatus() if rbsrt.collapsef: raise error.StateError(_(b"cannot stop in --collapse session")) allowunstable = obsolete.isenabled(repo, obsolete.allowunstableopt) if not (rbsrt.keepf or allowunstable): raise error.StateError( _( b"cannot remove original changesets with" b" unrebased descendants" ), hint=_( b'either enable obsmarkers to allow unstable ' b'revisions or use --keep to keep original ' b'changesets' ), ) # update to the current working revision # to clear interrupted merge mergemod.clean_update(repo[rbsrt.originalwd]) rbsrt._finishrebase() return 0 elif inmemory: try: # in-memory merge doesn't support conflicts, so if we hit any, abort # and re-run as an on-disk merge. overrides = {(b'rebase', b'singletransaction'): True} with ui.configoverride(overrides, b'rebase'): return _dorebase(ui, repo, action, opts, inmemory=inmemory) except error.InMemoryMergeConflictsError: if ui.configbool(b'devel', b'rebase.force-in-memory-merge'): raise ui.warn( _( b'hit merge conflicts; re-running rebase without in-memory' b' merge\n' ) ) clearstatus(repo) clearcollapsemsg(repo) return _dorebase(ui, repo, action, opts, inmemory=False) else: return _dorebase(ui, repo, action, opts) def _dryrunrebase(ui, repo, action, opts): rbsrt = rebaseruntime(repo, ui, inmemory=True, dryrun=True, opts=opts) confirm = opts.get('confirm') if confirm: ui.status(_(b'starting in-memory rebase\n')) else: ui.status( _(b'starting dry-run rebase; repository will not be changed\n') ) with repo.wlock(), repo.lock(): needsabort = True try: overrides = {(b'rebase', b'singletransaction'): True} with ui.configoverride(overrides, b'rebase'): res = _origrebase( ui, repo, action, opts, rbsrt, ) if res == _nothingtorebase(): needsabort = False return res except error.ConflictResolutionRequired: ui.status(_(b'hit a merge conflict\n')) return 1 except error.Abort: needsabort = False raise else: if confirm: ui.status(_(b'rebase completed successfully\n')) if not ui.promptchoice(_(b'apply changes (yn)?$$ &Yes $$ &No')): # finish unfinished rebase rbsrt._finishrebase() else: rbsrt._prepareabortorcontinue( isabort=True, backup=False, suppwarns=True, confirm=confirm, ) needsabort = False else: ui.status( _( b'dry-run rebase completed successfully; run without' b' -n/--dry-run to perform this rebase\n' ) ) return 0 finally: if needsabort: # no need to store backup in case of dryrun rbsrt._prepareabortorcontinue( isabort=True, backup=False, suppwarns=True, dryrun=opts.get('dry_run'), ) def _dorebase(ui, repo, action, opts, inmemory=False): rbsrt = rebaseruntime(repo, ui, inmemory, opts=opts) return _origrebase(ui, repo, action, opts, rbsrt) def _origrebase(ui, repo, action, opts, rbsrt): assert action != 'stop' with repo.wlock(), repo.lock(): if opts.get('interactive'): try: if extensions.find(b'histedit'): enablehistedit = b'' except KeyError: enablehistedit = b" --config extensions.histedit=" help = b"hg%s help -e histedit" % enablehistedit msg = ( _( b"interactive history editing is supported by the " b"'histedit' extension (see \"%s\")" ) % help ) raise error.InputError(msg) if rbsrt.collapsemsg and not rbsrt.collapsef: raise error.InputError( _(b'message can only be specified with collapse') ) if action: if rbsrt.collapsef: raise error.InputError( _(b'cannot use collapse with continue or abort') ) if action == 'abort' and opts.get('tool', False): ui.warn(_(b'tool option will be ignored\n')) if action == 'continue': ms = mergestatemod.mergestate.read(repo) mergeutil.checkunresolved(ms) retcode = rbsrt._prepareabortorcontinue(isabort=(action == 'abort')) if retcode is not None: return retcode else: # search default destination in this space # used in the 'hg pull --rebase' case, see issue 5214. destspace = opts.get('_destspace') destmap = _definedestmap( ui, repo, rbsrt.inmemory, opts.get('dest', None), opts.get('source', []), opts.get('base', []), opts.get('rev', []), destspace=destspace, ) retcode = rbsrt._preparenewrebase(destmap) if retcode is not None: return retcode storecollapsemsg(repo, rbsrt.collapsemsg) tr = None singletr = ui.configbool(b'rebase', b'singletransaction') if singletr: tr = repo.transaction(b'rebase') # If `rebase.singletransaction` is enabled, wrap the entire operation in # one transaction here. Otherwise, transactions are obtained when # committing each node, which is slower but allows partial success. with util.acceptintervention(tr): # Same logic for the dirstate guard, except we don't create one when # rebasing in-memory (it's not needed). dsguard = None if singletr and not rbsrt.inmemory: dsguard = dirstateguard.dirstateguard(repo, b'rebase') with util.acceptintervention(dsguard): rbsrt._performrebase(tr) if not rbsrt.dryrun: rbsrt._finishrebase() def _definedestmap(ui, repo, inmemory, destf, srcf, basef, revf, destspace): """use revisions argument to define destmap {srcrev: destrev}""" if revf is None: revf = [] # destspace is here to work around issues with `hg pull --rebase` see # issue5214 for details cmdutil.checkunfinished(repo) if not inmemory: cmdutil.bailifchanged(repo) if ui.configbool(b'commands', b'rebase.requiredest') and not destf: raise error.InputError( _(b'you must specify a destination'), hint=_(b'use: hg rebase -d REV'), ) dest = None if revf: rebaseset = logcmdutil.revrange(repo, revf) if not rebaseset: ui.status(_(b'empty "rev" revision set - nothing to rebase\n')) return None elif srcf: src = logcmdutil.revrange(repo, srcf) if not src: ui.status(_(b'empty "source" revision set - nothing to rebase\n')) return None # `+ (%ld)` to work around `wdir()::` being empty rebaseset = repo.revs(b'(%ld):: + (%ld)', src, src) else: base = logcmdutil.revrange(repo, basef or [b'.']) if not base: ui.status( _(b'empty "base" revision set - ' b"can't compute rebase set\n") ) return None if destf: # --base does not support multiple destinations dest = logcmdutil.revsingle(repo, destf) else: dest = repo[_destrebase(repo, base, destspace=destspace)] destf = bytes(dest) roots = [] # selected children of branching points bpbase = {} # {branchingpoint: [origbase]} for b in base: # group bases by branching points bp = repo.revs(b'ancestor(%d, %d)', b, dest.rev()).first() bpbase[bp] = bpbase.get(bp, []) + [b] if None in bpbase: # emulate the old behavior, showing "nothing to rebase" (a better # behavior may be abort with "cannot find branching point" error) bpbase.clear() for bp, bs in pycompat.iteritems(bpbase): # calculate roots roots += list(repo.revs(b'children(%d) & ancestors(%ld)', bp, bs)) rebaseset = repo.revs(b'%ld::', roots) if not rebaseset: # transform to list because smartsets are not comparable to # lists. This should be improved to honor laziness of # smartset. if list(base) == [dest.rev()]: if basef: ui.status( _( b'nothing to rebase - %s is both "base"' b' and destination\n' ) % dest ) else: ui.status( _( b'nothing to rebase - working directory ' b'parent is also destination\n' ) ) elif not repo.revs(b'%ld - ::%d', base, dest.rev()): if basef: ui.status( _( b'nothing to rebase - "base" %s is ' b'already an ancestor of destination ' b'%s\n' ) % (b'+'.join(bytes(repo[r]) for r in base), dest) ) else: ui.status( _( b'nothing to rebase - working ' b'directory parent is already an ' b'ancestor of destination %s\n' ) % dest ) else: # can it happen? ui.status( _(b'nothing to rebase from %s to %s\n') % (b'+'.join(bytes(repo[r]) for r in base), dest) ) return None if wdirrev in rebaseset: raise error.InputError(_(b'cannot rebase the working copy')) rebasingwcp = repo[b'.'].rev() in rebaseset ui.log( b"rebase", b"rebasing working copy parent: %r\n", rebasingwcp, rebase_rebasing_wcp=rebasingwcp, ) if inmemory and rebasingwcp: # Check these since we did not before. cmdutil.checkunfinished(repo) cmdutil.bailifchanged(repo) if not destf: dest = repo[_destrebase(repo, rebaseset, destspace=destspace)] destf = bytes(dest) allsrc = revsetlang.formatspec(b'%ld', rebaseset) alias = {b'ALLSRC': allsrc} if dest is None: try: # fast path: try to resolve dest without SRC alias dest = scmutil.revsingle(repo, destf, localalias=alias) except error.RepoLookupError: # multi-dest path: resolve dest for each SRC separately destmap = {} for r in rebaseset: alias[b'SRC'] = revsetlang.formatspec(b'%d', r) # use repo.anyrevs instead of scmutil.revsingle because we # don't want to abort if destset is empty. destset = repo.anyrevs([destf], user=True, localalias=alias) size = len(destset) if size == 1: destmap[r] = destset.first() elif size == 0: ui.note(_(b'skipping %s - empty destination\n') % repo[r]) else: raise error.InputError( _(b'rebase destination for %s is not unique') % repo[r] ) if dest is not None: # single-dest case: assign dest to each rev in rebaseset destrev = dest.rev() destmap = {r: destrev for r in rebaseset} # {srcrev: destrev} if not destmap: ui.status(_(b'nothing to rebase - empty destination\n')) return None return destmap def externalparent(repo, state, destancestors): """Return the revision that should be used as the second parent when the revisions in state is collapsed on top of destancestors. Abort if there is more than one parent. """ parents = set() source = min(state) for rev in state: if rev == source: continue for p in repo[rev].parents(): if p.rev() not in state and p.rev() not in destancestors: parents.add(p.rev()) if not parents: return nullrev if len(parents) == 1: return parents.pop() raise error.StateError( _( b'unable to collapse on top of %d, there is more ' b'than one external parent: %s' ) % (max(destancestors), b', '.join(b"%d" % p for p in sorted(parents))) ) def commitmemorynode(repo, wctx, editor, extra, user, date, commitmsg): """Commit the memory changes with parents p1 and p2. Return node of committed revision.""" # By convention, ``extra['branch']`` (set by extrafn) clobbers # ``branch`` (used when passing ``--keepbranches``). branch = None if b'branch' in extra: branch = extra[b'branch'] # FIXME: We call _compact() because it's required to correctly detect # changed files. This was added to fix a regression shortly before the 5.5 # release. A proper fix will be done in the default branch. wctx._compact() memctx = wctx.tomemctx( commitmsg, date=date, extra=extra, user=user, branch=branch, editor=editor, ) if memctx.isempty() and not repo.ui.configbool(b'ui', b'allowemptycommit'): return None commitres = repo.commitctx(memctx) wctx.clean() # Might be reused return commitres def commitnode(repo, editor, extra, user, date, commitmsg): """Commit the wd changes with parents p1 and p2. Return node of committed revision.""" dsguard = util.nullcontextmanager() if not repo.ui.configbool(b'rebase', b'singletransaction'): dsguard = dirstateguard.dirstateguard(repo, b'rebase') with dsguard: # Commit might fail if unresolved files exist newnode = repo.commit( text=commitmsg, user=user, date=date, extra=extra, editor=editor ) repo.dirstate.setbranch(repo[newnode].branch()) return newnode def rebasenode(repo, rev, p1, p2, base, collapse, wctx): """Rebase a single revision rev on top of p1 using base as merge ancestor""" # Merge phase # Update to destination and merge it with local p1ctx = repo[p1] if wctx.isinmemory(): wctx.setbase(p1ctx) else: if repo[b'.'].rev() != p1: repo.ui.debug(b" update to %d:%s\n" % (p1, p1ctx)) mergemod.clean_update(p1ctx) else: repo.ui.debug(b" already in destination\n") # This is, alas, necessary to invalidate workingctx's manifest cache, # as well as other data we litter on it in other places. wctx = repo[None] repo.dirstate.write(repo.currenttransaction()) ctx = repo[rev] repo.ui.debug(b" merge against %d:%s\n" % (rev, ctx)) if base is not None: repo.ui.debug(b" detach base %d:%s\n" % (base, repo[base])) # See explanation in merge.graft() mergeancestor = repo.changelog.isancestor(p1ctx.node(), ctx.node()) stats = mergemod._update( repo, rev, branchmerge=True, force=True, ancestor=base, mergeancestor=mergeancestor, labels=[b'dest', b'source', b'parent of source'], wc=wctx, ) wctx.setparents(p1ctx.node(), repo[p2].node()) if collapse: copies.graftcopies(wctx, ctx, p1ctx) else: # If we're not using --collapse, we need to # duplicate copies between the revision we're # rebasing and its first parent. copies.graftcopies(wctx, ctx, ctx.p1()) if stats.unresolvedcount > 0: if wctx.isinmemory(): raise error.InMemoryMergeConflictsError() else: raise error.ConflictResolutionRequired(b'rebase') def adjustdest(repo, rev, destmap, state, skipped): r"""adjust rebase destination given the current rebase state rev is what is being rebased. Return a list of two revs, which are the adjusted destinations for rev's p1 and p2, respectively. If a parent is nullrev, return dest without adjustment for it. For example, when doing rebasing B+E to F, C to G, rebase will first move B to B1, and E's destination will be adjusted from F to B1. B1 <- written during rebasing B | F <- original destination of B, E | | E <- rev, which is being rebased | | | D <- prev, one parent of rev being checked | | | x <- skipped, ex. no successor or successor in (::dest) | | | C <- rebased as C', different destination | | | B <- rebased as B1 C' |/ | A G <- destination of C, different Another example about merge changeset, rebase -r C+G+H -d K, rebase will first move C to C1, G to G1, and when it's checking H, the adjusted destinations will be [C1, G1]. H C1 G1 /| | / F G |/ K | | -> K | C D | | |/ | | B | ... |/ |/ A A Besides, adjust dest according to existing rebase information. For example, B C D B needs to be rebased on top of C, C needs to be rebased on top \|/ of D. We will rebase C first. A C' After rebasing C, when considering B's destination, use C' | instead of the original C. B D \ / A """ # pick already rebased revs with same dest from state as interesting source dest = destmap[rev] source = [ s for s, d in state.items() if d > 0 and destmap[s] == dest and s not in skipped ] result = [] for prev in repo.changelog.parentrevs(rev): adjusted = dest if prev != nullrev: candidate = repo.revs(b'max(%ld and (::%d))', source, prev).first() if candidate is not None: adjusted = state[candidate] if adjusted == dest and dest in state: adjusted = state[dest] if adjusted == revtodo: # sortsource should produce an order that makes this impossible raise error.ProgrammingError( b'rev %d should be rebased already at this time' % dest ) result.append(adjusted) return result def _checkobsrebase(repo, ui, rebaseobsrevs, rebaseobsskipped): """ Abort if rebase will create divergence or rebase is noop because of markers `rebaseobsrevs`: set of obsolete revision in source `rebaseobsskipped`: set of revisions from source skipped because they have successors in destination or no non-obsolete successor. """ # Obsolete node with successors not in dest leads to divergence divergenceok = obsolete.isenabled(repo, obsolete.allowdivergenceopt) divergencebasecandidates = rebaseobsrevs - rebaseobsskipped if divergencebasecandidates and not divergenceok: divhashes = (bytes(repo[r]) for r in divergencebasecandidates) msg = _(b"this rebase will cause divergences from: %s") h = _( b"to force the rebase please set " b"experimental.evolution.allowdivergence=True" ) raise error.StateError(msg % (b",".join(divhashes),), hint=h) def successorrevs(unfi, rev): """yield revision numbers for successors of rev""" assert unfi.filtername is None get_rev = unfi.changelog.index.get_rev for s in obsutil.allsuccessors(unfi.obsstore, [unfi[rev].node()]): r = get_rev(s) if r is not None: yield r def defineparents(repo, rev, destmap, state, skipped, obsskipped): """Return new parents and optionally a merge base for rev being rebased The destination specified by "dest" cannot always be used directly because previously rebase result could affect destination. For example, D E rebase -r C+D+E -d B |/ C will be rebased to C' B C D's new destination will be C' instead of B |/ E's new destination will be C' instead of B A The new parents of a merge is slightly more complicated. See the comment block below. """ # use unfiltered changelog since successorrevs may return filtered nodes assert repo.filtername is None cl = repo.changelog isancestor = cl.isancestorrev dest = destmap[rev] oldps = repo.changelog.parentrevs(rev) # old parents newps = [nullrev, nullrev] # new parents dests = adjustdest(repo, rev, destmap, state, skipped) bases = list(oldps) # merge base candidates, initially just old parents if all(r == nullrev for r in oldps[1:]): # For non-merge changeset, just move p to adjusted dest as requested. newps[0] = dests[0] else: # For merge changeset, if we move p to dests[i] unconditionally, both # parents may change and the end result looks like "the merge loses a # parent", which is a surprise. This is a limit because "--dest" only # accepts one dest per src. # # Therefore, only move p with reasonable conditions (in this order): # 1. use dest, if dest is a descendent of (p or one of p's successors) # 2. use p's rebased result, if p is rebased (state[p] > 0) # # Comparing with adjustdest, the logic here does some additional work: # 1. decide which parents will not be moved towards dest # 2. if the above decision is "no", should a parent still be moved # because it was rebased? # # For example: # # C # "rebase -r C -d D" is an error since none of the parents # /| # can be moved. "rebase -r B+C -d D" will move C's parent # A B D # B (using rule "2."), since B will be rebased. # # The loop tries to be not rely on the fact that a Mercurial node has # at most 2 parents. for i, p in enumerate(oldps): np = p # new parent if any(isancestor(x, dests[i]) for x in successorrevs(repo, p)): np = dests[i] elif p in state and state[p] > 0: np = state[p] # If one parent becomes an ancestor of the other, drop the ancestor for j, x in enumerate(newps[:i]): if x == nullrev: continue if isancestor(np, x): # CASE-1 np = nullrev elif isancestor(x, np): # CASE-2 newps[j] = np np = nullrev # New parents forming an ancestor relationship does not # mean the old parents have a similar relationship. Do not # set bases[x] to nullrev. bases[j], bases[i] = bases[i], bases[j] newps[i] = np # "rebasenode" updates to new p1, and the old p1 will be used as merge # base. If only p2 changes, merging using unchanged p1 as merge base is # suboptimal. Therefore swap parents to make the merge sane. if newps[1] != nullrev and oldps[0] == newps[0]: assert len(newps) == 2 and len(oldps) == 2 newps.reverse() bases.reverse() # No parent change might be an error because we fail to make rev a # descendent of requested dest. This can happen, for example: # # C # rebase -r C -d D # /| # None of A and B will be changed to D and rebase fails. # A B D if set(newps) == set(oldps) and dest not in newps: raise error.InputError( _( b'cannot rebase %d:%s without ' b'moving at least one of its parents' ) % (rev, repo[rev]) ) # Source should not be ancestor of dest. The check here guarantees it's # impossible. With multi-dest, the initial check does not cover complex # cases since we don't have abstractions to dry-run rebase cheaply. if any(p != nullrev and isancestor(rev, p) for p in newps): raise error.InputError(_(b'source is ancestor of destination')) # Check if the merge will contain unwanted changes. That may happen if # there are multiple special (non-changelog ancestor) merge bases, which # cannot be handled well by the 3-way merge algorithm. For example: # # F # /| # D E # "rebase -r D+E+F -d Z", when rebasing F, if "D" was chosen # | | # as merge base, the difference between D and F will include # B C # C, so the rebased F will contain C surprisingly. If "E" was # |/ # chosen, the rebased F will contain B. # A Z # # But our merge base candidates (D and E in above case) could still be # better than the default (ancestor(F, Z) == null). Therefore still # pick one (so choose p1 above). if sum(1 for b in set(bases) if b != nullrev and b not in newps) > 1: unwanted = [None, None] # unwanted[i]: unwanted revs if choose bases[i] for i, base in enumerate(bases): if base == nullrev or base in newps: continue # Revisions in the side (not chosen as merge base) branch that # might contain "surprising" contents other_bases = set(bases) - {base} siderevs = list( repo.revs(b'(%ld %% (%d+%d))', other_bases, base, dest) ) # If those revisions are covered by rebaseset, the result is good. # A merge in rebaseset would be considered to cover its ancestors. if siderevs: rebaseset = [ r for r, d in state.items() if d > 0 and r not in obsskipped ] merges = [ r for r in rebaseset if cl.parentrevs(r)[1] != nullrev ] unwanted[i] = list( repo.revs( b'%ld - (::%ld) - %ld', siderevs, merges, rebaseset ) ) if any(revs is not None for revs in unwanted): # Choose a merge base that has a minimal number of unwanted revs. l, i = min( (len(revs), i) for i, revs in enumerate(unwanted) if revs is not None ) # The merge will include unwanted revisions. Abort now. Revisit this if # we have a more advanced merge algorithm that handles multiple bases. if l > 0: unwanteddesc = _(b' or ').join( ( b', '.join(b'%d:%s' % (r, repo[r]) for r in revs) for revs in unwanted if revs is not None ) ) raise error.InputError( _(b'rebasing %d:%s will include unwanted changes from %s') % (rev, repo[rev], unwanteddesc) ) # newps[0] should match merge base if possible. Currently, if newps[i] # is nullrev, the only case is newps[i] and newps[j] (j < i), one is # the other's ancestor. In that case, it's fine to not swap newps here. # (see CASE-1 and CASE-2 above) if i != 0: if newps[i] != nullrev: newps[0], newps[i] = newps[i], newps[0] bases[0], bases[i] = bases[i], bases[0] # "rebasenode" updates to new p1, use the corresponding merge base. base = bases[0] repo.ui.debug(b" future parents are %d and %d\n" % tuple(newps)) return newps[0], newps[1], base def isagitpatch(repo, patchname): """Return true if the given patch is in git format""" mqpatch = os.path.join(repo.mq.path, patchname) for line in patch.linereader(open(mqpatch, b'rb')): if line.startswith(b'diff --git'): return True return False def updatemq(repo, state, skipped, **opts): """Update rebased mq patches - finalize and then import them""" mqrebase = {} mq = repo.mq original_series = mq.fullseries[:] skippedpatches = set() for p in mq.applied: rev = repo[p.node].rev() if rev in state: repo.ui.debug( b'revision %d is an mq patch (%s), finalize it.\n' % (rev, p.name) ) mqrebase[rev] = (p.name, isagitpatch(repo, p.name)) else: # Applied but not rebased, not sure this should happen skippedpatches.add(p.name) if mqrebase: mq.finish(repo, mqrebase.keys()) # We must start import from the newest revision for rev in sorted(mqrebase, reverse=True): if rev not in skipped: name, isgit = mqrebase[rev] repo.ui.note( _(b'updating mq patch %s to %d:%s\n') % (name, state[rev], repo[state[rev]]) ) mq.qimport( repo, (), patchname=name, git=isgit, rev=[b"%d" % state[rev]], ) else: # Rebased and skipped skippedpatches.add(mqrebase[rev][0]) # Patches were either applied and rebased and imported in # order, applied and removed or unapplied. Discard the removed # ones while preserving the original series order and guards. newseries = [ s for s in original_series if mq.guard_re.split(s, 1)[0] not in skippedpatches ] mq.fullseries[:] = newseries mq.seriesdirty = True mq.savedirty() def storecollapsemsg(repo, collapsemsg): """Store the collapse message to allow recovery""" collapsemsg = collapsemsg or b'' f = repo.vfs(b"last-message.txt", b"w") f.write(b"%s\n" % collapsemsg) f.close() def clearcollapsemsg(repo): """Remove collapse message file""" repo.vfs.unlinkpath(b"last-message.txt", ignoremissing=True) def restorecollapsemsg(repo, isabort): """Restore previously stored collapse message""" try: f = repo.vfs(b"last-message.txt") collapsemsg = f.readline().strip() f.close() except IOError as err: if err.errno != errno.ENOENT: raise if isabort: # Oh well, just abort like normal collapsemsg = b'' else: raise error.Abort(_(b'missing .hg/last-message.txt for rebase')) return collapsemsg def clearstatus(repo): """Remove the status files""" # Make sure the active transaction won't write the state file tr = repo.currenttransaction() if tr: tr.removefilegenerator(b'rebasestate') repo.vfs.unlinkpath(b"rebasestate", ignoremissing=True) def sortsource(destmap): """yield source revisions in an order that we only rebase things once If source and destination overlaps, we should filter out revisions depending on other revisions which hasn't been rebased yet. Yield a sorted list of revisions each time. For example, when rebasing A to B, B to C. This function yields [B], then [A], indicating B needs to be rebased first. Raise if there is a cycle so the rebase is impossible. """ srcset = set(destmap) while srcset: srclist = sorted(srcset) result = [] for r in srclist: if destmap[r] not in srcset: result.append(r) if not result: raise error.InputError(_(b'source and destination form a cycle')) srcset -= set(result) yield result def buildstate(repo, destmap, collapse): """Define which revisions are going to be rebased and where repo: repo destmap: {srcrev: destrev} """ rebaseset = destmap.keys() originalwd = repo[b'.'].rev() # This check isn't strictly necessary, since mq detects commits over an # applied patch. But it prevents messing up the working directory when # a partially completed rebase is blocked by mq. if b'qtip' in repo.tags(): mqapplied = {repo[s.node].rev() for s in repo.mq.applied} if set(destmap.values()) & mqapplied: raise error.StateError(_(b'cannot rebase onto an applied mq patch')) # Get "cycle" error early by exhausting the generator. sortedsrc = list(sortsource(destmap)) # a list of sorted revs if not sortedsrc: raise error.InputError(_(b'no matching revisions')) # Only check the first batch of revisions to rebase not depending on other # rebaseset. This means "source is ancestor of destination" for the second # (and following) batches of revisions are not checked here. We rely on # "defineparents" to do that check. roots = list(repo.set(b'roots(%ld)', sortedsrc[0])) if not roots: raise error.InputError(_(b'no matching revisions')) def revof(r): return r.rev() roots = sorted(roots, key=revof) state = dict.fromkeys(rebaseset, revtodo) emptyrebase = len(sortedsrc) == 1 for root in roots: dest = repo[destmap[root.rev()]] commonbase = root.ancestor(dest) if commonbase == root: raise error.InputError(_(b'source is ancestor of destination')) if commonbase == dest: wctx = repo[None] if dest == wctx.p1(): # when rebasing to '.', it will use the current wd branch name samebranch = root.branch() == wctx.branch() else: samebranch = root.branch() == dest.branch() if not collapse and samebranch and dest in root.parents(): # mark the revision as done by setting its new revision # equal to its old (current) revisions state[root.rev()] = root.rev() repo.ui.debug(b'source is a child of destination\n') continue emptyrebase = False repo.ui.debug(b'rebase onto %s starting from %s\n' % (dest, root)) if emptyrebase: return None for rev in sorted(state): parents = [p for p in repo.changelog.parentrevs(rev) if p != nullrev] # if all parents of this revision are done, then so is this revision if parents and all((state.get(p) == p for p in parents)): state[rev] = rev return originalwd, destmap, state def clearrebased( ui, repo, destmap, state, skipped, collapsedas=None, keepf=False, fm=None, backup=True, ): """dispose of rebased revision at the end of the rebase If `collapsedas` is not None, the rebase was a collapse whose result if the `collapsedas` node. If `keepf` is not True, the rebase has --keep set and no nodes should be removed (but bookmarks still need to be moved). If `backup` is False, no backup will be stored when stripping rebased revisions. """ tonode = repo.changelog.node replacements = {} moves = {} stripcleanup = not obsolete.isenabled(repo, obsolete.createmarkersopt) collapsednodes = [] for rev, newrev in sorted(state.items()): if newrev >= 0 and newrev != rev: oldnode = tonode(rev) newnode = collapsedas or tonode(newrev) moves[oldnode] = newnode succs = None if rev in skipped: if stripcleanup or not repo[rev].obsolete(): succs = () elif collapsedas: collapsednodes.append(oldnode) else: succs = (newnode,) if succs is not None: replacements[(oldnode,)] = succs if collapsednodes: replacements[tuple(collapsednodes)] = (collapsedas,) if fm: hf = fm.hexfunc fl = fm.formatlist fd = fm.formatdict changes = {} for oldns, newn in pycompat.iteritems(replacements): for oldn in oldns: changes[hf(oldn)] = fl([hf(n) for n in newn], name=b'node') nodechanges = fd(changes, key=b"oldnode", value=b"newnodes") fm.data(nodechanges=nodechanges) if keepf: replacements = {} scmutil.cleanupnodes(repo, replacements, b'rebase', moves, backup=backup) def pullrebase(orig, ui, repo, *args, **opts): """Call rebase after pull if the latter has been invoked with --rebase""" if opts.get('rebase'): if ui.configbool(b'commands', b'rebase.requiredest'): msg = _(b'rebase destination required by configuration') hint = _(b'use hg pull followed by hg rebase -d DEST') raise error.InputError(msg, hint=hint) with repo.wlock(), repo.lock(): if opts.get('update'): del opts['update'] ui.debug( b'--update and --rebase are not compatible, ignoring ' b'the update flag\n' ) cmdutil.checkunfinished(repo, skipmerge=True) cmdutil.bailifchanged( repo, hint=_( b'cannot pull with rebase: ' b'please commit or shelve your changes first' ), ) revsprepull = len(repo) origpostincoming = commands.postincoming def _dummy(*args, **kwargs): pass commands.postincoming = _dummy try: ret = orig(ui, repo, *args, **opts) finally: commands.postincoming = origpostincoming revspostpull = len(repo) if revspostpull > revsprepull: # --rev option from pull conflict with rebase own --rev # dropping it if 'rev' in opts: del opts['rev'] # positional argument from pull conflicts with rebase's own # --source. if 'source' in opts: del opts['source'] # revsprepull is the len of the repo, not revnum of tip. destspace = list(repo.changelog.revs(start=revsprepull)) opts['_destspace'] = destspace try: rebase(ui, repo, **opts) except error.NoMergeDestAbort: # we can maybe update instead rev, _a, _b = destutil.destupdate(repo) if rev == repo[b'.'].rev(): ui.status(_(b'nothing to rebase\n')) else: ui.status(_(b'nothing to rebase - updating instead\n')) # not passing argument to get the bare update behavior # with warning and trumpets commands.update(ui, repo) else: if opts.get('tool'): raise error.InputError(_(b'--tool can only be used with --rebase')) ret = orig(ui, repo, *args, **opts) return ret def _compute_obsolete_sets(repo, rebaseobsrevs, destmap): """Figure out what to do about about obsolete revisions `obsolete_with_successor_in_destination` is a mapping mapping obsolete => successor for all obsolete nodes to be rebased given in `rebaseobsrevs`. `obsolete_with_successor_in_rebase_set` is a set with obsolete revisions, without a successor in destination, that would cause divergence. """ obsolete_with_successor_in_destination = {} obsolete_with_successor_in_rebase_set = set() cl = repo.changelog get_rev = cl.index.get_rev extinctrevs = set(repo.revs(b'extinct()')) for srcrev in rebaseobsrevs: srcnode = cl.node(srcrev) # XXX: more advanced APIs are required to handle split correctly successors = set(obsutil.allsuccessors(repo.obsstore, [srcnode])) # obsutil.allsuccessors includes node itself successors.remove(srcnode) succrevs = {get_rev(s) for s in successors} succrevs.discard(None) if not successors or succrevs.issubset(extinctrevs): # no successor, or all successors are extinct obsolete_with_successor_in_destination[srcrev] = None else: dstrev = destmap[srcrev] for succrev in succrevs: if cl.isancestorrev(succrev, dstrev): obsolete_with_successor_in_destination[srcrev] = succrev break else: # If 'srcrev' has a successor in rebase set but none in # destination (which would be catched above), we shall skip it # and its descendants to avoid divergence. if srcrev in extinctrevs or any(s in destmap for s in succrevs): obsolete_with_successor_in_rebase_set.add(srcrev) return ( obsolete_with_successor_in_destination, obsolete_with_successor_in_rebase_set, ) def abortrebase(ui, repo): with repo.wlock(), repo.lock(): rbsrt = rebaseruntime(repo, ui) rbsrt._prepareabortorcontinue(isabort=True) def continuerebase(ui, repo): with repo.wlock(), repo.lock(): rbsrt = rebaseruntime(repo, ui) ms = mergestatemod.mergestate.read(repo) mergeutil.checkunresolved(ms) retcode = rbsrt._prepareabortorcontinue(isabort=False) if retcode is not None: return retcode rbsrt._performrebase(None) rbsrt._finishrebase() def summaryhook(ui, repo): if not repo.vfs.exists(b'rebasestate'): return try: rbsrt = rebaseruntime(repo, ui, {}) rbsrt.restorestatus() state = rbsrt.state except error.RepoLookupError: # i18n: column positioning for "hg summary" msg = _(b'rebase: (use "hg rebase --abort" to clear broken state)\n') ui.write(msg) return numrebased = len([i for i in pycompat.itervalues(state) if i >= 0]) # i18n: column positioning for "hg summary" ui.write( _(b'rebase: %s, %s (rebase --continue)\n') % ( ui.label(_(b'%d rebased'), b'rebase.rebased') % numrebased, ui.label(_(b'%d remaining'), b'rebase.remaining') % (len(state) - numrebased), ) ) def uisetup(ui): # Replace pull with a decorator to provide --rebase option entry = extensions.wrapcommand(commands.table, b'pull', pullrebase) entry[1].append( (b'', b'rebase', None, _(b"rebase working directory to branch head")) ) entry[1].append((b't', b'tool', b'', _(b"specify merge tool for rebase"))) cmdutil.summaryhooks.add(b'rebase', summaryhook) statemod.addunfinished( b'rebase', fname=b'rebasestate', stopflag=True, continueflag=True, abortfunc=abortrebase, continuefunc=continuerebase, )