Mercurial > hg
view mercurial/commit.py @ 52293:77b38c86915d
ci: add a small script one can run to purge older pipeline
We have over ten thousands old pipeline that take a huge space and that I
suspect to be the source of some slowdown in merge request.
However it seems that the only way to clear them is manually and through the
API, so lets do it.
The script was run today.
author | Pierre-Yves David <pierre-yves.david@octobus.net> |
---|---|
date | Tue, 12 Nov 2024 12:45:23 +0100 |
parents | f4733654f144 |
children |
line wrap: on
line source
# commit.py - fonction to perform commit # # This software may be used and distributed according to the terms of the # GNU General Public License version 2 or any later version. from __future__ import annotations from .i18n import _ from .node import ( hex, nullrev, ) from . import ( context, mergestate, metadata, phases, scmutil, subrepoutil, ) def _write_copy_meta(repo): """return a (changelog, filelog) boolean tuple changelog: copy related information should be stored in the changeset filelof: copy related information should be written in the file revision """ if repo.filecopiesmode == b'changeset-sidedata': writechangesetcopy = True writefilecopymeta = True else: writecopiesto = repo.ui.config(b'experimental', b'copies.write-to') writefilecopymeta = writecopiesto != b'changeset-only' writechangesetcopy = writecopiesto in ( b'changeset-only', b'compatibility', ) return writechangesetcopy, writefilecopymeta def commitctx(repo, ctx, error=False, origctx=None): """Add a new revision to the target repository. Revision information is passed via the context argument. ctx.files() should list all files involved in this commit, i.e. modified/added/removed files. On merge, it may be wider than the ctx.files() to be committed, since any file nodes derived directly from p1 or p2 are excluded from the committed ctx.files(). origctx is for convert to work around the problem that bug fixes to the files list in changesets change hashes. For convert to be the identity, it can pass an origctx and this function will use the same files list when it makes sense to do so. """ repo = repo.unfiltered() p1, p2 = ctx.p1(), ctx.p2() user = ctx.user() with repo.lock(), repo.transaction(b"commit") as tr: mn, files = _prepare_files(tr, ctx, error=error, origctx=origctx) extra = ctx.extra().copy() if extra is not None: for name in ( b'p1copies', b'p2copies', b'filesadded', b'filesremoved', ): extra.pop(name, None) if repo.changelog._copiesstorage == b'extra': extra = _extra_with_copies(repo, extra, files) # save the tip to check whether we actually committed anything oldtip = repo.changelog.tiprev() # update changelog repo.ui.note(_(b"committing changelog\n")) repo.changelog.delayupdate(tr) n = repo.changelog.add( mn, files, ctx.description(), tr, p1.node(), p2.node(), user, ctx.date(), extra, ) rev = repo[n].rev() if oldtip != repo.changelog.tiprev(): repo.register_changeset(rev, repo.changelog.changelogrevision(rev)) xp1, xp2 = p1.hex(), p2 and p2.hex() or b'' repo.hook( b'pretxncommit', throw=True, node=hex(n), parent1=xp1, parent2=xp2, ) # set the new commit is proper phase targetphase = subrepoutil.newcommitphase(repo.ui, ctx) # prevent unmarking changesets as public on recommit waspublic = oldtip == repo.changelog.tiprev() and not repo[rev].phase() if targetphase and not waspublic: # retract boundary do not alter parent changeset. # if a parent have higher the resulting phase will # be compliant anyway # # if minimal phase was 0 we don't need to retract anything phases.registernew(repo, tr, targetphase, [rev]) return n def _prepare_files(tr, ctx, error=False, origctx=None): repo = ctx.repo() p1 = ctx.p1() writechangesetcopy, writefilecopymeta = _write_copy_meta(repo) files = metadata.ChangingFiles() ms = mergestate.mergestate.read(repo) salvaged = _get_salvaged(repo, ms, ctx) for s in salvaged: files.mark_salvaged(s) narrow_files = {} if not ctx.repo().narrowmatch().always(): for f, e in ms.allextras().items(): action = e.get(b'outside-narrow-merge-action') if action is not None: narrow_files[f] = action if ctx.manifestnode() and not narrow_files: # reuse an existing manifest revision repo.ui.debug(b'reusing known manifest\n') mn = ctx.manifestnode() files.update_touched(ctx.files()) if writechangesetcopy: files.update_added(ctx.filesadded()) files.update_removed(ctx.filesremoved()) elif not ctx.files() and not narrow_files: repo.ui.debug(b'reusing manifest from p1 (no file change)\n') mn = p1.manifestnode() else: mn = _process_files(tr, ctx, ms, files, narrow_files, error=error) if origctx and origctx.manifestnode() == mn: origfiles = origctx.files() assert files.touched.issubset(origfiles) files.update_touched(origfiles) if writechangesetcopy: files.update_copies_from_p1(ctx.p1copies()) files.update_copies_from_p2(ctx.p2copies()) return mn, files def _get_salvaged(repo, ms, ctx): """returns a list of salvaged files returns empty list if config option which process salvaged files are not enabled""" salvaged = [] copy_sd = repo.filecopiesmode == b'changeset-sidedata' if copy_sd and len(ctx.parents()) > 1: if ms.active(): for fname in sorted(ms.allextras().keys()): might_removed = ms.extras(fname).get(b'merge-removal-candidate') if might_removed == b'yes': if fname in ctx: salvaged.append(fname) return salvaged def _process_files(tr, ctx, ms, files, narrow_files=None, error=False): repo = ctx.repo() p1 = ctx.p1() p2 = ctx.p2() writechangesetcopy, writefilecopymeta = _write_copy_meta(repo) m1ctx = p1.manifestctx() m2ctx = p2.manifestctx() mctx = m1ctx.copy() m = mctx.read() m1 = m1ctx.read() m2 = m2ctx.read() # check in files added = [] removed = list(ctx.removed()) linkrev = len(repo) repo.ui.note(_(b"committing files:\n")) uipathfn = scmutil.getuipathfn(repo) all_files = ctx.modified() + ctx.added() all_files.extend(narrow_files.keys()) all_files.sort() for f in all_files: repo.ui.note(uipathfn(f) + b"\n") if f in narrow_files: narrow_action = narrow_files.get(f) if narrow_action == mergestate.CHANGE_REMOVED: files.mark_removed(f) removed.append(f) elif narrow_action == mergestate.CHANGE_ADDED: files.mark_added(f) added.append(f) fnode = m2[f] flags = m2ctx.find(f)[1] or b'' m.set(f, fnode, flags) elif narrow_action == mergestate.CHANGE_MODIFIED: files.mark_touched(f) added.append(f) fnode = m2[f] flags = m2ctx.find(f)[1] or b'' m.set(f, fnode, flags) else: msg = _(b"corrupted mergestate, unknown narrow action: %b") hint = _(b"restart the merge") raise error.Abort(msg, hint=hint) continue try: fctx = ctx[f] if fctx is None: removed.append(f) else: added.append(f) fnode, is_touched = _filecommit( repo, fctx, m1, m2, linkrev, tr, writefilecopymeta, ms ) if is_touched: if is_touched == 'added': files.mark_added(f) elif is_touched == 'merged': files.mark_merged(f) else: files.mark_touched(f) m.set(f, fnode, fctx.flags()) except OSError: repo.ui.warn(_(b"trouble committing %s!\n") % uipathfn(f)) raise # update manifest removed = [f for f in removed if f in m1 or f in m2] drop = sorted([f for f in removed if f in m]) for f in drop: del m[f] if p2.rev() == nullrev: files.update_removed(removed) else: rf = metadata.get_removal_filter(ctx, (p1, p2, m1, m2)) for f in removed: if not rf(f): files.mark_removed(f) mn = _commit_manifest( tr, linkrev, ctx, mctx, m, files.touched, added, drop, bool(narrow_files), ) return mn def _filecommit( repo, fctx, manifest1, manifest2, linkrev, tr, includecopymeta, ms, ): """ commit an individual file as part of a larger transaction input: fctx: a file context with the content we are trying to commit manifest1: manifest of changeset first parent manifest2: manifest of changeset second parent linkrev: revision number of the changeset being created tr: current transation includecopymeta: boolean, set to False to skip storing the copy data (only used by the Google specific feature of using changeset extra as copy source of truth). ms: mergestate object output: (filenode, touched) filenode: the filenode that should be used by this changeset touched: one of: None (mean untouched), 'added' or 'modified' """ fname = fctx.path() fparent1 = manifest1.get(fname, repo.nullid) fparent2 = manifest2.get(fname, repo.nullid) touched = None if fparent1 == fparent2 == repo.nullid: touched = 'added' if isinstance(fctx, context.filectx): # This block fast path most comparisons which are usually done. It # assumes that bare filectx is used and no merge happened, hence no # need to create a new file revision in this case. node = fctx.filenode() if node in [fparent1, fparent2]: repo.ui.debug(b'reusing %s filelog entry\n' % fname) if ( fparent1 != repo.nullid and manifest1.flags(fname) != fctx.flags() ) or ( fparent2 != repo.nullid and manifest2.flags(fname) != fctx.flags() ): touched = 'modified' return node, touched flog = repo.file(fname) meta = {} cfname = fctx.copysource() fnode = None if cfname and cfname != fname: # Mark the new revision of this file as a copy of another # file. This copy data will effectively act as a parent # of this new revision. If this is a merge, the first # parent will be the nullid (meaning "look up the copy data") # and the second one will be the other parent. For example: # # 0 --- 1 --- 3 rev1 changes file foo # \ / rev2 renames foo to bar and changes it # \- 2 -/ rev3 should have bar with all changes and # should record that bar descends from # bar in rev2 and foo in rev1 # # this allows this merge to succeed: # # 0 --- 1 --- 3 rev4 reverts the content change from rev2 # \ / merging rev3 and rev4 should use bar@rev2 # \- 2 --- 4 as the merge base # cnode = manifest1.get(cfname) newfparent = fparent2 if manifest2: # branch merge if ( fparent2 == repo.nullid or cnode is None ): # copied on remote side if cfname in manifest2: cnode = manifest2[cfname] newfparent = fparent1 # Here, we used to search backwards through history to try to find # where the file copy came from if the source of a copy was not in # the parent directory. However, this doesn't actually make sense to # do (what does a copy from something not in your working copy even # mean?) and it causes bugs (eg, issue4476). Instead, we will warn # the user that copy information was dropped, so if they didn't # expect this outcome it can be fixed, but this is the correct # behavior in this circumstance. if cnode: repo.ui.debug(b" %s: copy %s:%s\n" % (fname, cfname, hex(cnode))) if includecopymeta: meta[b"copy"] = cfname meta[b"copyrev"] = hex(cnode) fparent1, fparent2 = repo.nullid, newfparent else: repo.ui.warn( _( b"warning: can't find ancestor for '%s' " b"copied from '%s'!\n" ) % (fname, cfname) ) elif fparent1 == repo.nullid: fparent1, fparent2 = fparent2, repo.nullid elif fparent2 != repo.nullid: if ms.active() and ms.extras(fname).get(b'filenode-source') == b'other': fparent1, fparent2 = fparent2, repo.nullid elif ms.active() and ms.extras(fname).get(b'merged') != b'yes': fparent1, fparent2 = fparent1, repo.nullid # is one parent an ancestor of the other? else: fparentancestors = flog.commonancestorsheads(fparent1, fparent2) if fparent1 in fparentancestors: fparent1, fparent2 = fparent2, repo.nullid elif fparent2 in fparentancestors: fparent2 = repo.nullid force_new_node = False # The file might have been deleted by merge code and user explicitly choose # to revert the file and keep it. The other case can be where there is # change-delete or delete-change conflict and user explicitly choose to keep # the file. The goal is to create a new filenode for users explicit choices if ( repo.ui.configbool(b'experimental', b'merge-track-salvaged') and ms.active() and ms.extras(fname).get(b'merge-removal-candidate') == b'yes' ): force_new_node = True # is the file changed? text = fctx.data() if ( fparent2 != repo.nullid or fparent1 == repo.nullid or meta or flog.cmp(fparent1, text) or force_new_node ): if touched is None: # do not overwrite added if fparent2 == repo.nullid: touched = 'modified' else: touched = 'merged' fnode = flog.add(text, meta, tr, linkrev, fparent1, fparent2) # are just the flags changed during merge? elif fname in manifest1 and manifest1.flags(fname) != fctx.flags(): touched = 'modified' fnode = fparent1 else: fnode = fparent1 return fnode, touched def _commit_manifest( tr, linkrev, ctx, mctx, manifest, files, added, drop, has_some_narrow_action=False, ): """make a new manifest entry (or reuse a new one) given an initialised manifest context and precomputed list of - files: files affected by the commit - added: new entries in the manifest - drop: entries present in parents but absent of this one Create a new manifest revision, reuse existing ones if possible. Return the nodeid of the manifest revision. """ repo = ctx.repo() md = None # all this is cached, so it is find to get them all from the ctx. p1 = ctx.p1() p2 = ctx.p2() m1ctx = p1.manifestctx() m1 = m1ctx.read() if not files: # if no "files" actually changed in terms of the changelog, # try hard to detect unmodified manifest entry so that the # exact same commit can be reproduced later on convert. md = m1.diff(manifest, scmutil.matchfiles(repo, ctx.files())) if not files and md: repo.ui.debug( b'not reusing manifest (no file change in ' b'changelog, but manifest differs)\n' ) if files or md: repo.ui.note(_(b"committing manifest\n")) # we're using narrowmatch here since it's already applied at # other stages (such as dirstate.walk), so we're already # ignoring things outside of narrowspec in most cases. The # one case where we might have files outside the narrowspec # at this point is merges, and we already error out in the # case where the merge has files outside of the narrowspec, # so this is safe. if has_some_narrow_action: match = None else: match = repo.narrowmatch() mn = mctx.write( tr, linkrev, p1.manifestnode(), p2.manifestnode(), added, drop, match=match, ) else: repo.ui.debug( b'reusing manifest from p1 (listed files ' b'actually unchanged)\n' ) mn = p1.manifestnode() return mn def _extra_with_copies(repo, extra, files): """encode copy information into a `extra` dictionnary""" p1copies = files.copied_from_p1 p2copies = files.copied_from_p2 filesadded = files.added filesremoved = files.removed files = sorted(files.touched) if not _write_copy_meta(repo)[1]: # If writing only to changeset extras, use None to indicate that # no entry should be written. If writing to both, write an empty # entry to prevent the reader from falling back to reading # filelogs. p1copies = p1copies or None p2copies = p2copies or None filesadded = filesadded or None filesremoved = filesremoved or None extrasentries = p1copies, p2copies, filesadded, filesremoved if extra is None and any(x is not None for x in extrasentries): extra = {} if p1copies is not None: p1copies = metadata.encodecopies(files, p1copies) extra[b'p1copies'] = p1copies if p2copies is not None: p2copies = metadata.encodecopies(files, p2copies) extra[b'p2copies'] = p2copies if filesadded is not None: filesadded = metadata.encodefileindices(files, filesadded) extra[b'filesadded'] = filesadded if filesremoved is not None: filesremoved = metadata.encodefileindices(files, filesremoved) extra[b'filesremoved'] = filesremoved return extra