exchange: improve computation of relevant markers for large repos
Compute the candidate nodes with relevant markers directly
from keys of the predecessors/successors/children dictionaries of
obsstore. This is faster than iterating over all nodes directly.
This test could be further improved for repositories with relative
few markers compared to the repository size, but this is no longer
hot already. With the current loop structure, the obshashrange use
works as well as before as it passes lists with a single node.
Adjust the interface by allowing revision lists as well as node lists.
This helps cases that computes ancestors as it reduces the
materialisation cost. Use this in _pushdiscoveryobsmarker and
_getbundleobsmarkerpart. Improve the latter further by directly using
ancestors().
Performance benchmarks show notable and welcome improvement to no-op push and
pull (that would also apply to other push/pull). This apply to push and pull
done without evolve.
### push/pull Benchmark parameter
# bin-env-vars.hg.flavor = default
# benchmark.variants.explicit-rev = none
# benchmark.variants.protocol = ssh
# benchmark.variants.revs = none
## benchmark.name = hg.command.pull
# data-env-vars.name = mercurial-devel-2024-03-22-zstd-sparse-revlog
before: 5.968537 seconds
after: 5.668507 seconds (-5.03%, -0.30)
# data-env-vars.name = tryton-devel-2024-03-22-zstd-sparse-revlog
before: 1.446232 seconds
after: 0.835553 seconds (-42.23%, -0.61)
# data-env-vars.name = netbsd-src-draft-2024-09-19-zstd-sparse-revlog
before: 5.777412 seconds
after: 2.523454 seconds (-56.32%, -3.25)
## benchmark.name = hg.command.push
# data-env-vars.name = mercurial-devel-2024-03-22-zstd-sparse-revlog
before: 6.155501 seconds
after: 5.885072 seconds (-4.39%, -0.27)
# data-env-vars.name = tryton-devel-2024-03-22-zstd-sparse-revlog
before: 1.491054 seconds
after: 0.934882 seconds (-37.30%, -0.56)
# data-env-vars.name = netbsd-src-draft-2024-09-19-zstd-sparse-revlog
before: 5.902494 seconds
after: 2.957644 seconds (-49.89%, -2.94)
There is not notable different in these result using the "rust" flavor instead
of the "default". The performance impact on the same operation when using
evolve were also tested and no impact was noted.
# patch.py - patch file parsing routines
#
# Copyright 2006 Brendan Cully <brendan@kublai.com>
# Copyright 2007 Chris Mason <chris.mason@oracle.com>
#
# This software may be used and distributed according to the terms of the
# GNU General Public License version 2 or any later version.
from __future__ import annotations
import collections
import contextlib
import copy
import os
import re
import shutil
import zlib
from .i18n import _
from .node import (
hex,
sha1nodeconstants,
short,
)
from .pycompat import open
from . import (
copies,
diffhelper,
diffutil,
encoding,
error,
mail,
mdiff,
pathutil,
pycompat,
scmutil,
similar,
util,
vfs as vfsmod,
)
from .utils import (
dateutil,
hashutil,
procutil,
stringutil,
)
stringio = util.stringio
gitre = re.compile(br'diff --git a/(.*) b/(.*)')
tabsplitter = re.compile(br'(\t+|[^\t]+)')
wordsplitter = re.compile(
br'(\t+| +|[a-zA-Z0-9_\x80-\xff]+|[^ \ta-zA-Z0-9_\x80-\xff])'
)
PatchError = error.PatchError
PatchParseError = error.PatchParseError
PatchApplicationError = error.PatchApplicationError
# public functions
def split(stream):
'''return an iterator of individual patches from a stream'''
def isheader(line, inheader):
if inheader and line.startswith((b' ', b'\t')):
# continuation
return True
if line.startswith((b' ', b'-', b'+')):
# diff line - don't check for header pattern in there
return False
l = line.split(b': ', 1)
return len(l) == 2 and b' ' not in l[0]
def chunk(lines):
return stringio(b''.join(lines))
def hgsplit(stream, cur):
inheader = True
for line in stream:
if not line.strip():
inheader = False
if not inheader and line.startswith(b'# HG changeset patch'):
yield chunk(cur)
cur = []
inheader = True
cur.append(line)
if cur:
yield chunk(cur)
def mboxsplit(stream, cur):
for line in stream:
if line.startswith(b'From '):
for c in split(chunk(cur[1:])):
yield c
cur = []
cur.append(line)
if cur:
for c in split(chunk(cur[1:])):
yield c
def mimesplit(stream, cur):
def msgfp(m):
fp = stringio()
# pytype: disable=wrong-arg-types
g = mail.Generator(fp, mangle_from_=False)
# pytype: enable=wrong-arg-types
g.flatten(m)
fp.seek(0)
return fp
for line in stream:
cur.append(line)
c = chunk(cur)
m = mail.parse(c)
if not m.is_multipart():
yield msgfp(m)
else:
ok_types = ('text/plain', 'text/x-diff', 'text/x-patch')
for part in m.walk():
ct = part.get_content_type()
if ct not in ok_types:
continue
yield msgfp(part)
def headersplit(stream, cur):
inheader = False
for line in stream:
if not inheader and isheader(line, inheader):
yield chunk(cur)
cur = []
inheader = True
if inheader and not isheader(line, inheader):
inheader = False
cur.append(line)
if cur:
yield chunk(cur)
def remainder(cur):
yield chunk(cur)
class fiter:
def __init__(self, fp):
self.fp = fp
def __iter__(self):
return self
def next(self):
l = self.fp.readline()
if not l:
raise StopIteration
return l
__next__ = next
inheader = False
cur = []
mimeheaders = [b'content-type']
if not hasattr(stream, 'next'):
# http responses, for example, have readline but not next
stream = fiter(stream)
for line in stream:
cur.append(line)
if line.startswith(b'# HG changeset patch'):
return hgsplit(stream, cur)
elif line.startswith(b'From '):
return mboxsplit(stream, cur)
elif isheader(line, inheader):
inheader = True
if line.split(b':', 1)[0].lower() in mimeheaders:
# let email parser handle this
return mimesplit(stream, cur)
elif line.startswith(b'--- ') and inheader:
# No evil headers seen by diff start, split by hand
return headersplit(stream, cur)
# Not enough info, keep reading
# if we are here, we have a very plain patch
return remainder(cur)
## Some facility for extensible patch parsing:
# list of pairs ("header to match", "data key")
patchheadermap = [
(b'Date', b'date'),
(b'Branch', b'branch'),
(b'Node ID', b'nodeid'),
]
@contextlib.contextmanager
def extract(ui, fileobj):
"""extract patch from data read from fileobj.
patch can be a normal patch or contained in an email message.
return a dictionary. Standard keys are:
- filename,
- message,
- user,
- date,
- branch,
- node,
- p1,
- p2.
Any item can be missing from the dictionary. If filename is missing,
fileobj did not contain a patch. Caller must unlink filename when done."""
fd, tmpname = pycompat.mkstemp(prefix=b'hg-patch-')
tmpfp = os.fdopen(fd, 'wb')
try:
yield _extract(ui, fileobj, tmpname, tmpfp)
finally:
tmpfp.close()
os.unlink(tmpname)
def _extract(ui, fileobj, tmpname, tmpfp):
# attempt to detect the start of a patch
# (this heuristic is borrowed from quilt)
diffre = re.compile(
br'^(?:Index:[ \t]|diff[ \t]-|RCS file: |'
br'retrieving revision [0-9]+(\.[0-9]+)*$|'
br'---[ \t].*?^\+\+\+[ \t]|'
br'\*\*\*[ \t].*?^---[ \t])',
re.MULTILINE | re.DOTALL,
)
data = {}
msg = mail.parse(fileobj)
subject = msg['Subject'] and mail.headdecode(msg['Subject'])
data[b'user'] = msg['From'] and mail.headdecode(msg['From'])
if not subject and not data[b'user']:
# Not an email, restore parsed headers if any
subject = (
b'\n'.join(
b': '.join(map(encoding.strtolocal, h)) for h in msg.items()
)
+ b'\n'
)
# should try to parse msg['Date']
parents = []
nodeid = msg['X-Mercurial-Node']
if nodeid:
data[b'nodeid'] = nodeid = mail.headdecode(nodeid)
ui.debug(b'Node ID: %s\n' % nodeid)
if subject:
if subject.startswith(b'[PATCH'):
pend = subject.find(b']')
if pend >= 0:
subject = subject[pend + 1 :].lstrip()
subject = re.sub(br'\n[ \t]+', b' ', subject)
ui.debug(b'Subject: %s\n' % subject)
if data[b'user']:
ui.debug(b'From: %s\n' % data[b'user'])
diffs_seen = 0
ok_types = (b'text/plain', b'text/x-diff', b'text/x-patch')
message = b''
for part in msg.walk():
content_type = pycompat.bytestr(part.get_content_type())
ui.debug(b'Content-Type: %s\n' % content_type)
if content_type not in ok_types:
continue
payload = part.get_payload(decode=True)
m = diffre.search(payload)
if m:
hgpatch = False
hgpatchheader = False
ignoretext = False
ui.debug(b'found patch at byte %d\n' % m.start(0))
diffs_seen += 1
cfp = stringio()
for line in payload[: m.start(0)].splitlines():
if line.startswith(b'# HG changeset patch') and not hgpatch:
ui.debug(b'patch generated by hg export\n')
hgpatch = True
hgpatchheader = True
# drop earlier commit message content
cfp.seek(0)
cfp.truncate()
subject = None
elif hgpatchheader:
if line.startswith(b'# User '):
data[b'user'] = line[7:]
ui.debug(b'From: %s\n' % data[b'user'])
elif line.startswith(b"# Parent "):
parents.append(line[9:].lstrip())
elif line.startswith(b"# "):
for header, key in patchheadermap:
prefix = b'# %s ' % header
if line.startswith(prefix):
data[key] = line[len(prefix) :]
ui.debug(b'%s: %s\n' % (header, data[key]))
else:
hgpatchheader = False
elif line == b'---':
ignoretext = True
if not hgpatchheader and not ignoretext:
cfp.write(line)
cfp.write(b'\n')
message = cfp.getvalue()
if tmpfp:
tmpfp.write(payload)
if not payload.endswith(b'\n'):
tmpfp.write(b'\n')
elif not diffs_seen and message and content_type == b'text/plain':
message += b'\n' + payload
if subject and not message.startswith(subject):
message = b'%s\n%s' % (subject, message)
data[b'message'] = message
tmpfp.close()
if parents:
data[b'p1'] = parents.pop(0)
if parents:
data[b'p2'] = parents.pop(0)
if diffs_seen:
data[b'filename'] = tmpname
return data
class patchmeta:
"""Patched file metadata
'op' is the performed operation within ADD, DELETE, RENAME, MODIFY
or COPY. 'path' is patched file path. 'oldpath' is set to the
origin file when 'op' is either COPY or RENAME, None otherwise. If
file mode is changed, 'mode' is a tuple (islink, isexec) where
'islink' is True if the file is a symlink and 'isexec' is True if
the file is executable. Otherwise, 'mode' is None.
"""
def __init__(self, path):
self.path = path
self.oldpath = None
self.mode = None
self.op = b'MODIFY'
self.binary = False
def setmode(self, mode):
islink = mode & 0o20000
isexec = mode & 0o100
self.mode = (islink, isexec)
def copy(self):
other = patchmeta(self.path)
other.oldpath = self.oldpath
other.mode = self.mode
other.op = self.op
other.binary = self.binary
return other
def _ispatchinga(self, afile):
if afile == b'/dev/null':
return self.op == b'ADD'
return afile == b'a/' + (self.oldpath or self.path)
def _ispatchingb(self, bfile):
if bfile == b'/dev/null':
return self.op == b'DELETE'
return bfile == b'b/' + self.path
def ispatching(self, afile, bfile):
return self._ispatchinga(afile) and self._ispatchingb(bfile)
def __repr__(self):
return "<patchmeta %s %r>" % (self.op, self.path)
def readgitpatch(lr):
"""extract git-style metadata about patches from <patchname>"""
# Filter patch for git information
gp = None
gitpatches = []
for line in lr:
line = line.rstrip(b'\r\n')
if line.startswith(b'diff --git a/'):
m = gitre.match(line)
if m:
if gp:
gitpatches.append(gp)
dst = m.group(2)
gp = patchmeta(dst)
elif gp:
if line.startswith(b'--- '):
gitpatches.append(gp)
gp = None
continue
if line.startswith(b'rename from '):
gp.op = b'RENAME'
gp.oldpath = line[12:]
elif line.startswith(b'rename to '):
gp.path = line[10:]
elif line.startswith(b'copy from '):
gp.op = b'COPY'
gp.oldpath = line[10:]
elif line.startswith(b'copy to '):
gp.path = line[8:]
elif line.startswith(b'deleted file'):
gp.op = b'DELETE'
elif line.startswith(b'new file mode '):
gp.op = b'ADD'
gp.setmode(int(line[-6:], 8))
elif line.startswith(b'new mode '):
gp.setmode(int(line[-6:], 8))
elif line.startswith(b'GIT binary patch'):
gp.binary = True
if gp:
gitpatches.append(gp)
return gitpatches
class linereader:
# simple class to allow pushing lines back into the input stream
def __init__(self, fp):
self.fp = fp
self.buf = []
def push(self, line):
if line is not None:
self.buf.append(line)
def readline(self):
if self.buf:
l = self.buf[0]
del self.buf[0]
return l
return self.fp.readline()
def __iter__(self):
return iter(self.readline, b'')
class abstractbackend:
def __init__(self, ui):
self.ui = ui
def getfile(self, fname):
"""Return target file data and flags as a (data, (islink,
isexec)) tuple. Data is None if file is missing/deleted.
"""
raise NotImplementedError
def setfile(self, fname, data, mode, copysource):
"""Write data to target file fname and set its mode. mode is a
(islink, isexec) tuple. If data is None, the file content should
be left unchanged. If the file is modified after being copied,
copysource is set to the original file name.
"""
raise NotImplementedError
def unlink(self, fname):
"""Unlink target file."""
raise NotImplementedError
def writerej(self, fname, failed, total, lines):
"""Write rejected lines for fname. total is the number of hunks
which failed to apply and total the total number of hunks for this
files.
"""
def exists(self, fname):
raise NotImplementedError
def close(self):
raise NotImplementedError
class fsbackend(abstractbackend):
def __init__(self, ui, basedir):
super(fsbackend, self).__init__(ui)
self.opener = vfsmod.vfs(basedir)
def getfile(self, fname):
if self.opener.islink(fname):
return (self.opener.readlink(fname), (True, False))
isexec = False
try:
isexec = self.opener.lstat(fname).st_mode & 0o100 != 0
except FileNotFoundError:
pass
try:
return (self.opener.read(fname), (False, isexec))
except FileNotFoundError:
return None, None
def setfile(self, fname, data, mode, copysource):
islink, isexec = mode
if data is None:
self.opener.setflags(fname, islink, isexec)
return
if islink:
self.opener.symlink(data, fname)
else:
self.opener.write(fname, data)
if isexec:
self.opener.setflags(fname, False, True)
def unlink(self, fname):
rmdir = self.ui.configbool(b'experimental', b'removeemptydirs')
self.opener.unlinkpath(fname, ignoremissing=True, rmdir=rmdir)
def writerej(self, fname, failed, total, lines):
fname = fname + b".rej"
self.ui.warn(
_(b"%d out of %d hunks FAILED -- saving rejects to file %s\n")
% (failed, total, fname)
)
fp = self.opener(fname, b'w')
fp.writelines(lines)
fp.close()
def exists(self, fname):
return self.opener.lexists(fname)
class workingbackend(fsbackend):
def __init__(self, ui, repo, similarity):
super(workingbackend, self).__init__(ui, repo.root)
self.repo = repo
self.similarity = similarity
self.removed = set()
self.changed = set()
self.copied = []
def _checkknown(self, fname):
if not self.repo.dirstate.get_entry(fname).any_tracked and self.exists(
fname
):
raise PatchApplicationError(
_(b'cannot patch %s: file is not tracked') % fname
)
def setfile(self, fname, data, mode, copysource):
self._checkknown(fname)
super(workingbackend, self).setfile(fname, data, mode, copysource)
if copysource is not None:
self.copied.append((copysource, fname))
self.changed.add(fname)
def unlink(self, fname):
self._checkknown(fname)
super(workingbackend, self).unlink(fname)
self.removed.add(fname)
self.changed.add(fname)
def close(self):
with self.repo.dirstate.changing_files(self.repo):
wctx = self.repo[None]
changed = set(self.changed)
for src, dst in self.copied:
scmutil.dirstatecopy(self.ui, self.repo, wctx, src, dst)
if self.removed:
wctx.forget(sorted(self.removed))
for f in self.removed:
if f not in self.repo.dirstate:
# File was deleted and no longer belongs to the
# dirstate, it was probably marked added then
# deleted, and should not be considered by
# marktouched().
changed.discard(f)
if changed:
scmutil.marktouched(self.repo, changed, self.similarity)
return sorted(self.changed)
class filestore:
def __init__(self, maxsize=None):
self.opener = None
self.files = {}
self.created = 0
self.maxsize = maxsize
if self.maxsize is None:
self.maxsize = 4 * (2**20)
self.size = 0
self.data = {}
def setfile(self, fname, data, mode, copied=None):
if self.maxsize < 0 or (len(data) + self.size) <= self.maxsize:
self.data[fname] = (data, mode, copied)
self.size += len(data)
else:
if self.opener is None:
root = pycompat.mkdtemp(prefix=b'hg-patch-')
self.opener = vfsmod.vfs(root)
# Avoid filename issues with these simple names
fn = b'%d' % self.created
self.opener.write(fn, data)
self.created += 1
self.files[fname] = (fn, mode, copied)
def getfile(self, fname):
if fname in self.data:
return self.data[fname]
if not self.opener or fname not in self.files:
return None, None, None
fn, mode, copied = self.files[fname]
return self.opener.read(fn), mode, copied
def close(self):
if self.opener:
shutil.rmtree(self.opener.base)
class repobackend(abstractbackend):
def __init__(self, ui, repo, ctx, store):
super(repobackend, self).__init__(ui)
self.repo = repo
self.ctx = ctx
self.store = store
self.changed = set()
self.removed = set()
self.copied = {}
def _checkknown(self, fname):
if fname not in self.ctx:
raise PatchApplicationError(
_(b'cannot patch %s: file is not tracked') % fname
)
def getfile(self, fname):
try:
fctx = self.ctx[fname]
except error.LookupError:
return None, None
flags = fctx.flags()
return fctx.data(), (b'l' in flags, b'x' in flags)
def setfile(self, fname, data, mode, copysource):
if copysource:
self._checkknown(copysource)
if data is None:
data = self.ctx[fname].data()
self.store.setfile(fname, data, mode, copysource)
self.changed.add(fname)
if copysource:
self.copied[fname] = copysource
def unlink(self, fname):
self._checkknown(fname)
self.removed.add(fname)
def exists(self, fname):
return fname in self.ctx
def close(self):
return self.changed | self.removed
# @@ -start,len +start,len @@ or @@ -start +start @@ if len is 1
unidesc = re.compile(br'@@ -(\d+)(?:,(\d+))? \+(\d+)(?:,(\d+))? @@')
contextdesc = re.compile(br'(?:---|\*\*\*) (\d+)(?:,(\d+))? (?:---|\*\*\*)')
eolmodes = [b'strict', b'crlf', b'lf', b'auto']
class patchfile:
def __init__(self, ui, gp, backend, store, eolmode=b'strict'):
self.fname = gp.path
self.eolmode = eolmode
self.eol = None
self.backend = backend
self.ui = ui
self.lines = []
self.exists = False
self.missing = True
self.mode = gp.mode
self.copysource = gp.oldpath
self.create = gp.op in (b'ADD', b'COPY', b'RENAME')
self.remove = gp.op == b'DELETE'
if self.copysource is None:
data, mode = backend.getfile(self.fname)
else:
data, mode = store.getfile(self.copysource)[:2]
if data is not None:
self.exists = self.copysource is None or backend.exists(self.fname)
self.missing = False
if data:
self.lines = mdiff.splitnewlines(data)
if self.mode is None:
self.mode = mode
if self.lines:
# Normalize line endings
if self.lines[0].endswith(b'\r\n'):
self.eol = b'\r\n'
elif self.lines[0].endswith(b'\n'):
self.eol = b'\n'
if eolmode != b'strict':
nlines = []
for l in self.lines:
if l.endswith(b'\r\n'):
l = l[:-2] + b'\n'
nlines.append(l)
self.lines = nlines
else:
if self.create:
self.missing = False
if self.mode is None:
self.mode = (False, False)
if self.missing:
self.ui.warn(_(b"unable to find '%s' for patching\n") % self.fname)
self.ui.warn(
_(
b"(use '--prefix' to apply patch relative to the "
b"current directory)\n"
)
)
self.hash = {}
self.dirty = 0
self.offset = 0
self.skew = 0
self.rej = []
self.fileprinted = False
self.printfile(False)
self.hunks = 0
def writelines(self, fname, lines, mode):
if self.eolmode == b'auto':
eol = self.eol
elif self.eolmode == b'crlf':
eol = b'\r\n'
else:
eol = b'\n'
if self.eolmode != b'strict' and eol and eol != b'\n':
rawlines = []
for l in lines:
if l and l.endswith(b'\n'):
l = l[:-1] + eol
rawlines.append(l)
lines = rawlines
self.backend.setfile(fname, b''.join(lines), mode, self.copysource)
def printfile(self, warn):
if self.fileprinted:
return
if warn or self.ui.verbose:
self.fileprinted = True
s = _(b"patching file %s\n") % self.fname
if warn:
self.ui.warn(s)
else:
self.ui.note(s)
def findlines(self, l, linenum):
# looks through the hash and finds candidate lines. The
# result is a list of line numbers sorted based on distance
# from linenum
cand = self.hash.get(l, [])
if len(cand) > 1:
# resort our list of potentials forward then back.
cand.sort(key=lambda x: abs(x - linenum))
return cand
def write_rej(self):
# our rejects are a little different from patch(1). This always
# creates rejects in the same form as the original patch. A file
# header is inserted so that you can run the reject through patch again
# without having to type the filename.
if not self.rej:
return
base = os.path.basename(self.fname)
lines = [b"--- %s\n+++ %s\n" % (base, base)]
for x in self.rej:
for l in x.hunk:
lines.append(l)
if l[-1:] != b'\n':
lines.append(b'\n' + diffhelper.MISSING_NEWLINE_MARKER)
self.backend.writerej(self.fname, len(self.rej), self.hunks, lines)
def apply(self, h):
if not h.complete():
raise PatchParseError(
_(b"bad hunk #%d %s (%d %d %d %d)")
% (h.number, h.desc, len(h.a), h.lena, len(h.b), h.lenb)
)
self.hunks += 1
if self.missing:
self.rej.append(h)
return -1
if self.exists and self.create:
if self.copysource:
self.ui.warn(
_(b"cannot create %s: destination already exists\n")
% self.fname
)
else:
self.ui.warn(_(b"file %s already exists\n") % self.fname)
self.rej.append(h)
return -1
if isinstance(h, binhunk):
if self.remove:
self.backend.unlink(self.fname)
else:
l = h.new(self.lines)
self.lines[:] = l
self.offset += len(l)
self.dirty = True
return 0
horig = h
if (
self.eolmode in (b'crlf', b'lf')
or self.eolmode == b'auto'
and self.eol
):
# If new eols are going to be normalized, then normalize
# hunk data before patching. Otherwise, preserve input
# line-endings.
h = h.getnormalized()
# fast case first, no offsets, no fuzz
old, oldstart, new, newstart = h.fuzzit(0, False)
oldstart += self.offset
orig_start = oldstart
# if there's skew we want to emit the "(offset %d lines)" even
# when the hunk cleanly applies at start + skew, so skip the
# fast case code
if self.skew == 0 and diffhelper.testhunk(old, self.lines, oldstart):
if self.remove:
self.backend.unlink(self.fname)
else:
self.lines[oldstart : oldstart + len(old)] = new
self.offset += len(new) - len(old)
self.dirty = True
return 0
# ok, we couldn't match the hunk. Lets look for offsets and fuzz it
self.hash = {}
for x, s in enumerate(self.lines):
self.hash.setdefault(s, []).append(x)
for fuzzlen in range(self.ui.configint(b"patch", b"fuzz") + 1):
for toponly in [True, False]:
old, oldstart, new, newstart = h.fuzzit(fuzzlen, toponly)
oldstart = oldstart + self.offset + self.skew
oldstart = min(oldstart, len(self.lines))
if old:
cand = self.findlines(old[0][1:], oldstart)
else:
# Only adding lines with no or fuzzed context, just
# take the skew in account
cand = [oldstart]
for l in cand:
if not old or diffhelper.testhunk(old, self.lines, l):
self.lines[l : l + len(old)] = new
self.offset += len(new) - len(old)
self.skew = l - orig_start
self.dirty = True
offset = l - orig_start - fuzzlen
if fuzzlen:
msg = _(
b"Hunk #%d succeeded at %d "
b"with fuzz %d "
b"(offset %d lines).\n"
)
self.printfile(True)
self.ui.warn(
msg % (h.number, l + 1, fuzzlen, offset)
)
else:
msg = _(
b"Hunk #%d succeeded at %d "
b"(offset %d lines).\n"
)
self.ui.note(msg % (h.number, l + 1, offset))
return fuzzlen
self.printfile(True)
self.ui.warn(_(b"Hunk #%d FAILED at %d\n") % (h.number, orig_start))
self.rej.append(horig)
return -1
def close(self):
if self.dirty:
self.writelines(self.fname, self.lines, self.mode)
self.write_rej()
return len(self.rej)
class header:
"""patch header"""
diffgit_re = re.compile(b'diff --git a/(.*) b/(.*)$')
diff_re = re.compile(b'diff -r .* (.*)$')
allhunks_re = re.compile(b'(?:index|deleted file) ')
pretty_re = re.compile(b'(?:new file|deleted file) ')
special_re = re.compile(b'(?:index|deleted|copy|rename|new mode) ')
newfile_re = re.compile(b'(?:new file|copy to|rename to)')
def __init__(self, header):
self.header = header
self.hunks = []
def binary(self):
return any(h.startswith(b'index ') for h in self.header)
def pretty(self, fp):
for h in self.header:
if h.startswith(b'index '):
fp.write(_(b'this modifies a binary file (all or nothing)\n'))
break
if self.pretty_re.match(h):
fp.write(h)
if self.binary():
fp.write(_(b'this is a binary file\n'))
break
if h.startswith(b'---'):
fp.write(
_(b'%d hunks, %d lines changed\n')
% (
len(self.hunks),
sum([max(h.added, h.removed) for h in self.hunks]),
)
)
break
fp.write(h)
def write(self, fp):
fp.write(b''.join(self.header))
def allhunks(self):
return any(self.allhunks_re.match(h) for h in self.header)
def files(self):
match = self.diffgit_re.match(self.header[0])
if match:
fromfile, tofile = match.groups()
if fromfile == tofile:
return [fromfile]
return [fromfile, tofile]
else:
return self.diff_re.match(self.header[0]).groups()
def filename(self):
return self.files()[-1]
def __repr__(self):
return '<header %s>' % (
' '.join(pycompat.rapply(pycompat.fsdecode, self.files()))
)
def isnewfile(self):
return any(self.newfile_re.match(h) for h in self.header)
def special(self):
# Special files are shown only at the header level and not at the hunk
# level for example a file that has been deleted is a special file.
# The user cannot change the content of the operation, in the case of
# the deleted file he has to take the deletion or not take it, he
# cannot take some of it.
# Newly added files are special if they are empty, they are not special
# if they have some content as we want to be able to change it
nocontent = len(self.header) == 2
emptynewfile = self.isnewfile() and nocontent
return emptynewfile or any(
self.special_re.match(h) for h in self.header
)
class recordhunk:
"""patch hunk
XXX shouldn't we merge this with the other hunk class?
"""
def __init__(
self,
header,
fromline,
toline,
proc,
before,
hunk,
after,
maxcontext=None,
):
def trimcontext(lines, reverse=False):
if maxcontext is not None:
delta = len(lines) - maxcontext
if delta > 0:
if reverse:
return delta, lines[delta:]
else:
return delta, lines[:maxcontext]
return 0, lines
self.header = header
trimedbefore, self.before = trimcontext(before, True)
self.fromline = fromline + trimedbefore
self.toline = toline + trimedbefore
_trimedafter, self.after = trimcontext(after, False)
self.proc = proc
self.hunk = hunk
self.added, self.removed = self.countchanges(self.hunk)
def __eq__(self, v):
if not isinstance(v, recordhunk):
return False
return (
(v.hunk == self.hunk)
and (v.proc == self.proc)
and (self.fromline == v.fromline)
and (self.header.files() == v.header.files())
)
def __hash__(self):
return hash(
(
tuple(self.hunk),
tuple(self.header.files()),
self.fromline,
self.proc,
)
)
def countchanges(self, hunk):
"""hunk -> (n+,n-)"""
add = len([h for h in hunk if h.startswith(b'+')])
rem = len([h for h in hunk if h.startswith(b'-')])
return add, rem
def reversehunk(self):
"""return another recordhunk which is the reverse of the hunk
If this hunk is diff(A, B), the returned hunk is diff(B, A). To do
that, swap fromline/toline and +/- signs while keep other things
unchanged.
"""
m = {b'+': b'-', b'-': b'+', b'\\': b'\\'}
hunk = [b'%s%s' % (m[l[0:1]], l[1:]) for l in self.hunk]
return recordhunk(
self.header,
self.toline,
self.fromline,
self.proc,
self.before,
hunk,
self.after,
)
def write(self, fp):
delta = len(self.before) + len(self.after)
if self.after and self.after[-1] == diffhelper.MISSING_NEWLINE_MARKER:
delta -= 1
fromlen = delta + self.removed
tolen = delta + self.added
fp.write(
b'@@ -%d,%d +%d,%d @@%s\n'
% (
self.fromline,
fromlen,
self.toline,
tolen,
self.proc and (b' ' + self.proc),
)
)
fp.write(b''.join(self.before + self.hunk + self.after))
pretty = write
def filename(self):
return self.header.filename()
@encoding.strmethod
def __repr__(self):
return b'<hunk %r@%d>' % (self.filename(), self.fromline)
def getmessages():
return {
b'multiple': {
b'apply': _(b"apply change %d/%d to '%s'?"),
b'discard': _(b"discard change %d/%d to '%s'?"),
b'keep': _(b"keep change %d/%d to '%s'?"),
b'record': _(b"record change %d/%d to '%s'?"),
},
b'single': {
b'apply': _(b"apply this change to '%s'?"),
b'discard': _(b"discard this change to '%s'?"),
b'keep': _(b"keep this change to '%s'?"),
b'record': _(b"record this change to '%s'?"),
},
b'help': {
b'apply': _(
b'[Ynesfdaq?]'
b'$$ &Yes, apply this change'
b'$$ &No, skip this change'
b'$$ &Edit this change manually'
b'$$ &Skip remaining changes to this file'
b'$$ Apply remaining changes to this &file'
b'$$ &Done, skip remaining changes and files'
b'$$ Apply &all changes to all remaining files'
b'$$ &Quit, applying no changes'
b'$$ &? (display help)'
),
b'discard': _(
b'[Ynesfdaq?]'
b'$$ &Yes, discard this change'
b'$$ &No, skip this change'
b'$$ &Edit this change manually'
b'$$ &Skip remaining changes to this file'
b'$$ Discard remaining changes to this &file'
b'$$ &Done, skip remaining changes and files'
b'$$ Discard &all changes to all remaining files'
b'$$ &Quit, discarding no changes'
b'$$ &? (display help)'
),
b'keep': _(
b'[Ynesfdaq?]'
b'$$ &Yes, keep this change'
b'$$ &No, skip this change'
b'$$ &Edit this change manually'
b'$$ &Skip remaining changes to this file'
b'$$ Keep remaining changes to this &file'
b'$$ &Done, skip remaining changes and files'
b'$$ Keep &all changes to all remaining files'
b'$$ &Quit, keeping all changes'
b'$$ &? (display help)'
),
b'record': _(
b'[Ynesfdaq?]'
b'$$ &Yes, record this change'
b'$$ &No, skip this change'
b'$$ &Edit this change manually'
b'$$ &Skip remaining changes to this file'
b'$$ Record remaining changes to this &file'
b'$$ &Done, skip remaining changes and files'
b'$$ Record &all changes to all remaining files'
b'$$ &Quit, recording no changes'
b'$$ &? (display help)'
),
},
}
def filterpatch(ui, headers, match, operation=None):
"""Interactively filter patch chunks into applied-only chunks"""
messages = getmessages()
if operation is None:
operation = b'record'
def prompt(skipfile, skipall, query, chunk):
"""prompt query, and process base inputs
- y/n for the rest of file
- y/n for the rest
- ? (help)
- q (quit)
Return True/False and possibly updated skipfile and skipall.
"""
newpatches = None
if skipall is not None:
return skipall, skipfile, skipall, newpatches
if skipfile is not None:
return skipfile, skipfile, skipall, newpatches
while True:
resps = messages[b'help'][operation]
# IMPORTANT: keep the last line of this prompt short (<40 english
# chars is a good target) because of issue6158.
r = ui.promptchoice(b"%s\n(enter ? for help) %s" % (query, resps))
ui.write(b"\n")
if r == 8: # ?
for c, t in ui.extractchoices(resps)[1]:
ui.write(b'%s - %s\n' % (c, encoding.lower(t)))
continue
elif r == 0: # yes
ret = True
elif r == 1: # no
ret = False
elif r == 2: # Edit patch
if chunk is None:
ui.write(_(b'cannot edit patch for whole file'))
ui.write(b"\n")
continue
if chunk.header.binary():
ui.write(_(b'cannot edit patch for binary file'))
ui.write(b"\n")
continue
# Patch comment based on the Git one (based on comment at end of
# https://mercurial-scm.org/wiki/RecordExtension)
phelp = b'---' + _(
b"""
To remove '-' lines, make them ' ' lines (context).
To remove '+' lines, delete them.
Lines starting with # will be removed from the patch.
If the patch applies cleanly, the edited hunk will immediately be
added to the record list. If it does not apply cleanly, a rejects
file will be generated: you can use that when you try again. If
all lines of the hunk are removed, then the edit is aborted and
the hunk is left unchanged.
"""
)
(patchfd, patchfn) = pycompat.mkstemp(
prefix=b"hg-editor-", suffix=b".diff"
)
ncpatchfp = None
try:
# Write the initial patch
f = util.nativeeolwriter(os.fdopen(patchfd, 'wb'))
chunk.header.write(f)
chunk.write(f)
f.write(
b''.join(
[b'# ' + i + b'\n' for i in phelp.splitlines()]
)
)
f.close()
# Start the editor and wait for it to complete
editor = ui.geteditor()
ret = ui.system(
b"%s \"%s\"" % (editor, patchfn),
environ={b'HGUSER': ui.username()},
blockedtag=b'filterpatch',
)
if ret != 0:
ui.warn(_(b"editor exited with exit code %d\n") % ret)
continue
# Remove comment lines
patchfp = open(patchfn, 'rb')
ncpatchfp = stringio()
for line in patchfp:
line = util.fromnativeeol(line)
if not line.startswith(b'#'):
ncpatchfp.write(line)
patchfp.close()
ncpatchfp.seek(0)
newpatches = parsepatch(ncpatchfp)
finally:
os.unlink(patchfn)
del ncpatchfp
# Signal that the chunk shouldn't be applied as-is, but
# provide the new patch to be used instead.
ret = False
elif r == 3: # Skip
ret = skipfile = False
elif r == 4: # file (Record remaining)
ret = skipfile = True
elif r == 5: # done, skip remaining
ret = skipall = False
elif r == 6: # all
ret = skipall = True
elif r == 7: # quit
raise error.CanceledError(_(b'user quit'))
return ret, skipfile, skipall, newpatches
seen = set()
applied = {} # 'filename' -> [] of chunks
skipfile, skipall = None, None
pos, total = 1, sum(len(h.hunks) for h in headers)
for h in headers:
pos += len(h.hunks)
skipfile = None
fixoffset = 0
hdr = b''.join(h.header)
if hdr in seen:
continue
seen.add(hdr)
if skipall is None:
h.pretty(ui)
files = h.files()
msg = _(b'examine changes to %s?') % _(b' and ').join(
b"'%s'" % f for f in files
)
if all(match.exact(f) for f in files):
r, skipall, np = True, None, None
else:
r, skipfile, skipall, np = prompt(skipfile, skipall, msg, None)
if not r:
continue
applied[h.filename()] = [h]
if h.allhunks():
applied[h.filename()] += h.hunks
continue
for i, chunk in enumerate(h.hunks):
if skipfile is None and skipall is None:
chunk.pretty(ui)
if total == 1:
msg = messages[b'single'][operation] % chunk.filename()
else:
idx = pos - len(h.hunks) + i
msg = messages[b'multiple'][operation] % (
idx,
total,
chunk.filename(),
)
r, skipfile, skipall, newpatches = prompt(
skipfile, skipall, msg, chunk
)
if r:
if fixoffset:
chunk = copy.copy(chunk)
chunk.toline += fixoffset
applied[chunk.filename()].append(chunk)
elif newpatches is not None:
for newpatch in newpatches:
for newhunk in newpatch.hunks:
if fixoffset:
newhunk.toline += fixoffset
applied[newhunk.filename()].append(newhunk)
else:
fixoffset += chunk.removed - chunk.added
return (
sum(
[h for h in applied.values() if h[0].special() or len(h) > 1],
[],
),
{},
)
class hunk:
def __init__(self, desc, num, lr, context):
self.number = num
self.desc = desc
self.hunk = [desc]
self.a = []
self.b = []
self.starta = self.lena = None
self.startb = self.lenb = None
if lr is not None:
if context:
self.read_context_hunk(lr)
else:
self.read_unified_hunk(lr)
def getnormalized(self):
"""Return a copy with line endings normalized to LF."""
def normalize(lines):
nlines = []
for line in lines:
if line.endswith(b'\r\n'):
line = line[:-2] + b'\n'
nlines.append(line)
return nlines
# Dummy object, it is rebuilt manually
nh = hunk(self.desc, self.number, None, None)
nh.number = self.number
nh.desc = self.desc
nh.hunk = self.hunk
nh.a = normalize(self.a)
nh.b = normalize(self.b)
nh.starta = self.starta
nh.startb = self.startb
nh.lena = self.lena
nh.lenb = self.lenb
return nh
def read_unified_hunk(self, lr):
m = unidesc.match(self.desc)
if not m:
raise PatchParseError(_(b"bad hunk #%d") % self.number)
self.starta, self.lena, self.startb, self.lenb = m.groups()
if self.lena is None:
self.lena = 1
else:
self.lena = int(self.lena)
if self.lenb is None:
self.lenb = 1
else:
self.lenb = int(self.lenb)
self.starta = int(self.starta)
self.startb = int(self.startb)
try:
diffhelper.addlines(
lr, self.hunk, self.lena, self.lenb, self.a, self.b
)
except error.ParseError as e:
raise PatchParseError(_(b"bad hunk #%d: %s") % (self.number, e))
# if we hit eof before finishing out the hunk, the last line will
# be zero length. Lets try to fix it up.
while len(self.hunk[-1]) == 0:
del self.hunk[-1]
del self.a[-1]
del self.b[-1]
self.lena -= 1
self.lenb -= 1
self._fixnewline(lr)
def read_context_hunk(self, lr):
self.desc = lr.readline()
m = contextdesc.match(self.desc)
if not m:
raise PatchParseError(_(b"bad hunk #%d") % self.number)
self.starta, aend = m.groups()
self.starta = int(self.starta)
if aend is None:
aend = self.starta
self.lena = int(aend) - self.starta
if self.starta:
self.lena += 1
for x in range(self.lena):
l = lr.readline()
if l.startswith(b'---'):
# lines addition, old block is empty
lr.push(l)
break
s = l[2:]
if l.startswith(b'- ') or l.startswith(b'! '):
u = b'-' + s
elif l.startswith(b' '):
u = b' ' + s
else:
raise PatchParseError(
_(b"bad hunk #%d old text line %d") % (self.number, x)
)
self.a.append(u)
self.hunk.append(u)
l = lr.readline()
if l.startswith(br'\ '):
s = self.a[-1][:-1]
self.a[-1] = s
self.hunk[-1] = s
l = lr.readline()
m = contextdesc.match(l)
if not m:
raise PatchParseError(_(b"bad hunk #%d") % self.number)
self.startb, bend = m.groups()
self.startb = int(self.startb)
if bend is None:
bend = self.startb
self.lenb = int(bend) - self.startb
if self.startb:
self.lenb += 1
hunki = 1
for x in range(self.lenb):
l = lr.readline()
if l.startswith(br'\ '):
# XXX: the only way to hit this is with an invalid line range.
# The no-eol marker is not counted in the line range, but I
# guess there are diff(1) out there which behave differently.
s = self.b[-1][:-1]
self.b[-1] = s
self.hunk[hunki - 1] = s
continue
if not l:
# line deletions, new block is empty and we hit EOF
lr.push(l)
break
s = l[2:]
if l.startswith(b'+ ') or l.startswith(b'! '):
u = b'+' + s
elif l.startswith(b' '):
u = b' ' + s
elif len(self.b) == 0:
# line deletions, new block is empty
lr.push(l)
break
else:
raise PatchParseError(
_(b"bad hunk #%d old text line %d") % (self.number, x)
)
self.b.append(s)
while True:
if hunki >= len(self.hunk):
h = b""
else:
h = self.hunk[hunki]
hunki += 1
if h == u:
break
elif h.startswith(b'-'):
continue
else:
self.hunk.insert(hunki - 1, u)
break
if not self.a:
# this happens when lines were only added to the hunk
for x in self.hunk:
if x.startswith(b'-') or x.startswith(b' '):
self.a.append(x)
if not self.b:
# this happens when lines were only deleted from the hunk
for x in self.hunk:
if x.startswith(b'+') or x.startswith(b' '):
self.b.append(x[1:])
# @@ -start,len +start,len @@
self.desc = b"@@ -%d,%d +%d,%d @@\n" % (
self.starta,
self.lena,
self.startb,
self.lenb,
)
self.hunk[0] = self.desc
self._fixnewline(lr)
def _fixnewline(self, lr):
l = lr.readline()
if l.startswith(br'\ '):
diffhelper.fixnewline(self.hunk, self.a, self.b)
else:
lr.push(l)
def complete(self):
return len(self.a) == self.lena and len(self.b) == self.lenb
def _fuzzit(self, old, new, fuzz, toponly):
# this removes context lines from the top and bottom of list 'l'. It
# checks the hunk to make sure only context lines are removed, and then
# returns a new shortened list of lines.
fuzz = min(fuzz, len(old))
if fuzz:
top = 0
bot = 0
hlen = len(self.hunk)
for x in range(hlen - 1):
# the hunk starts with the @@ line, so use x+1
if self.hunk[x + 1].startswith(b' '):
top += 1
else:
break
if not toponly:
for x in range(hlen - 1):
if self.hunk[hlen - bot - 1].startswith(b' '):
bot += 1
else:
break
bot = min(fuzz, bot)
top = min(fuzz, top)
return old[top : len(old) - bot], new[top : len(new) - bot], top
return old, new, 0
def fuzzit(self, fuzz, toponly):
old, new, top = self._fuzzit(self.a, self.b, fuzz, toponly)
oldstart = self.starta + top
newstart = self.startb + top
# zero length hunk ranges already have their start decremented
if self.lena and oldstart > 0:
oldstart -= 1
if self.lenb and newstart > 0:
newstart -= 1
return old, oldstart, new, newstart
class binhunk:
"""A binary patch file."""
def __init__(self, lr, fname):
self.text = None
self.delta = False
self.hunk = [b'GIT binary patch\n']
self._fname = fname
self._read(lr)
def complete(self):
return self.text is not None
def new(self, lines):
if self.delta:
return [applybindelta(self.text, b''.join(lines))]
return [self.text]
def _read(self, lr):
def getline(lr, hunk):
l = lr.readline()
hunk.append(l)
return l.rstrip(b'\r\n')
while True:
line = getline(lr, self.hunk)
if not line:
raise PatchParseError(
_(b'could not extract "%s" binary data') % self._fname
)
if line.startswith(b'literal '):
size = int(line[8:].rstrip())
break
if line.startswith(b'delta '):
size = int(line[6:].rstrip())
self.delta = True
break
dec = []
line = getline(lr, self.hunk)
while len(line) > 1:
l = line[0:1]
if l <= b'Z' and l >= b'A':
l = ord(l) - ord(b'A') + 1
else:
l = ord(l) - ord(b'a') + 27
try:
dec.append(util.b85decode(line[1:])[:l])
except ValueError as e:
raise PatchParseError(
_(b'could not decode "%s" binary patch: %s')
% (self._fname, stringutil.forcebytestr(e))
)
line = getline(lr, self.hunk)
text = zlib.decompress(b''.join(dec))
if len(text) != size:
raise PatchParseError(
_(b'"%s" length is %d bytes, should be %d')
% (self._fname, len(text), size)
)
self.text = text
def parsefilename(str):
# --- filename \t|space stuff
s = str[4:].rstrip(b'\r\n')
i = s.find(b'\t')
if i < 0:
i = s.find(b' ')
if i < 0:
return s
return s[:i]
def reversehunks(hunks):
'''reverse the signs in the hunks given as argument
This function operates on hunks coming out of patch.filterpatch, that is
a list of the form: [header1, hunk1, hunk2, header2...]. Example usage:
>>> rawpatch = b"""diff --git a/folder1/g b/folder1/g
... --- a/folder1/g
... +++ b/folder1/g
... @@ -1,7 +1,7 @@
... +firstline
... c
... 1
... 2
... + 3
... -4
... 5
... d
... +lastline"""
>>> hunks = parsepatch([rawpatch])
>>> hunkscomingfromfilterpatch = []
>>> for h in hunks:
... hunkscomingfromfilterpatch.append(h)
... hunkscomingfromfilterpatch.extend(h.hunks)
>>> reversedhunks = reversehunks(hunkscomingfromfilterpatch)
>>> from . import util
>>> fp = util.stringio()
>>> for c in reversedhunks:
... c.write(fp)
>>> fp.seek(0) or None
>>> reversedpatch = fp.read()
>>> print(pycompat.sysstr(reversedpatch))
diff --git a/folder1/g b/folder1/g
--- a/folder1/g
+++ b/folder1/g
@@ -1,4 +1,3 @@
-firstline
c
1
2
@@ -2,6 +1,6 @@
c
1
2
- 3
+4
5
d
@@ -6,3 +5,2 @@
5
d
-lastline
'''
newhunks = []
for c in hunks:
if hasattr(c, 'reversehunk'):
c = c.reversehunk()
newhunks.append(c)
return newhunks
def parsepatch(originalchunks, maxcontext=None):
"""patch -> [] of headers -> [] of hunks
If maxcontext is not None, trim context lines if necessary.
>>> rawpatch = b'''diff --git a/folder1/g b/folder1/g
... --- a/folder1/g
... +++ b/folder1/g
... @@ -1,8 +1,10 @@
... 1
... 2
... -3
... 4
... 5
... 6
... +6.1
... +6.2
... 7
... 8
... +9'''
>>> out = util.stringio()
>>> headers = parsepatch([rawpatch], maxcontext=1)
>>> for header in headers:
... header.write(out)
... for hunk in header.hunks:
... hunk.write(out)
>>> print(pycompat.sysstr(out.getvalue()))
diff --git a/folder1/g b/folder1/g
--- a/folder1/g
+++ b/folder1/g
@@ -2,3 +2,2 @@
2
-3
4
@@ -6,2 +5,4 @@
6
+6.1
+6.2
7
@@ -8,1 +9,2 @@
8
+9
"""
class parser:
"""patch parsing state machine"""
def __init__(self):
self.fromline = 0
self.toline = 0
self.proc = b''
self.header = None
self.context = []
self.before = []
self.hunk = []
self.headers = []
def addrange(self, limits):
self.addcontext([])
fromstart, fromend, tostart, toend, proc = limits
self.fromline = int(fromstart)
self.toline = int(tostart)
self.proc = proc
def addcontext(self, context):
if self.hunk:
h = recordhunk(
self.header,
self.fromline,
self.toline,
self.proc,
self.before,
self.hunk,
context,
maxcontext,
)
self.header.hunks.append(h)
self.fromline += len(self.before) + h.removed
self.toline += len(self.before) + h.added
self.before = []
self.hunk = []
self.context = context
def addhunk(self, hunk):
if self.context:
self.before = self.context
self.context = []
if self.hunk:
self.addcontext([])
self.hunk = hunk
def newfile(self, hdr):
self.addcontext([])
h = header(hdr)
self.headers.append(h)
self.header = h
def addother(self, line):
pass # 'other' lines are ignored
def finished(self):
self.addcontext([])
return self.headers
transitions = {
b'file': {
b'context': addcontext,
b'file': newfile,
b'hunk': addhunk,
b'range': addrange,
},
b'context': {
b'file': newfile,
b'hunk': addhunk,
b'range': addrange,
b'other': addother,
},
b'hunk': {
b'context': addcontext,
b'file': newfile,
b'range': addrange,
},
b'range': {b'context': addcontext, b'hunk': addhunk},
b'other': {b'other': addother},
}
p = parser()
fp = stringio()
fp.write(b''.join(originalchunks))
fp.seek(0)
state = b'context'
for newstate, data in scanpatch(fp):
try:
p.transitions[state][newstate](p, data)
except KeyError:
raise PatchParseError(
b'unhandled transition: %s -> %s' % (state, newstate)
)
state = newstate
del fp
return p.finished()
def pathtransform(path, strip, prefix):
"""turn a path from a patch into a path suitable for the repository
prefix, if not empty, is expected to be normalized with a / at the end.
Returns (stripped components, path in repository).
>>> pathtransform(b'a/b/c', 0, b'')
('', 'a/b/c')
>>> pathtransform(b' a/b/c ', 0, b'')
('', ' a/b/c')
>>> pathtransform(b' a/b/c ', 2, b'')
('a/b/', 'c')
>>> pathtransform(b'a/b/c', 0, b'd/e/')
('', 'd/e/a/b/c')
>>> pathtransform(b' a//b/c ', 2, b'd/e/')
('a//b/', 'd/e/c')
>>> pathtransform(b'a/b/c', 3, b'')
Traceback (most recent call last):
PatchApplicationError: unable to strip away 1 of 3 dirs from a/b/c
"""
pathlen = len(path)
i = 0
if strip == 0:
return b'', prefix + path.rstrip()
count = strip
while count > 0:
i = path.find(b'/', i)
if i == -1:
raise PatchApplicationError(
_(b"unable to strip away %d of %d dirs from %s")
% (count, strip, path)
)
i += 1
# consume '//' in the path
while i < pathlen - 1 and path[i : i + 1] == b'/':
i += 1
count -= 1
return path[:i].lstrip(), prefix + path[i:].rstrip()
def makepatchmeta(backend, afile_orig, bfile_orig, hunk, strip, prefix):
nulla = afile_orig == b"/dev/null"
nullb = bfile_orig == b"/dev/null"
create = nulla and hunk.starta == 0 and hunk.lena == 0
remove = nullb and hunk.startb == 0 and hunk.lenb == 0
abase, afile = pathtransform(afile_orig, strip, prefix)
gooda = not nulla and backend.exists(afile)
bbase, bfile = pathtransform(bfile_orig, strip, prefix)
if afile == bfile:
goodb = gooda
else:
goodb = not nullb and backend.exists(bfile)
missing = not goodb and not gooda and not create
# some diff programs apparently produce patches where the afile is
# not /dev/null, but afile starts with bfile
abasedir = afile[: afile.rfind(b'/') + 1]
bbasedir = bfile[: bfile.rfind(b'/') + 1]
if (
missing
and abasedir == bbasedir
and afile.startswith(bfile)
and hunk.starta == 0
and hunk.lena == 0
):
create = True
missing = False
# If afile is "a/b/foo" and bfile is "a/b/foo.orig" we assume the
# diff is between a file and its backup. In this case, the original
# file should be patched (see original mpatch code).
isbackup = abase == bbase and bfile.startswith(afile)
fname = None
if not missing:
if gooda and goodb:
if isbackup:
fname = afile
else:
fname = bfile
elif gooda:
fname = afile
if not fname:
if not nullb:
if isbackup:
fname = afile
else:
fname = bfile
elif not nulla:
fname = afile
else:
raise PatchParseError(_(b"undefined source and destination files"))
gp = patchmeta(fname)
if create:
gp.op = b'ADD'
elif remove:
gp.op = b'DELETE'
return gp
def scanpatch(fp):
"""like patch.iterhunks, but yield different events
- ('file', [header_lines + fromfile + tofile])
- ('context', [context_lines])
- ('hunk', [hunk_lines])
- ('range', (-start,len, +start,len, proc))
"""
lines_re = re.compile(br'@@ -(\d+),(\d+) \+(\d+),(\d+) @@\s*(.*)')
lr = linereader(fp)
def scanwhile(first, p):
"""scan lr while predicate holds"""
lines = [first]
for line in iter(lr.readline, b''):
if p(line):
lines.append(line)
else:
lr.push(line)
break
return lines
for line in iter(lr.readline, b''):
if line.startswith(b'diff --git a/') or line.startswith(b'diff -r '):
def notheader(line):
s = line.split(None, 1)
return not s or s[0] not in (b'---', b'diff')
header = scanwhile(line, notheader)
fromfile = lr.readline()
if fromfile.startswith(b'---'):
tofile = lr.readline()
header += [fromfile, tofile]
else:
lr.push(fromfile)
yield b'file', header
elif line.startswith(b' '):
cs = (b' ', b'\\')
yield b'context', scanwhile(line, lambda l: l.startswith(cs))
elif line.startswith((b'-', b'+')):
cs = (b'-', b'+', b'\\')
yield b'hunk', scanwhile(line, lambda l: l.startswith(cs))
else:
m = lines_re.match(line)
if m:
yield b'range', m.groups()
else:
yield b'other', line
def scangitpatch(lr, firstline):
"""
Git patches can emit:
- rename a to b
- change b
- copy a to c
- change c
We cannot apply this sequence as-is, the renamed 'a' could not be
found for it would have been renamed already. And we cannot copy
from 'b' instead because 'b' would have been changed already. So
we scan the git patch for copy and rename commands so we can
perform the copies ahead of time.
"""
pos = 0
try:
pos = lr.fp.tell()
fp = lr.fp
except IOError:
fp = stringio(lr.fp.read())
gitlr = linereader(fp)
gitlr.push(firstline)
gitpatches = readgitpatch(gitlr)
fp.seek(pos)
return gitpatches
def iterhunks(fp):
"""Read a patch and yield the following events:
- ("file", afile, bfile, firsthunk): select a new target file.
- ("hunk", hunk): a new hunk is ready to be applied, follows a
"file" event.
- ("git", gitchanges): current diff is in git format, gitchanges
maps filenames to gitpatch records. Unique event.
"""
afile = b""
bfile = b""
state = None
hunknum = 0
emitfile = newfile = False
gitpatches = None
# our states
BFILE = 1
context = None
lr = linereader(fp)
for x in iter(lr.readline, b''):
if state == BFILE and (
(not context and x.startswith(b'@'))
or (context is not False and x.startswith(b'***************'))
or x.startswith(b'GIT binary patch')
):
gp = None
if gitpatches and gitpatches[-1].ispatching(afile, bfile):
gp = gitpatches.pop()
if x.startswith(b'GIT binary patch'):
h = binhunk(lr, gp.path)
else:
if context is None and x.startswith(b'***************'):
context = True
h = hunk(x, hunknum + 1, lr, context)
hunknum += 1
if emitfile:
emitfile = False
yield b'file', (afile, bfile, h, gp and gp.copy() or None)
yield b'hunk', h
elif x.startswith(b'diff --git a/'):
m = gitre.match(x.rstrip(b'\r\n'))
if not m:
continue
if gitpatches is None:
# scan whole input for git metadata
gitpatches = scangitpatch(lr, x)
yield b'git', [
g.copy() for g in gitpatches if g.op in (b'COPY', b'RENAME')
]
gitpatches.reverse()
afile = b'a/' + m.group(1)
bfile = b'b/' + m.group(2)
while gitpatches and not gitpatches[-1].ispatching(afile, bfile):
gp = gitpatches.pop()
yield b'file', (
b'a/' + gp.path,
b'b/' + gp.path,
None,
gp.copy(),
)
if not gitpatches:
raise PatchParseError(
_(b'failed to synchronize metadata for "%s"') % afile[2:]
)
newfile = True
elif x.startswith(b'---'):
# check for a unified diff
l2 = lr.readline()
if not l2.startswith(b'+++'):
lr.push(l2)
continue
newfile = True
context = False
afile = parsefilename(x)
bfile = parsefilename(l2)
elif x.startswith(b'***'):
# check for a context diff
l2 = lr.readline()
if not l2.startswith(b'---'):
lr.push(l2)
continue
l3 = lr.readline()
lr.push(l3)
if not l3.startswith(b"***************"):
lr.push(l2)
continue
newfile = True
context = True
afile = parsefilename(x)
bfile = parsefilename(l2)
if newfile:
newfile = False
emitfile = True
state = BFILE
hunknum = 0
while gitpatches:
gp = gitpatches.pop()
yield b'file', (b'a/' + gp.path, b'b/' + gp.path, None, gp.copy())
def applybindelta(binchunk, data):
"""Apply a binary delta hunk
The algorithm used is the algorithm from git's patch-delta.c
"""
def deltahead(binchunk):
i = 0
for c in pycompat.bytestr(binchunk):
i += 1
if not (ord(c) & 0x80):
return i
return i
out = b""
s = deltahead(binchunk)
binchunk = binchunk[s:]
s = deltahead(binchunk)
binchunk = binchunk[s:]
i = 0
while i < len(binchunk):
cmd = ord(binchunk[i : i + 1])
i += 1
if cmd & 0x80:
offset = 0
size = 0
if cmd & 0x01:
offset = ord(binchunk[i : i + 1])
i += 1
if cmd & 0x02:
offset |= ord(binchunk[i : i + 1]) << 8
i += 1
if cmd & 0x04:
offset |= ord(binchunk[i : i + 1]) << 16
i += 1
if cmd & 0x08:
offset |= ord(binchunk[i : i + 1]) << 24
i += 1
if cmd & 0x10:
size = ord(binchunk[i : i + 1])
i += 1
if cmd & 0x20:
size |= ord(binchunk[i : i + 1]) << 8
i += 1
if cmd & 0x40:
size |= ord(binchunk[i : i + 1]) << 16
i += 1
if size == 0:
size = 0x10000
offset_end = offset + size
out += data[offset:offset_end]
elif cmd != 0:
offset_end = i + cmd
out += binchunk[i:offset_end]
i += cmd
else:
raise PatchApplicationError(_(b'unexpected delta opcode 0'))
return out
def applydiff(ui, fp, backend, store, strip=1, prefix=b'', eolmode=b'strict'):
"""Reads a patch from fp and tries to apply it.
Returns 0 for a clean patch, -1 if any rejects were found and 1 if
there was any fuzz.
If 'eolmode' is 'strict', the patch content and patched file are
read in binary mode. Otherwise, line endings are ignored when
patching then normalized according to 'eolmode'.
"""
return _applydiff(
ui,
fp,
patchfile,
backend,
store,
strip=strip,
prefix=prefix,
eolmode=eolmode,
)
def _canonprefix(repo, prefix):
if prefix:
prefix = pathutil.canonpath(repo.root, repo.getcwd(), prefix)
if prefix != b'':
prefix += b'/'
return prefix
def _applydiff(
ui, fp, patcher, backend, store, strip=1, prefix=b'', eolmode=b'strict'
):
prefix = _canonprefix(backend.repo, prefix)
def pstrip(p):
return pathtransform(p, strip - 1, prefix)[1]
rejects = 0
err = 0
current_file = None
for state, values in iterhunks(fp):
if state == b'hunk':
if not current_file:
continue
ret = current_file.apply(values)
if ret > 0:
err = 1
elif state == b'file':
if current_file:
rejects += current_file.close()
current_file = None
afile, bfile, first_hunk, gp = values
if gp:
gp.path = pstrip(gp.path)
if gp.oldpath:
gp.oldpath = pstrip(gp.oldpath)
else:
gp = makepatchmeta(
backend, afile, bfile, first_hunk, strip, prefix
)
if gp.op == b'RENAME':
backend.unlink(gp.oldpath)
if not first_hunk:
if gp.op == b'DELETE':
backend.unlink(gp.path)
continue
data, mode = None, None
if gp.op in (b'RENAME', b'COPY'):
data, mode = store.getfile(gp.oldpath)[:2]
if data is None:
# This means that the old path does not exist
raise PatchApplicationError(
_(b"source file '%s' does not exist") % gp.oldpath
)
if gp.mode:
mode = gp.mode
if gp.op == b'ADD':
# Added files without content have no hunk and
# must be created
data = b''
if data or mode:
if gp.op in (b'ADD', b'RENAME', b'COPY') and backend.exists(
gp.path
):
raise PatchApplicationError(
_(
b"cannot create %s: destination "
b"already exists"
)
% gp.path
)
backend.setfile(gp.path, data, mode, gp.oldpath)
continue
try:
current_file = patcher(ui, gp, backend, store, eolmode=eolmode)
except PatchError as inst:
ui.warn(stringutil.forcebytestr(inst) + b'\n')
current_file = None
rejects += 1
continue
elif state == b'git':
for gp in values:
path = pstrip(gp.oldpath)
data, mode = backend.getfile(path)
if data is None:
# The error ignored here will trigger a getfile()
# error in a place more appropriate for error
# handling, and will not interrupt the patching
# process.
pass
else:
store.setfile(path, data, mode)
else:
raise error.Abort(_(b'unsupported parser state: %s') % state)
if current_file:
rejects += current_file.close()
if rejects:
return -1
return err
def _externalpatch(ui, repo, patcher, patchname, strip, files, similarity):
"""use <patcher> to apply <patchname> to the working directory.
returns whether patch was applied with fuzz factor."""
fuzz = False
args = []
cwd = repo.root
if cwd:
args.append(b'-d %s' % procutil.shellquote(cwd))
cmd = b'%s %s -p%d < %s' % (
patcher,
b' '.join(args),
strip,
procutil.shellquote(patchname),
)
ui.debug(b'Using external patch tool: %s\n' % cmd)
fp = procutil.popen(cmd, b'rb')
try:
for line in fp:
line = line.rstrip()
ui.note(line + b'\n')
if line.startswith(b'patching file '):
pf = util.parsepatchoutput(line)
printed_file = False
files.add(pf)
elif line.find(b'with fuzz') >= 0:
fuzz = True
if not printed_file:
ui.warn(pf + b'\n')
printed_file = True
ui.warn(line + b'\n')
elif line.find(b'saving rejects to file') >= 0:
ui.warn(line + b'\n')
elif line.find(b'FAILED') >= 0:
if not printed_file:
ui.warn(pf + b'\n')
printed_file = True
ui.warn(line + b'\n')
finally:
if files:
scmutil.marktouched(repo, files, similarity)
code = fp.close()
if code:
raise PatchApplicationError(
_(b"patch command failed: %s") % procutil.explainexit(code)
)
return fuzz
def patchbackend(
ui, backend, patchobj, strip, prefix, files=None, eolmode=b'strict'
):
if files is None:
files = set()
if eolmode is None:
eolmode = ui.config(b'patch', b'eol')
if eolmode.lower() not in eolmodes:
raise error.Abort(_(b'unsupported line endings type: %s') % eolmode)
eolmode = eolmode.lower()
store = filestore()
try:
fp = open(patchobj, b'rb')
except TypeError:
fp = patchobj
try:
ret = applydiff(
ui, fp, backend, store, strip=strip, prefix=prefix, eolmode=eolmode
)
finally:
if fp != patchobj:
fp.close()
files.update(backend.close())
store.close()
if ret < 0:
raise PatchApplicationError(_(b'patch failed to apply'))
return ret > 0
def internalpatch(
ui,
repo,
patchobj,
strip,
prefix=b'',
files=None,
eolmode=b'strict',
similarity=0,
):
"""use builtin patch to apply <patchobj> to the working directory.
returns whether patch was applied with fuzz factor."""
backend = workingbackend(ui, repo, similarity)
return patchbackend(ui, backend, patchobj, strip, prefix, files, eolmode)
def patchrepo(
ui, repo, ctx, store, patchobj, strip, prefix, files=None, eolmode=b'strict'
):
backend = repobackend(ui, repo, ctx, store)
return patchbackend(ui, backend, patchobj, strip, prefix, files, eolmode)
def patch(
ui,
repo,
patchname,
strip=1,
prefix=b'',
files=None,
eolmode=b'strict',
similarity=0,
):
"""Apply <patchname> to the working directory.
'eolmode' specifies how end of lines should be handled. It can be:
- 'strict': inputs are read in binary mode, EOLs are preserved
- 'crlf': EOLs are ignored when patching and reset to CRLF
- 'lf': EOLs are ignored when patching and reset to LF
- None: get it from user settings, default to 'strict'
'eolmode' is ignored when using an external patcher program.
Returns whether patch was applied with fuzz factor.
"""
patcher = ui.config(b'ui', b'patch')
if files is None:
files = set()
if patcher:
return _externalpatch(
ui, repo, patcher, patchname, strip, files, similarity
)
return internalpatch(
ui, repo, patchname, strip, prefix, files, eolmode, similarity
)
def changedfiles(ui, repo, patchpath, strip=1, prefix=b''):
backend = fsbackend(ui, repo.root)
prefix = _canonprefix(repo, prefix)
with open(patchpath, b'rb') as fp:
changed = set()
for state, values in iterhunks(fp):
if state == b'file':
afile, bfile, first_hunk, gp = values
if gp:
gp.path = pathtransform(gp.path, strip - 1, prefix)[1]
if gp.oldpath:
gp.oldpath = pathtransform(
gp.oldpath, strip - 1, prefix
)[1]
else:
gp = makepatchmeta(
backend, afile, bfile, first_hunk, strip, prefix
)
changed.add(gp.path)
if gp.op == b'RENAME':
changed.add(gp.oldpath)
elif state not in (b'hunk', b'git'):
raise error.Abort(_(b'unsupported parser state: %s') % state)
return changed
class GitDiffRequired(Exception):
pass
diffopts = diffutil.diffallopts
diffallopts = diffutil.diffallopts
difffeatureopts = diffutil.difffeatureopts
def diff(
repo,
node1=None,
node2=None,
match=None,
changes=None,
opts=None,
losedatafn=None,
pathfn=None,
copy=None,
copysourcematch=None,
hunksfilterfn=None,
):
"""yields diff of changes to files between two nodes, or node and
working directory.
if node1 is None, use first dirstate parent instead.
if node2 is None, compare node1 with working directory.
losedatafn(**kwarg) is a callable run when opts.upgrade=True and
every time some change cannot be represented with the current
patch format. Return False to upgrade to git patch format, True to
accept the loss or raise an exception to abort the diff. It is
called with the name of current file being diffed as 'fn'. If set
to None, patches will always be upgraded to git format when
necessary.
prefix is a filename prefix that is prepended to all filenames on
display (used for subrepos).
relroot, if not empty, must be normalized with a trailing /. Any match
patterns that fall outside it will be ignored.
copy, if not empty, should contain mappings {dst@y: src@x} of copy
information.
if copysourcematch is not None, then copy sources will be filtered by this
matcher
hunksfilterfn, if not None, should be a function taking a filectx and
hunks generator that may yield filtered hunks.
"""
if not node1 and not node2:
node1 = repo.dirstate.p1()
ctx1 = repo[node1]
ctx2 = repo[node2]
for fctx1, fctx2, hdr, hunks in diffhunks(
repo,
ctx1=ctx1,
ctx2=ctx2,
match=match,
changes=changes,
opts=opts,
losedatafn=losedatafn,
pathfn=pathfn,
copy=copy,
copysourcematch=copysourcematch,
):
if hunksfilterfn is not None:
# If the file has been removed, fctx2 is None; but this should
# not occur here since we catch removed files early in
# logcmdutil.getlinerangerevs() for 'hg log -L'.
assert (
fctx2 is not None
), b'fctx2 unexpectly None in diff hunks filtering'
hunks = hunksfilterfn(fctx2, hunks)
text = b''.join(b''.join(hlines) for hrange, hlines in hunks)
if hdr and (text or len(hdr) > 1):
yield b'\n'.join(hdr) + b'\n'
if text:
yield text
def diffhunks(
repo,
ctx1,
ctx2,
match=None,
changes=None,
opts=None,
losedatafn=None,
pathfn=None,
copy=None,
copysourcematch=None,
):
"""Yield diff of changes to files in the form of (`header`, `hunks`) tuples
where `header` is a list of diff headers and `hunks` is an iterable of
(`hunkrange`, `hunklines`) tuples.
See diff() for the meaning of parameters.
"""
if opts is None:
opts = mdiff.defaultopts
def lrugetfilectx():
cache = {}
order = collections.deque()
def getfilectx(f, ctx):
fctx = ctx.filectx(f, filelog=cache.get(f))
if f not in cache:
if len(cache) > 20:
del cache[order.popleft()]
cache[f] = fctx.filelog()
else:
order.remove(f)
order.append(f)
return fctx
return getfilectx
getfilectx = lrugetfilectx()
if not changes:
changes = ctx1.status(ctx2, match=match)
if isinstance(changes, list):
modified, added, removed = changes[:3]
else:
modified, added, removed = (
changes.modified,
changes.added,
changes.removed,
)
if not modified and not added and not removed:
return []
if repo.ui.debugflag:
hexfunc = hex
else:
hexfunc = short
revs = [hexfunc(node) for node in [ctx1.node(), ctx2.node()] if node]
if copy is None:
copy = {}
if opts.git or opts.upgrade:
copy = copies.pathcopies(ctx1, ctx2, match=match)
if copysourcematch:
# filter out copies where source side isn't inside the matcher
# (copies.pathcopies() already filtered out the destination)
copy = {dst: src for dst, src in copy.items() if copysourcematch(src)}
modifiedset = set(modified)
addedset = set(added)
removedset = set(removed)
for f in modified:
if f not in ctx1:
# Fix up added, since merged-in additions appear as
# modifications during merges
modifiedset.remove(f)
addedset.add(f)
for f in removed:
if f not in ctx1:
# Merged-in additions that are then removed are reported as removed.
# They are not in ctx1, so We don't want to show them in the diff.
removedset.remove(f)
modified = sorted(modifiedset)
added = sorted(addedset)
removed = sorted(removedset)
for dst, src in list(copy.items()):
if src not in ctx1:
# Files merged in during a merge and then copied/renamed are
# reported as copies. We want to show them in the diff as additions.
del copy[dst]
prefetchmatch = scmutil.matchfiles(
repo, list(modifiedset | addedset | removedset)
)
revmatches = [
(ctx1.rev(), prefetchmatch),
(ctx2.rev(), prefetchmatch),
]
scmutil.prefetchfiles(repo, revmatches)
def difffn(opts, losedata):
return trydiff(
repo,
revs,
ctx1,
ctx2,
modified,
added,
removed,
copy,
getfilectx,
opts,
losedata,
pathfn,
)
if opts.upgrade and not opts.git:
try:
def losedata(fn):
if not losedatafn or not losedatafn(fn=fn):
raise GitDiffRequired
# Buffer the whole output until we are sure it can be generated
return list(difffn(opts.copy(git=False), losedata))
except GitDiffRequired:
return difffn(opts.copy(git=True), None)
else:
return difffn(opts, None)
def diffsinglehunk(hunklines):
"""yield tokens for a list of lines in a single hunk"""
for line in hunklines:
# chomp
chompline = line.rstrip(b'\r\n')
# highlight tabs and trailing whitespace
stripline = chompline.rstrip()
if line.startswith(b'-'):
label = b'diff.deleted'
elif line.startswith(b'+'):
label = b'diff.inserted'
else:
raise error.ProgrammingError(b'unexpected hunk line: %s' % line)
for token in tabsplitter.findall(stripline):
if token.startswith(b'\t'):
yield (token, b'diff.tab')
else:
yield (token, label)
if chompline != stripline:
yield (chompline[len(stripline) :], b'diff.trailingwhitespace')
if chompline != line:
yield (line[len(chompline) :], b'')
def diffsinglehunkinline(hunklines):
"""yield tokens for a list of lines in a single hunk, with inline colors"""
# prepare deleted, and inserted content
a = bytearray()
b = bytearray()
for line in hunklines:
if line[0:1] == b'-':
a += line[1:]
elif line[0:1] == b'+':
b += line[1:]
else:
raise error.ProgrammingError(b'unexpected hunk line: %s' % line)
# fast path: if either side is empty, use diffsinglehunk
if not a or not b:
for t in diffsinglehunk(hunklines):
yield t
return
# re-split the content into words
al = wordsplitter.findall(bytes(a))
bl = wordsplitter.findall(bytes(b))
# re-arrange the words to lines since the diff algorithm is line-based
aln = [s if s == b'\n' else s + b'\n' for s in al]
bln = [s if s == b'\n' else s + b'\n' for s in bl]
an = b''.join(aln)
bn = b''.join(bln)
# run the diff algorithm, prepare atokens and btokens
atokens = []
btokens = []
blocks = mdiff.allblocks(an, bn, lines1=aln, lines2=bln)
for (a1, a2, b1, b2), btype in blocks:
changed = btype == b'!'
for token in mdiff.splitnewlines(b''.join(al[a1:a2])):
atokens.append((changed, token))
for token in mdiff.splitnewlines(b''.join(bl[b1:b2])):
btokens.append((changed, token))
# yield deleted tokens, then inserted ones
for prefix, label, tokens in [
(b'-', b'diff.deleted', atokens),
(b'+', b'diff.inserted', btokens),
]:
nextisnewline = True
for changed, token in tokens:
if nextisnewline:
yield (prefix, label)
nextisnewline = False
# special handling line end
isendofline = token.endswith(b'\n')
if isendofline:
chomp = token[:-1] # chomp
if chomp.endswith(b'\r'):
chomp = chomp[:-1]
endofline = token[len(chomp) :]
token = chomp.rstrip() # detect spaces at the end
endspaces = chomp[len(token) :]
# scan tabs
for maybetab in tabsplitter.findall(token):
if b'\t' == maybetab[0:1]:
currentlabel = b'diff.tab'
else:
if changed:
currentlabel = label + b'.changed'
else:
currentlabel = label + b'.unchanged'
yield (maybetab, currentlabel)
if isendofline:
if endspaces:
yield (endspaces, b'diff.trailingwhitespace')
yield (endofline, b'')
nextisnewline = True
def difflabel(func, *args, **kw):
'''yields 2-tuples of (output, label) based on the output of func()'''
if kw.get('opts') and kw['opts'].worddiff:
dodiffhunk = diffsinglehunkinline
else:
dodiffhunk = diffsinglehunk
headprefixes = [
(b'diff', b'diff.diffline'),
(b'copy', b'diff.extended'),
(b'rename', b'diff.extended'),
(b'old', b'diff.extended'),
(b'new', b'diff.extended'),
(b'deleted', b'diff.extended'),
(b'index', b'diff.extended'),
(b'similarity', b'diff.extended'),
(b'---', b'diff.file_a'),
(b'+++', b'diff.file_b'),
]
textprefixes = [
(b'@', b'diff.hunk'),
# - and + are handled by diffsinglehunk
]
head = False
# buffers a hunk, i.e. adjacent "-", "+" lines without other changes.
hunkbuffer = []
def consumehunkbuffer():
if hunkbuffer:
for token in dodiffhunk(hunkbuffer):
yield token
hunkbuffer[:] = []
for chunk in func(*args, **kw):
lines = chunk.split(b'\n')
linecount = len(lines)
for i, line in enumerate(lines):
if head:
if line.startswith(b'@'):
head = False
else:
if line and not line.startswith(
(b' ', b'+', b'-', b'@', b'\\')
):
head = True
diffline = False
if not head and line and line.startswith((b'+', b'-')):
diffline = True
prefixes = textprefixes
if head:
prefixes = headprefixes
if diffline:
# buffered
bufferedline = line
if i + 1 < linecount:
bufferedline += b"\n"
hunkbuffer.append(bufferedline)
else:
# unbuffered
for token in consumehunkbuffer():
yield token
stripline = line.rstrip()
for prefix, label in prefixes:
if stripline.startswith(prefix):
yield (stripline, label)
if line != stripline:
yield (
line[len(stripline) :],
b'diff.trailingwhitespace',
)
break
else:
yield (line, b'')
if i + 1 < linecount:
yield (b'\n', b'')
for token in consumehunkbuffer():
yield token
def diffui(*args, **kw):
'''like diff(), but yields 2-tuples of (output, label) for ui.write()'''
return difflabel(diff, *args, **kw)
def _filepairs(modified, added, removed, copy, opts):
"""generates tuples (f1, f2, copyop), where f1 is the name of the file
before and f2 is the the name after. For added files, f1 will be None,
and for removed files, f2 will be None. copyop may be set to None, 'copy'
or 'rename' (the latter two only if opts.git is set)."""
gone = set()
copyto = {v: k for k, v in copy.items()}
addedset, removedset = set(added), set(removed)
for f in sorted(modified + added + removed):
copyop = None
f1, f2 = f, f
if f in addedset:
f1 = None
if f in copy:
if opts.git:
f1 = copy[f]
if f1 in removedset and f1 not in gone:
copyop = b'rename'
gone.add(f1)
else:
copyop = b'copy'
elif f in removedset:
f2 = None
if opts.git:
# have we already reported a copy above?
if (
f in copyto
and copyto[f] in addedset
and copy[copyto[f]] == f
):
continue
yield f1, f2, copyop
def _gitindex(text):
if not text:
text = b""
l = len(text)
s = hashutil.sha1(b'blob %d\0' % l)
s.update(text)
return hex(s.digest())
_gitmode = {b'l': b'120000', b'x': b'100755', b'': b'100644'}
def trydiff(
repo,
revs,
ctx1,
ctx2,
modified,
added,
removed,
copy,
getfilectx,
opts,
losedatafn,
pathfn,
):
"""given input data, generate a diff and yield it in blocks
If generating a diff would lose data like flags or binary data and
losedatafn is not None, it will be called.
pathfn is applied to every path in the diff output.
"""
if opts.noprefix:
aprefix = bprefix = b''
else:
aprefix = b'a/'
bprefix = b'b/'
def diffline(f, revs):
revinfo = b' '.join([b"-r %s" % rev for rev in revs])
return b'diff %s %s' % (revinfo, f)
def isempty(fctx):
return fctx is None or fctx.size() == 0
date1 = dateutil.datestr(ctx1.date())
date2 = dateutil.datestr(ctx2.date())
if not pathfn:
pathfn = lambda f: f
for f1, f2, copyop in _filepairs(modified, added, removed, copy, opts):
content1 = None
content2 = None
fctx1 = None
fctx2 = None
flag1 = None
flag2 = None
if f1:
fctx1 = getfilectx(f1, ctx1)
if opts.git or losedatafn:
flag1 = ctx1.flags(f1)
if f2:
fctx2 = getfilectx(f2, ctx2)
if opts.git or losedatafn:
flag2 = ctx2.flags(f2)
# if binary is True, output "summary" or "base85", but not "text diff"
if opts.text:
binary = False
else:
binary = any(f.isbinary() for f in [fctx1, fctx2] if f is not None)
if losedatafn and not opts.git:
if (
binary
or
# copy/rename
f2 in copy
or
# empty file creation
(not f1 and isempty(fctx2))
or
# empty file deletion
(isempty(fctx1) and not f2)
or
# create with flags
(not f1 and flag2)
or
# change flags
(f1 and f2 and flag1 != flag2)
):
losedatafn(f2 or f1)
path1 = pathfn(f1 or f2)
path2 = pathfn(f2 or f1)
header = []
if opts.git:
header.append(
b'diff --git %s%s %s%s' % (aprefix, path1, bprefix, path2)
)
if not f1: # added
header.append(b'new file mode %s' % _gitmode[flag2])
elif not f2: # removed
header.append(b'deleted file mode %s' % _gitmode[flag1])
else: # modified/copied/renamed
mode1, mode2 = _gitmode[flag1], _gitmode[flag2]
if mode1 != mode2:
header.append(b'old mode %s' % mode1)
header.append(b'new mode %s' % mode2)
if copyop is not None:
if opts.showsimilarity:
sim = similar.score(ctx1[path1], ctx2[path2]) * 100
header.append(b'similarity index %d%%' % sim)
header.append(b'%s from %s' % (copyop, path1))
header.append(b'%s to %s' % (copyop, path2))
elif revs:
header.append(diffline(path1, revs))
# fctx.is | diffopts | what to | is fctx.data()
# binary() | text nobinary git index | output? | outputted?
# ------------------------------------|----------------------------
# yes | no no no * | summary | no
# yes | no no yes * | base85 | yes
# yes | no yes no * | summary | no
# yes | no yes yes 0 | summary | no
# yes | no yes yes >0 | summary | semi [1]
# yes | yes * * * | text diff | yes
# no | * * * * | text diff | yes
# [1]: hash(fctx.data()) is outputted. so fctx.data() cannot be faked
if binary and (
not opts.git or (opts.git and opts.nobinary and not opts.index)
):
# fast path: no binary content will be displayed, content1 and
# content2 are only used for equivalent test. cmp() could have a
# fast path.
if fctx1 is not None:
content1 = b'\0'
if fctx2 is not None:
if fctx1 is not None and not fctx1.cmp(fctx2):
content2 = b'\0' # not different
else:
content2 = b'\0\0'
else:
# normal path: load contents
if fctx1 is not None:
content1 = fctx1.data()
if fctx2 is not None:
content2 = fctx2.data()
data1 = (ctx1, fctx1, path1, flag1, content1, date1)
data2 = (ctx2, fctx2, path2, flag2, content2, date2)
yield diffcontent(data1, data2, header, binary, opts)
def diffcontent(data1, data2, header, binary, opts):
"""diffs two versions of a file.
data1 and data2 are tuples containg:
* ctx: changeset for the file
* fctx: file context for that file
* path1: name of the file
* flag: flags of the file
* content: full content of the file (can be null in case of binary)
* date: date of the changeset
header: the patch header
binary: whether the any of the version of file is binary or not
opts: user passed options
It exists as a separate function so that extensions like extdiff can wrap
it and use the file content directly.
"""
ctx1, fctx1, path1, flag1, content1, date1 = data1
ctx2, fctx2, path2, flag2, content2, date2 = data2
index1 = _gitindex(content1) if path1 in ctx1 else sha1nodeconstants.nullhex
index2 = _gitindex(content2) if path2 in ctx2 else sha1nodeconstants.nullhex
if binary and opts.git and not opts.nobinary:
text = mdiff.b85diff(content1, content2)
if text:
header.append(b'index %s..%s' % (index1, index2))
hunks = ((None, [text]),)
else:
if opts.git and opts.index > 0:
flag = flag1
if flag is None:
flag = flag2
header.append(
b'index %s..%s %s'
% (
index1[0 : opts.index],
index2[0 : opts.index],
_gitmode[flag],
)
)
uheaders, hunks = mdiff.unidiff(
content1,
date1,
content2,
date2,
path1,
path2,
binary=binary,
opts=opts,
)
header.extend(uheaders)
return fctx1, fctx2, header, hunks
def diffstatsum(stats):
maxfile, maxtotal, addtotal, removetotal, binary = 0, 0, 0, 0, False
for f, a, r, b in stats:
maxfile = max(maxfile, encoding.colwidth(f))
maxtotal = max(maxtotal, a + r)
addtotal += a
removetotal += r
binary = binary or b
return maxfile, maxtotal, addtotal, removetotal, binary
def diffstatdata(lines):
diffre = re.compile(br'^diff .*-r [a-z0-9]+\s(.*)$')
results = []
filename, adds, removes, isbinary = None, 0, 0, False
def addresult():
if filename:
results.append((filename, adds, removes, isbinary))
# inheader is used to track if a line is in the
# header portion of the diff. This helps properly account
# for lines that start with '--' or '++'
inheader = False
for line in lines:
if line.startswith(b'diff'):
addresult()
# starting a new file diff
# set numbers to 0 and reset inheader
inheader = True
adds, removes, isbinary = 0, 0, False
if line.startswith(b'diff --git a/'):
filename = gitre.search(line).group(2)
elif line.startswith(b'diff -r'):
# format: "diff -r ... -r ... filename"
filename = diffre.search(line).group(1)
elif line.startswith(b'@@'):
inheader = False
elif line.startswith(b'+') and not inheader:
adds += 1
elif line.startswith(b'-') and not inheader:
removes += 1
elif line.startswith(b'GIT binary patch') or line.startswith(
b'Binary file'
):
isbinary = True
elif line.startswith(b'rename from'):
filename = line[12:]
elif line.startswith(b'rename to'):
filename += b' => %s' % line[10:]
addresult()
return results
def diffstat(lines, width=80):
output = []
stats = diffstatdata(lines)
maxname, maxtotal, totaladds, totalremoves, hasbinary = diffstatsum(stats)
countwidth = len(str(maxtotal))
if hasbinary and countwidth < 3:
countwidth = 3
graphwidth = width - countwidth - maxname - 6
if graphwidth < 10:
graphwidth = 10
def scale(i):
if maxtotal <= graphwidth:
return i
# If diffstat runs out of room it doesn't print anything,
# which isn't very useful, so always print at least one + or -
# if there were at least some changes.
return max(i * graphwidth // maxtotal, int(bool(i)))
for filename, adds, removes, isbinary in stats:
if isbinary:
count = b'Bin'
else:
count = b'%d' % (adds + removes)
pluses = b'+' * scale(adds)
minuses = b'-' * scale(removes)
output.append(
b' %s%s | %*s %s%s\n'
% (
filename,
b' ' * (maxname - encoding.colwidth(filename)),
countwidth,
count,
pluses,
minuses,
)
)
if stats:
output.append(
_(b' %d files changed, %d insertions(+), %d deletions(-)\n')
% (len(stats), totaladds, totalremoves)
)
return b''.join(output)
def diffstatui(*args, **kw):
"""like diffstat(), but yields 2-tuples of (output, label) for
ui.write()
"""
for line in diffstat(*args, **kw).splitlines():
if line and line[-1] in b'+-':
name, graph = line.rsplit(b' ', 1)
yield (name + b' ', b'')
m = re.search(br'\++', graph)
if m:
yield (m.group(0), b'diffstat.inserted')
m = re.search(br'-+', graph)
if m:
yield (m.group(0), b'diffstat.deleted')
else:
yield (line, b'')
yield (b'\n', b'')