mercurial/match.py
author Pierre-Yves David <pierre-yves.david@octobus.net>
Mon, 14 Dec 2020 11:32:24 +0100
changeset 46149 294d5aca4ff5
parent 46064 1f0ed7e63c2a
child 46819 d4ba4d51f85f
permissions -rw-r--r--
copies: iterate over children directly (instead of parents) Before this change we would gather all parent → child edges and iterate on all parent, gathering copy information for children and aggregating them from there. They are not strict requirement for edges to be processed in that specific order. We could also simply iterate over all "children" revision and aggregate data from both parents at the same time. This patch does that. It make various things simpler: * since both parents are processed at the same time, we no longer need to cache data for merge (see next changeset for details), * we no longer need nested loop to process data, * we no longer need to store partial merge data for a rev from distinct loop interaction to another when processing merges, * we no longer need to build a full parent -> children mapping (we only rely on a simpler "parent -> number of children" map (for memory efficiency), * the data access pattern is now simpler (from lower revisions to higher revisions) and entirely predicable. That predictability open the way to prefetching and parallel processing. So that new iterations order requires simpler code and open the way to interesting optimisation. The effect on performance is quite good. In the worse case, we don't see any significant negative impact. And in the best case, the reduction of roundtrip to Python provide us with a significant speed. Some example below: Repo Case Source-Rev Dest-Rev # of revisions old time new time Difference Factor time per rev --------------------------------------------------------------------------------------------------------------------------------------------------------------- mozilla-try x00000_revs_x00000_added_0_copies dc8a3ca7010e d16fde900c9c : 34414 revs, 0.962867 s, 0.502584 s, -0.460283 s, × 0.5220, 14 µs/rev mozilla-try x0000_revs_xx000_added_x_copies 156f6e2674f2 4d0f2c178e66 : 8598 revs, 0.110717 s, 0.076323 s, -0.034394 s, × 0.6894, 8 µs/rev # full comparison between the previous changeset and this one Repo Case Source-Rev Dest-Rev # of revisions old time new time Difference Factor time per rev --------------------------------------------------------------------------------------------------------------------------------------------------------------- mercurial x_revs_x_added_0_copies ad6b123de1c7 39cfcef4f463 : 1 revs, 0.000048 s, 0.000041 s, -0.000007 s, × 0.8542, 41 µs/rev mercurial x_revs_x_added_x_copies 2b1c78674230 0c1d10351869 : 6 revs, 0.000153 s, 0.000102 s, -0.000051 s, × 0.6667, 17 µs/rev mercurial x000_revs_x000_added_x_copies 81f8ff2a9bf2 dd3267698d84 : 1032 revs, 0.004209 s, 0.004254 s, +0.000045 s, × 1.0107, 4 µs/rev pypy x_revs_x_added_0_copies aed021ee8ae8 099ed31b181b : 9 revs, 0.000203 s, 0.000282 s, +0.000079 s, × 1.3892, 31 µs/rev pypy x_revs_x000_added_0_copies 4aa4e1f8e19a 359343b9ac0e : 1 revs, 0.000059 s, 0.000048 s, -0.000011 s, × 0.8136, 48 µs/rev pypy x_revs_x_added_x_copies ac52eb7bbbb0 72e022663155 : 7 revs, 0.000194 s, 0.000211 s, +0.000017 s, × 1.0876, 30 µs/rev pypy x_revs_x00_added_x_copies c3b14617fbd7 ace7255d9a26 : 1 revs, 0.000380 s, 0.000375 s, -0.000005 s, × 0.9868, 375 µs/rev pypy x_revs_x000_added_x000_copies df6f7a526b60 a83dc6a2d56f : 6 revs, 0.010588 s, 0.010574 s, -0.000014 s, × 0.9987, 1762 µs/rev pypy x000_revs_xx00_added_0_copies 89a76aede314 2f22446ff07e : 4785 revs, 0.048961 s, 0.049974 s, +0.001013 s, × 1.0207, 10 µs/rev pypy x000_revs_x000_added_x_copies 8a3b5bfd266e 2c68e87c3efe : 6780 revs, 0.083612 s, 0.084300 s, +0.000688 s, × 1.0082, 12 µs/rev pypy x000_revs_x000_added_x000_copies 89a76aede314 7b3dda341c84 : 5441 revs, 0.058579 s, 0.060128 s, +0.001549 s, × 1.0264, 11 µs/rev pypy x0000_revs_x_added_0_copies d1defd0dc478 c9cb1334cc78 : 43645 revs, 0.736783 s, 0.686542 s, -0.050241 s, × 0.9318, 15 µs/rev pypy x0000_revs_xx000_added_0_copies bf2c629d0071 4ffed77c095c : 2 revs, 0.022050 s, 0.009277 s, -0.012773 s, × 0.4207, 4638 µs/rev pypy x0000_revs_xx000_added_x000_copies 08ea3258278e d9fa043f30c0 : 11316 revs, 0.120800 s, 0.114733 s, -0.006067 s, × 0.9498, 10 µs/rev netbeans x_revs_x_added_0_copies fb0955ffcbcd a01e9239f9e7 : 2 revs, 0.000140 s, 0.000081 s, -0.000059 s, × 0.5786, 40 µs/rev netbeans x_revs_x000_added_0_copies 6f360122949f 20eb231cc7d0 : 2 revs, 0.000114 s, 0.000107 s, -0.000007 s, × 0.9386, 53 µs/rev netbeans x_revs_x_added_x_copies 1ada3faf6fb6 5a39d12eecf4 : 3 revs, 0.000224 s, 0.000173 s, -0.000051 s, × 0.7723, 57 µs/rev netbeans x_revs_x00_added_x_copies 35be93ba1e2c 9eec5e90c05f : 9 revs, 0.000723 s, 0.000698 s, -0.000025 s, × 0.9654, 77 µs/rev netbeans x000_revs_xx00_added_0_copies eac3045b4fdd 51d4ae7f1290 : 1421 revs, 0.009665 s, 0.009248 s, -0.000417 s, × 0.9569, 6 µs/rev netbeans x000_revs_x000_added_x_copies e2063d266acd 6081d72689dc : 1533 revs, 0.014820 s, 0.015446 s, +0.000626 s, × 1.0422, 10 µs/rev netbeans x000_revs_x000_added_x000_copies ff453e9fee32 411350406ec2 : 5750 revs, 0.076049 s, 0.074373 s, -0.001676 s, × 0.9780, 12 µs/rev netbeans x0000_revs_xx000_added_x000_copies 588c2d1ced70 1aad62e59ddd : 66949 revs, 0.683603 s, 0.639870 s, -0.043733 s, × 0.9360, 9 µs/rev mozilla-central x_revs_x_added_0_copies 3697f962bb7b 7015fcdd43a2 : 2 revs, 0.000161 s, 0.000088 s, -0.000073 s, × 0.5466, 44 µs/rev mozilla-central x_revs_x000_added_0_copies dd390860c6c9 40d0c5bed75d : 8 revs, 0.000234 s, 0.000199 s, -0.000035 s, × 0.8504, 24 µs/rev mozilla-central x_revs_x_added_x_copies 8d198483ae3b 14207ffc2b2f : 9 revs, 0.000247 s, 0.000171 s, -0.000076 s, × 0.6923, 19 µs/rev mozilla-central x_revs_x00_added_x_copies 98cbc58cc6bc 446a150332c3 : 7 revs, 0.000630 s, 0.000592 s, -0.000038 s, × 0.9397, 84 µs/rev mozilla-central x_revs_x000_added_x000_copies 3c684b4b8f68 0a5e72d1b479 : 3 revs, 0.003286 s, 0.003151 s, -0.000135 s, × 0.9589, 1050 µs/rev mozilla-central x_revs_x0000_added_x0000_copies effb563bb7e5 c07a39dc4e80 : 6 revs, 0.062441 s, 0.061612 s, -0.000829 s, × 0.9867, 10268 µs/rev mozilla-central x000_revs_xx00_added_0_copies 6100d773079a 04a55431795e : 1593 revs, 0.005423 s, 0.005381 s, -0.000042 s, × 0.9923, 3 µs/rev mozilla-central x000_revs_x000_added_x_copies 9f17a6fc04f9 2d37b966abed : 41 revs, 0.005919 s, 0.003742 s, -0.002177 s, × 0.6322, 91 µs/rev mozilla-central x000_revs_x000_added_x000_copies 7c97034feb78 4407bd0c6330 : 7839 revs, 0.062597 s, 0.061983 s, -0.000614 s, × 0.9902, 7 µs/rev mozilla-central x0000_revs_xx000_added_0_copies 9eec5917337d 67118cc6dcad : 615 revs, 0.043551 s, 0.019861 s, -0.023690 s, × 0.4560, 32 µs/rev mozilla-central x0000_revs_xx000_added_x000_copies f78c615a656c 96a38b690156 : 30263 revs, 0.192475 s, 0.188101 s, -0.004374 s, × 0.9773, 6 µs/rev mozilla-central x00000_revs_x0000_added_x0000_copies 6832ae71433c 4c222a1d9a00 : 153721 revs, 1.955575 s, 1.806696 s, -0.148879 s, × 0.9239, 11 µs/rev mozilla-central x00000_revs_x00000_added_x000_copies 76caed42cf7c 1daa622bbe42 : 204976 revs, 2.886501 s, 2.682987 s, -0.203514 s, × 0.9295, 13 µs/rev mozilla-try x_revs_x_added_0_copies aaf6dde0deb8 9790f499805a : 2 revs, 0.001181 s, 0.000852 s, -0.000329 s, × 0.7214, 426 µs/rev mozilla-try x_revs_x000_added_0_copies d8d0222927b4 5bb8ce8c7450 : 2 revs, 0.001189 s, 0.000859 s, -0.000330 s, × 0.7225, 429 µs/rev mozilla-try x_revs_x_added_x_copies 092fcca11bdb 936255a0384a : 4 revs, 0.000563 s, 0.000150 s, -0.000413 s, × 0.2664, 37 µs/rev mozilla-try x_revs_x00_added_x_copies b53d2fadbdb5 017afae788ec : 2 revs, 0.001548 s, 0.001158 s, -0.000390 s, × 0.7481, 579 µs/rev mozilla-try x_revs_x000_added_x000_copies 20408ad61ce5 6f0ee96e21ad : 1 revs, 0.027782 s, 0.027240 s, -0.000542 s, × 0.9805, 27240 µs/rev mozilla-try x_revs_x0000_added_x0000_copies effb563bb7e5 c07a39dc4e80 : 6 revs, 0.062781 s, 0.062824 s, +0.000043 s, × 1.0007, 10470 µs/rev mozilla-try x000_revs_xx00_added_0_copies 6100d773079a 04a55431795e : 1593 revs, 0.005778 s, 0.005463 s, -0.000315 s, × 0.9455, 3 µs/rev mozilla-try x000_revs_x000_added_x_copies 9f17a6fc04f9 2d37b966abed : 41 revs, 0.006192 s, 0.004238 s, -0.001954 s, × 0.6844, 103 µs/rev mozilla-try x000_revs_x000_added_x000_copies 1346fd0130e4 4c65cbdabc1f : 6657 revs, 0.065391 s, 0.064113 s, -0.001278 s, × 0.9805, 9 µs/rev mozilla-try x0000_revs_x_added_0_copies 63519bfd42ee a36a2a865d92 : 40314 revs, 0.317216 s, 0.294063 s, -0.023153 s, × 0.9270, 7 µs/rev mozilla-try x0000_revs_x_added_x_copies 9fe69ff0762d bcabf2a78927 : 38690 revs, 0.303119 s, 0.281493 s, -0.021626 s, × 0.9287, 7 µs/rev mozilla-try x0000_revs_xx000_added_x_copies 156f6e2674f2 4d0f2c178e66 : 8598 revs, 0.110717 s, 0.076323 s, -0.034394 s, × 0.6894, 8 µs/rev mozilla-try x0000_revs_xx000_added_0_copies 9eec5917337d 67118cc6dcad : 615 revs, 0.045739 s, 0.020390 s, -0.025349 s, × 0.4458, 33 µs/rev mozilla-try x0000_revs_xx000_added_x000_copies 89294cd501d9 7ccb2fc7ccb5 : 97052 revs, 3.098021 s, 3.023879 s, -0.074142 s, × 0.9761, 31 µs/rev mozilla-try x0000_revs_x0000_added_x0000_copies e928c65095ed e951f4ad123a : 52031 revs, 0.771480 s, 0.735549 s, -0.035931 s, × 0.9534, 14 µs/rev mozilla-try x00000_revs_x_added_0_copies 6a320851d377 1ebb79acd503 : 363753 revs, 18.813422 s, 18.568900 s, -0.244522 s, × 0.9870, 51 µs/rev mozilla-try x00000_revs_x00000_added_0_copies dc8a3ca7010e d16fde900c9c : 34414 revs, 0.962867 s, 0.502584 s, -0.460283 s, × 0.5220, 14 µs/rev mozilla-try x00000_revs_x_added_x_copies 5173c4b6f97c 95d83ee7242d : 362229 revs, 18.684923 s, 18.356645 s, -0.328278 s, × 0.9824, 50 µs/rev mozilla-try x00000_revs_x000_added_x_copies 9126823d0e9c ca82787bb23c : 359344 revs, 18.296305 s, 18.250393 s, -0.045912 s, × 0.9975, 50 µs/rev mozilla-try x00000_revs_x0000_added_x0000_copies 8d3fafa80d4b eb884023b810 : 192665 revs, 3.061887 s, 2.792459 s, -0.269428 s, × 0.9120, 14 µs/rev mozilla-try x00000_revs_x00000_added_x0000_copies 1b661134e2ca 1ae03d022d6d : 228985 revs, 103.869641 s, 107.697264 s, +3.827623 s, × 1.0369, 470 µs/rev mozilla-try x00000_revs_x00000_added_x000_copies 9b2a99adc05e 8e29777b48e6 : 382065 revs, 64.262957 s, 63.961040 s, -0.301917 s, × 0.9953, 167 µs/rev Differential Revision: https://phab.mercurial-scm.org/D9422

# match.py - filename matching
#
#  Copyright 2008, 2009 Matt Mackall <mpm@selenic.com> and others
#
# This software may be used and distributed according to the terms of the
# GNU General Public License version 2 or any later version.

from __future__ import absolute_import, print_function

import copy
import itertools
import os
import re

from .i18n import _
from .pycompat import open
from . import (
    encoding,
    error,
    pathutil,
    policy,
    pycompat,
    util,
)
from .utils import stringutil

rustmod = policy.importrust('dirstate')

allpatternkinds = (
    b're',
    b'glob',
    b'path',
    b'relglob',
    b'relpath',
    b'relre',
    b'rootglob',
    b'listfile',
    b'listfile0',
    b'set',
    b'include',
    b'subinclude',
    b'rootfilesin',
)
cwdrelativepatternkinds = (b'relpath', b'glob')

propertycache = util.propertycache


def _rematcher(regex):
    """compile the regexp with the best available regexp engine and return a
    matcher function"""
    m = util.re.compile(regex)
    try:
        # slightly faster, provided by facebook's re2 bindings
        return m.test_match
    except AttributeError:
        return m.match


def _expandsets(cwd, kindpats, ctx=None, listsubrepos=False, badfn=None):
    '''Returns the kindpats list with the 'set' patterns expanded to matchers'''
    matchers = []
    other = []

    for kind, pat, source in kindpats:
        if kind == b'set':
            if ctx is None:
                raise error.ProgrammingError(
                    b"fileset expression with no context"
                )
            matchers.append(ctx.matchfileset(cwd, pat, badfn=badfn))

            if listsubrepos:
                for subpath in ctx.substate:
                    sm = ctx.sub(subpath).matchfileset(cwd, pat, badfn=badfn)
                    pm = prefixdirmatcher(subpath, sm, badfn=badfn)
                    matchers.append(pm)

            continue
        other.append((kind, pat, source))
    return matchers, other


def _expandsubinclude(kindpats, root):
    """Returns the list of subinclude matcher args and the kindpats without the
    subincludes in it."""
    relmatchers = []
    other = []

    for kind, pat, source in kindpats:
        if kind == b'subinclude':
            sourceroot = pathutil.dirname(util.normpath(source))
            pat = util.pconvert(pat)
            path = pathutil.join(sourceroot, pat)

            newroot = pathutil.dirname(path)
            matcherargs = (newroot, b'', [], [b'include:%s' % path])

            prefix = pathutil.canonpath(root, root, newroot)
            if prefix:
                prefix += b'/'
            relmatchers.append((prefix, matcherargs))
        else:
            other.append((kind, pat, source))

    return relmatchers, other


def _kindpatsalwaysmatch(kindpats):
    """Checks whether the kindspats match everything, as e.g.
    'relpath:.' does.
    """
    for kind, pat, source in kindpats:
        if pat != b'' or kind not in [b'relpath', b'glob']:
            return False
    return True


def _buildkindpatsmatcher(
    matchercls,
    root,
    cwd,
    kindpats,
    ctx=None,
    listsubrepos=False,
    badfn=None,
):
    matchers = []
    fms, kindpats = _expandsets(
        cwd,
        kindpats,
        ctx=ctx,
        listsubrepos=listsubrepos,
        badfn=badfn,
    )
    if kindpats:
        m = matchercls(root, kindpats, badfn=badfn)
        matchers.append(m)
    if fms:
        matchers.extend(fms)
    if not matchers:
        return nevermatcher(badfn=badfn)
    if len(matchers) == 1:
        return matchers[0]
    return unionmatcher(matchers)


def match(
    root,
    cwd,
    patterns=None,
    include=None,
    exclude=None,
    default=b'glob',
    auditor=None,
    ctx=None,
    listsubrepos=False,
    warn=None,
    badfn=None,
    icasefs=False,
):
    r"""build an object to match a set of file patterns

    arguments:
    root - the canonical root of the tree you're matching against
    cwd - the current working directory, if relevant
    patterns - patterns to find
    include - patterns to include (unless they are excluded)
    exclude - patterns to exclude (even if they are included)
    default - if a pattern in patterns has no explicit type, assume this one
    auditor - optional path auditor
    ctx - optional changecontext
    listsubrepos - if True, recurse into subrepositories
    warn - optional function used for printing warnings
    badfn - optional bad() callback for this matcher instead of the default
    icasefs - make a matcher for wdir on case insensitive filesystems, which
        normalizes the given patterns to the case in the filesystem

    a pattern is one of:
    'glob:<glob>' - a glob relative to cwd
    're:<regexp>' - a regular expression
    'path:<path>' - a path relative to repository root, which is matched
                    recursively
    'rootfilesin:<path>' - a path relative to repository root, which is
                    matched non-recursively (will not match subdirectories)
    'relglob:<glob>' - an unrooted glob (*.c matches C files in all dirs)
    'relpath:<path>' - a path relative to cwd
    'relre:<regexp>' - a regexp that needn't match the start of a name
    'set:<fileset>' - a fileset expression
    'include:<path>' - a file of patterns to read and include
    'subinclude:<path>' - a file of patterns to match against files under
                          the same directory
    '<something>' - a pattern of the specified default type

    >>> def _match(root, *args, **kwargs):
    ...     return match(util.localpath(root), *args, **kwargs)

    Usually a patternmatcher is returned:
    >>> _match(b'/foo', b'.', [b're:.*\.c$', b'path:foo/a', b'*.py'])
    <patternmatcher patterns='.*\\.c$|foo/a(?:/|$)|[^/]*\\.py$'>

    Combining 'patterns' with 'include' (resp. 'exclude') gives an
    intersectionmatcher (resp. a differencematcher):
    >>> type(_match(b'/foo', b'.', [b're:.*\.c$'], include=[b'path:lib']))
    <class 'mercurial.match.intersectionmatcher'>
    >>> type(_match(b'/foo', b'.', [b're:.*\.c$'], exclude=[b'path:build']))
    <class 'mercurial.match.differencematcher'>

    Notice that, if 'patterns' is empty, an alwaysmatcher is returned:
    >>> _match(b'/foo', b'.', [])
    <alwaysmatcher>

    The 'default' argument determines which kind of pattern is assumed if a
    pattern has no prefix:
    >>> _match(b'/foo', b'.', [b'.*\.c$'], default=b're')
    <patternmatcher patterns='.*\\.c$'>
    >>> _match(b'/foo', b'.', [b'main.py'], default=b'relpath')
    <patternmatcher patterns='main\\.py(?:/|$)'>
    >>> _match(b'/foo', b'.', [b'main.py'], default=b're')
    <patternmatcher patterns='main.py'>

    The primary use of matchers is to check whether a value (usually a file
    name) matches againset one of the patterns given at initialization. There
    are two ways of doing this check.

    >>> m = _match(b'/foo', b'', [b're:.*\.c$', b'relpath:a'])

    1. Calling the matcher with a file name returns True if any pattern
    matches that file name:
    >>> m(b'a')
    True
    >>> m(b'main.c')
    True
    >>> m(b'test.py')
    False

    2. Using the exact() method only returns True if the file name matches one
    of the exact patterns (i.e. not re: or glob: patterns):
    >>> m.exact(b'a')
    True
    >>> m.exact(b'main.c')
    False
    """
    assert os.path.isabs(root)
    cwd = os.path.join(root, util.localpath(cwd))
    normalize = _donormalize
    if icasefs:
        dirstate = ctx.repo().dirstate
        dsnormalize = dirstate.normalize

        def normalize(patterns, default, root, cwd, auditor, warn):
            kp = _donormalize(patterns, default, root, cwd, auditor, warn)
            kindpats = []
            for kind, pats, source in kp:
                if kind not in (b're', b'relre'):  # regex can't be normalized
                    p = pats
                    pats = dsnormalize(pats)

                    # Preserve the original to handle a case only rename.
                    if p != pats and p in dirstate:
                        kindpats.append((kind, p, source))

                kindpats.append((kind, pats, source))
            return kindpats

    if patterns:
        kindpats = normalize(patterns, default, root, cwd, auditor, warn)
        if _kindpatsalwaysmatch(kindpats):
            m = alwaysmatcher(badfn)
        else:
            m = _buildkindpatsmatcher(
                patternmatcher,
                root,
                cwd,
                kindpats,
                ctx=ctx,
                listsubrepos=listsubrepos,
                badfn=badfn,
            )
    else:
        # It's a little strange that no patterns means to match everything.
        # Consider changing this to match nothing (probably using nevermatcher).
        m = alwaysmatcher(badfn)

    if include:
        kindpats = normalize(include, b'glob', root, cwd, auditor, warn)
        im = _buildkindpatsmatcher(
            includematcher,
            root,
            cwd,
            kindpats,
            ctx=ctx,
            listsubrepos=listsubrepos,
            badfn=None,
        )
        m = intersectmatchers(m, im)
    if exclude:
        kindpats = normalize(exclude, b'glob', root, cwd, auditor, warn)
        em = _buildkindpatsmatcher(
            includematcher,
            root,
            cwd,
            kindpats,
            ctx=ctx,
            listsubrepos=listsubrepos,
            badfn=None,
        )
        m = differencematcher(m, em)
    return m


def exact(files, badfn=None):
    return exactmatcher(files, badfn=badfn)


def always(badfn=None):
    return alwaysmatcher(badfn)


def never(badfn=None):
    return nevermatcher(badfn)


def badmatch(match, badfn):
    """Make a copy of the given matcher, replacing its bad method with the given
    one.
    """
    m = copy.copy(match)
    m.bad = badfn
    return m


def _donormalize(patterns, default, root, cwd, auditor=None, warn=None):
    """Convert 'kind:pat' from the patterns list to tuples with kind and
    normalized and rooted patterns and with listfiles expanded."""
    kindpats = []
    for kind, pat in [_patsplit(p, default) for p in patterns]:
        if kind in cwdrelativepatternkinds:
            pat = pathutil.canonpath(root, cwd, pat, auditor=auditor)
        elif kind in (b'relglob', b'path', b'rootfilesin', b'rootglob'):
            pat = util.normpath(pat)
        elif kind in (b'listfile', b'listfile0'):
            try:
                files = util.readfile(pat)
                if kind == b'listfile0':
                    files = files.split(b'\0')
                else:
                    files = files.splitlines()
                files = [f for f in files if f]
            except EnvironmentError:
                raise error.Abort(_(b"unable to read file list (%s)") % pat)
            for k, p, source in _donormalize(
                files, default, root, cwd, auditor, warn
            ):
                kindpats.append((k, p, pat))
            continue
        elif kind == b'include':
            try:
                fullpath = os.path.join(root, util.localpath(pat))
                includepats = readpatternfile(fullpath, warn)
                for k, p, source in _donormalize(
                    includepats, default, root, cwd, auditor, warn
                ):
                    kindpats.append((k, p, source or pat))
            except error.Abort as inst:
                raise error.Abort(
                    b'%s: %s'
                    % (
                        pat,
                        inst.message,
                    )  # pytype: disable=unsupported-operands
                )
            except IOError as inst:
                if warn:
                    warn(
                        _(b"skipping unreadable pattern file '%s': %s\n")
                        % (pat, stringutil.forcebytestr(inst.strerror))
                    )
            continue
        # else: re or relre - which cannot be normalized
        kindpats.append((kind, pat, b''))
    return kindpats


class basematcher(object):
    def __init__(self, badfn=None):
        if badfn is not None:
            self.bad = badfn

    def __call__(self, fn):
        return self.matchfn(fn)

    # Callbacks related to how the matcher is used by dirstate.walk.
    # Subscribers to these events must monkeypatch the matcher object.
    def bad(self, f, msg):
        """Callback from dirstate.walk for each explicit file that can't be
        found/accessed, with an error message."""

    # If an traversedir is set, it will be called when a directory discovered
    # by recursive traversal is visited.
    traversedir = None

    @propertycache
    def _files(self):
        return []

    def files(self):
        """Explicitly listed files or patterns or roots:
        if no patterns or .always(): empty list,
        if exact: list exact files,
        if not .anypats(): list all files and dirs,
        else: optimal roots"""
        return self._files

    @propertycache
    def _fileset(self):
        return set(self._files)

    def exact(self, f):
        '''Returns True if f is in .files().'''
        return f in self._fileset

    def matchfn(self, f):
        return False

    def visitdir(self, dir):
        """Decides whether a directory should be visited based on whether it
        has potential matches in it or one of its subdirectories. This is
        based on the match's primary, included, and excluded patterns.

        Returns the string 'all' if the given directory and all subdirectories
        should be visited. Otherwise returns True or False indicating whether
        the given directory should be visited.
        """
        return True

    def visitchildrenset(self, dir):
        """Decides whether a directory should be visited based on whether it
        has potential matches in it or one of its subdirectories, and
        potentially lists which subdirectories of that directory should be
        visited. This is based on the match's primary, included, and excluded
        patterns.

        This function is very similar to 'visitdir', and the following mapping
        can be applied:

             visitdir | visitchildrenlist
            ----------+-------------------
             False    | set()
             'all'    | 'all'
             True     | 'this' OR non-empty set of subdirs -or files- to visit

        Example:
          Assume matchers ['path:foo/bar', 'rootfilesin:qux'], we would return
          the following values (assuming the implementation of visitchildrenset
          is capable of recognizing this; some implementations are not).

          '' -> {'foo', 'qux'}
          'baz' -> set()
          'foo' -> {'bar'}
          # Ideally this would be 'all', but since the prefix nature of matchers
          # is applied to the entire matcher, we have to downgrade this to
          # 'this' due to the non-prefix 'rootfilesin'-kind matcher being mixed
          # in.
          'foo/bar' -> 'this'
          'qux' -> 'this'

        Important:
          Most matchers do not know if they're representing files or
          directories. They see ['path:dir/f'] and don't know whether 'f' is a
          file or a directory, so visitchildrenset('dir') for most matchers will
          return {'f'}, but if the matcher knows it's a file (like exactmatcher
          does), it may return 'this'. Do not rely on the return being a set
          indicating that there are no files in this dir to investigate (or
          equivalently that if there are files to investigate in 'dir' that it
          will always return 'this').
        """
        return b'this'

    def always(self):
        """Matcher will match everything and .files() will be empty --
        optimization might be possible."""
        return False

    def isexact(self):
        """Matcher will match exactly the list of files in .files() --
        optimization might be possible."""
        return False

    def prefix(self):
        """Matcher will match the paths in .files() recursively --
        optimization might be possible."""
        return False

    def anypats(self):
        """None of .always(), .isexact(), and .prefix() is true --
        optimizations will be difficult."""
        return not self.always() and not self.isexact() and not self.prefix()


class alwaysmatcher(basematcher):
    '''Matches everything.'''

    def __init__(self, badfn=None):
        super(alwaysmatcher, self).__init__(badfn)

    def always(self):
        return True

    def matchfn(self, f):
        return True

    def visitdir(self, dir):
        return b'all'

    def visitchildrenset(self, dir):
        return b'all'

    def __repr__(self):
        return r'<alwaysmatcher>'


class nevermatcher(basematcher):
    '''Matches nothing.'''

    def __init__(self, badfn=None):
        super(nevermatcher, self).__init__(badfn)

    # It's a little weird to say that the nevermatcher is an exact matcher
    # or a prefix matcher, but it seems to make sense to let callers take
    # fast paths based on either. There will be no exact matches, nor any
    # prefixes (files() returns []), so fast paths iterating over them should
    # be efficient (and correct).
    def isexact(self):
        return True

    def prefix(self):
        return True

    def visitdir(self, dir):
        return False

    def visitchildrenset(self, dir):
        return set()

    def __repr__(self):
        return r'<nevermatcher>'


class predicatematcher(basematcher):
    """A matcher adapter for a simple boolean function"""

    def __init__(self, predfn, predrepr=None, badfn=None):
        super(predicatematcher, self).__init__(badfn)
        self.matchfn = predfn
        self._predrepr = predrepr

    @encoding.strmethod
    def __repr__(self):
        s = stringutil.buildrepr(self._predrepr) or pycompat.byterepr(
            self.matchfn
        )
        return b'<predicatenmatcher pred=%s>' % s


def path_or_parents_in_set(path, prefix_set):
    """Returns True if `path` (or any parent of `path`) is in `prefix_set`."""
    l = len(prefix_set)
    if l == 0:
        return False
    if path in prefix_set:
        return True
    # If there's more than 5 paths in prefix_set, it's *probably* quicker to
    # "walk up" the directory hierarchy instead, with the assumption that most
    # directory hierarchies are relatively shallow and hash lookup is cheap.
    if l > 5:
        return any(
            parentdir in prefix_set for parentdir in pathutil.finddirs(path)
        )

    # FIXME: Ideally we'd never get to this point if this is the case - we'd
    # recognize ourselves as an 'always' matcher and skip this.
    if b'' in prefix_set:
        return True

    if pycompat.ispy3:
        sl = ord(b'/')
    else:
        sl = '/'

    # We already checked that path isn't in prefix_set exactly, so
    # `path[len(pf)] should never raise IndexError.
    return any(path.startswith(pf) and path[len(pf)] == sl for pf in prefix_set)


class patternmatcher(basematcher):
    r"""Matches a set of (kind, pat, source) against a 'root' directory.

    >>> kindpats = [
    ...     (b're', br'.*\.c$', b''),
    ...     (b'path', b'foo/a', b''),
    ...     (b'relpath', b'b', b''),
    ...     (b'glob', b'*.h', b''),
    ... ]
    >>> m = patternmatcher(b'foo', kindpats)
    >>> m(b'main.c')  # matches re:.*\.c$
    True
    >>> m(b'b.txt')
    False
    >>> m(b'foo/a')  # matches path:foo/a
    True
    >>> m(b'a')  # does not match path:b, since 'root' is 'foo'
    False
    >>> m(b'b')  # matches relpath:b, since 'root' is 'foo'
    True
    >>> m(b'lib.h')  # matches glob:*.h
    True

    >>> m.files()
    ['', 'foo/a', 'b', '']
    >>> m.exact(b'foo/a')
    True
    >>> m.exact(b'b')
    True
    >>> m.exact(b'lib.h')  # exact matches are for (rel)path kinds
    False
    """

    def __init__(self, root, kindpats, badfn=None):
        super(patternmatcher, self).__init__(badfn)

        self._files = _explicitfiles(kindpats)
        self._prefix = _prefix(kindpats)
        self._pats, self.matchfn = _buildmatch(kindpats, b'$', root)

    @propertycache
    def _dirs(self):
        return set(pathutil.dirs(self._fileset))

    def visitdir(self, dir):
        if self._prefix and dir in self._fileset:
            return b'all'
        return dir in self._dirs or path_or_parents_in_set(dir, self._fileset)

    def visitchildrenset(self, dir):
        ret = self.visitdir(dir)
        if ret is True:
            return b'this'
        elif not ret:
            return set()
        assert ret == b'all'
        return b'all'

    def prefix(self):
        return self._prefix

    @encoding.strmethod
    def __repr__(self):
        return b'<patternmatcher patterns=%r>' % pycompat.bytestr(self._pats)


# This is basically a reimplementation of pathutil.dirs that stores the
# children instead of just a count of them, plus a small optional optimization
# to avoid some directories we don't need.
class _dirchildren(object):
    def __init__(self, paths, onlyinclude=None):
        self._dirs = {}
        self._onlyinclude = onlyinclude or []
        addpath = self.addpath
        for f in paths:
            addpath(f)

    def addpath(self, path):
        if path == b'':
            return
        dirs = self._dirs
        findsplitdirs = _dirchildren._findsplitdirs
        for d, b in findsplitdirs(path):
            if d not in self._onlyinclude:
                continue
            dirs.setdefault(d, set()).add(b)

    @staticmethod
    def _findsplitdirs(path):
        # yields (dirname, basename) tuples, walking back to the root.  This is
        # very similar to pathutil.finddirs, except:
        #  - produces a (dirname, basename) tuple, not just 'dirname'
        # Unlike manifest._splittopdir, this does not suffix `dirname` with a
        # slash.
        oldpos = len(path)
        pos = path.rfind(b'/')
        while pos != -1:
            yield path[:pos], path[pos + 1 : oldpos]
            oldpos = pos
            pos = path.rfind(b'/', 0, pos)
        yield b'', path[:oldpos]

    def get(self, path):
        return self._dirs.get(path, set())


class includematcher(basematcher):
    def __init__(self, root, kindpats, badfn=None):
        super(includematcher, self).__init__(badfn)
        if rustmod is not None:
            # We need to pass the patterns to Rust because they can contain
            # patterns from the user interface
            self._kindpats = kindpats
        self._pats, self.matchfn = _buildmatch(kindpats, b'(?:/|$)', root)
        self._prefix = _prefix(kindpats)
        roots, dirs, parents = _rootsdirsandparents(kindpats)
        # roots are directories which are recursively included.
        self._roots = set(roots)
        # dirs are directories which are non-recursively included.
        self._dirs = set(dirs)
        # parents are directories which are non-recursively included because
        # they are needed to get to items in _dirs or _roots.
        self._parents = parents

    def visitdir(self, dir):
        if self._prefix and dir in self._roots:
            return b'all'
        return (
            dir in self._dirs
            or dir in self._parents
            or path_or_parents_in_set(dir, self._roots)
        )

    @propertycache
    def _allparentschildren(self):
        # It may seem odd that we add dirs, roots, and parents, and then
        # restrict to only parents. This is to catch the case of:
        #   dirs = ['foo/bar']
        #   parents = ['foo']
        # if we asked for the children of 'foo', but had only added
        # self._parents, we wouldn't be able to respond ['bar'].
        return _dirchildren(
            itertools.chain(self._dirs, self._roots, self._parents),
            onlyinclude=self._parents,
        )

    def visitchildrenset(self, dir):
        if self._prefix and dir in self._roots:
            return b'all'
        # Note: this does *not* include the 'dir in self._parents' case from
        # visitdir, that's handled below.
        if (
            b'' in self._roots
            or dir in self._dirs
            or path_or_parents_in_set(dir, self._roots)
        ):
            return b'this'

        if dir in self._parents:
            return self._allparentschildren.get(dir) or set()
        return set()

    @encoding.strmethod
    def __repr__(self):
        return b'<includematcher includes=%r>' % pycompat.bytestr(self._pats)


class exactmatcher(basematcher):
    r"""Matches the input files exactly. They are interpreted as paths, not
    patterns (so no kind-prefixes).

    >>> m = exactmatcher([b'a.txt', br're:.*\.c$'])
    >>> m(b'a.txt')
    True
    >>> m(b'b.txt')
    False

    Input files that would be matched are exactly those returned by .files()
    >>> m.files()
    ['a.txt', 're:.*\\.c$']

    So pattern 're:.*\.c$' is not considered as a regex, but as a file name
    >>> m(b'main.c')
    False
    >>> m(br're:.*\.c$')
    True
    """

    def __init__(self, files, badfn=None):
        super(exactmatcher, self).__init__(badfn)

        if isinstance(files, list):
            self._files = files
        else:
            self._files = list(files)

    matchfn = basematcher.exact

    @propertycache
    def _dirs(self):
        return set(pathutil.dirs(self._fileset))

    def visitdir(self, dir):
        return dir in self._dirs

    def visitchildrenset(self, dir):
        if not self._fileset or dir not in self._dirs:
            return set()

        candidates = self._fileset | self._dirs - {b''}
        if dir != b'':
            d = dir + b'/'
            candidates = {c[len(d) :] for c in candidates if c.startswith(d)}
        # self._dirs includes all of the directories, recursively, so if
        # we're attempting to match foo/bar/baz.txt, it'll have '', 'foo',
        # 'foo/bar' in it. Thus we can safely ignore a candidate that has a
        # '/' in it, indicating a it's for a subdir-of-a-subdir; the
        # immediate subdir will be in there without a slash.
        ret = {c for c in candidates if b'/' not in c}
        # We really do not expect ret to be empty, since that would imply that
        # there's something in _dirs that didn't have a file in _fileset.
        assert ret
        return ret

    def isexact(self):
        return True

    @encoding.strmethod
    def __repr__(self):
        return b'<exactmatcher files=%r>' % self._files


class differencematcher(basematcher):
    """Composes two matchers by matching if the first matches and the second
    does not.

    The second matcher's non-matching-attributes (bad, traversedir) are ignored.
    """

    def __init__(self, m1, m2):
        super(differencematcher, self).__init__()
        self._m1 = m1
        self._m2 = m2
        self.bad = m1.bad
        self.traversedir = m1.traversedir

    def matchfn(self, f):
        return self._m1(f) and not self._m2(f)

    @propertycache
    def _files(self):
        if self.isexact():
            return [f for f in self._m1.files() if self(f)]
        # If m1 is not an exact matcher, we can't easily figure out the set of
        # files, because its files() are not always files. For example, if
        # m1 is "path:dir" and m2 is "rootfileins:.", we don't
        # want to remove "dir" from the set even though it would match m2,
        # because the "dir" in m1 may not be a file.
        return self._m1.files()

    def visitdir(self, dir):
        if self._m2.visitdir(dir) == b'all':
            return False
        elif not self._m2.visitdir(dir):
            # m2 does not match dir, we can return 'all' here if possible
            return self._m1.visitdir(dir)
        return bool(self._m1.visitdir(dir))

    def visitchildrenset(self, dir):
        m2_set = self._m2.visitchildrenset(dir)
        if m2_set == b'all':
            return set()
        m1_set = self._m1.visitchildrenset(dir)
        # Possible values for m1: 'all', 'this', set(...), set()
        # Possible values for m2:        'this', set(...), set()
        # If m2 has nothing under here that we care about, return m1, even if
        # it's 'all'. This is a change in behavior from visitdir, which would
        # return True, not 'all', for some reason.
        if not m2_set:
            return m1_set
        if m1_set in [b'all', b'this']:
            # Never return 'all' here if m2_set is any kind of non-empty (either
            # 'this' or set(foo)), since m2 might return set() for a
            # subdirectory.
            return b'this'
        # Possible values for m1:         set(...), set()
        # Possible values for m2: 'this', set(...)
        # We ignore m2's set results. They're possibly incorrect:
        #  m1 = path:dir/subdir, m2=rootfilesin:dir, visitchildrenset(''):
        #    m1 returns {'dir'}, m2 returns {'dir'}, if we subtracted we'd
        #    return set(), which is *not* correct, we still need to visit 'dir'!
        return m1_set

    def isexact(self):
        return self._m1.isexact()

    @encoding.strmethod
    def __repr__(self):
        return b'<differencematcher m1=%r, m2=%r>' % (self._m1, self._m2)


def intersectmatchers(m1, m2):
    """Composes two matchers by matching if both of them match.

    The second matcher's non-matching-attributes (bad, traversedir) are ignored.
    """
    if m1 is None or m2 is None:
        return m1 or m2
    if m1.always():
        m = copy.copy(m2)
        # TODO: Consider encapsulating these things in a class so there's only
        # one thing to copy from m1.
        m.bad = m1.bad
        m.traversedir = m1.traversedir
        return m
    if m2.always():
        m = copy.copy(m1)
        return m
    return intersectionmatcher(m1, m2)


class intersectionmatcher(basematcher):
    def __init__(self, m1, m2):
        super(intersectionmatcher, self).__init__()
        self._m1 = m1
        self._m2 = m2
        self.bad = m1.bad
        self.traversedir = m1.traversedir

    @propertycache
    def _files(self):
        if self.isexact():
            m1, m2 = self._m1, self._m2
            if not m1.isexact():
                m1, m2 = m2, m1
            return [f for f in m1.files() if m2(f)]
        # It neither m1 nor m2 is an exact matcher, we can't easily intersect
        # the set of files, because their files() are not always files. For
        # example, if intersecting a matcher "-I glob:foo.txt" with matcher of
        # "path:dir2", we don't want to remove "dir2" from the set.
        return self._m1.files() + self._m2.files()

    def matchfn(self, f):
        return self._m1(f) and self._m2(f)

    def visitdir(self, dir):
        visit1 = self._m1.visitdir(dir)
        if visit1 == b'all':
            return self._m2.visitdir(dir)
        # bool() because visit1=True + visit2='all' should not be 'all'
        return bool(visit1 and self._m2.visitdir(dir))

    def visitchildrenset(self, dir):
        m1_set = self._m1.visitchildrenset(dir)
        if not m1_set:
            return set()
        m2_set = self._m2.visitchildrenset(dir)
        if not m2_set:
            return set()

        if m1_set == b'all':
            return m2_set
        elif m2_set == b'all':
            return m1_set

        if m1_set == b'this' or m2_set == b'this':
            return b'this'

        assert isinstance(m1_set, set) and isinstance(m2_set, set)
        return m1_set.intersection(m2_set)

    def always(self):
        return self._m1.always() and self._m2.always()

    def isexact(self):
        return self._m1.isexact() or self._m2.isexact()

    @encoding.strmethod
    def __repr__(self):
        return b'<intersectionmatcher m1=%r, m2=%r>' % (self._m1, self._m2)


class subdirmatcher(basematcher):
    """Adapt a matcher to work on a subdirectory only.

    The paths are remapped to remove/insert the path as needed:

    >>> from . import pycompat
    >>> m1 = match(util.localpath(b'/root'), b'', [b'a.txt', b'sub/b.txt'], auditor=lambda name: None)
    >>> m2 = subdirmatcher(b'sub', m1)
    >>> m2(b'a.txt')
    False
    >>> m2(b'b.txt')
    True
    >>> m2.matchfn(b'a.txt')
    False
    >>> m2.matchfn(b'b.txt')
    True
    >>> m2.files()
    ['b.txt']
    >>> m2.exact(b'b.txt')
    True
    >>> def bad(f, msg):
    ...     print(pycompat.sysstr(b"%s: %s" % (f, msg)))
    >>> m1.bad = bad
    >>> m2.bad(b'x.txt', b'No such file')
    sub/x.txt: No such file
    """

    def __init__(self, path, matcher):
        super(subdirmatcher, self).__init__()
        self._path = path
        self._matcher = matcher
        self._always = matcher.always()

        self._files = [
            f[len(path) + 1 :]
            for f in matcher._files
            if f.startswith(path + b"/")
        ]

        # If the parent repo had a path to this subrepo and the matcher is
        # a prefix matcher, this submatcher always matches.
        if matcher.prefix():
            self._always = any(f == path for f in matcher._files)

    def bad(self, f, msg):
        self._matcher.bad(self._path + b"/" + f, msg)

    def matchfn(self, f):
        # Some information is lost in the superclass's constructor, so we
        # can not accurately create the matching function for the subdirectory
        # from the inputs. Instead, we override matchfn() and visitdir() to
        # call the original matcher with the subdirectory path prepended.
        return self._matcher.matchfn(self._path + b"/" + f)

    def visitdir(self, dir):
        if dir == b'':
            dir = self._path
        else:
            dir = self._path + b"/" + dir
        return self._matcher.visitdir(dir)

    def visitchildrenset(self, dir):
        if dir == b'':
            dir = self._path
        else:
            dir = self._path + b"/" + dir
        return self._matcher.visitchildrenset(dir)

    def always(self):
        return self._always

    def prefix(self):
        return self._matcher.prefix() and not self._always

    @encoding.strmethod
    def __repr__(self):
        return b'<subdirmatcher path=%r, matcher=%r>' % (
            self._path,
            self._matcher,
        )


class prefixdirmatcher(basematcher):
    """Adapt a matcher to work on a parent directory.

    The matcher's non-matching-attributes (bad, traversedir) are ignored.

    The prefix path should usually be the relative path from the root of
    this matcher to the root of the wrapped matcher.

    >>> m1 = match(util.localpath(b'/root/d/e'), b'f', [b'../a.txt', b'b.txt'], auditor=lambda name: None)
    >>> m2 = prefixdirmatcher(b'd/e', m1)
    >>> m2(b'a.txt')
    False
    >>> m2(b'd/e/a.txt')
    True
    >>> m2(b'd/e/b.txt')
    False
    >>> m2.files()
    ['d/e/a.txt', 'd/e/f/b.txt']
    >>> m2.exact(b'd/e/a.txt')
    True
    >>> m2.visitdir(b'd')
    True
    >>> m2.visitdir(b'd/e')
    True
    >>> m2.visitdir(b'd/e/f')
    True
    >>> m2.visitdir(b'd/e/g')
    False
    >>> m2.visitdir(b'd/ef')
    False
    """

    def __init__(self, path, matcher, badfn=None):
        super(prefixdirmatcher, self).__init__(badfn)
        if not path:
            raise error.ProgrammingError(b'prefix path must not be empty')
        self._path = path
        self._pathprefix = path + b'/'
        self._matcher = matcher

    @propertycache
    def _files(self):
        return [self._pathprefix + f for f in self._matcher._files]

    def matchfn(self, f):
        if not f.startswith(self._pathprefix):
            return False
        return self._matcher.matchfn(f[len(self._pathprefix) :])

    @propertycache
    def _pathdirs(self):
        return set(pathutil.finddirs(self._path))

    def visitdir(self, dir):
        if dir == self._path:
            return self._matcher.visitdir(b'')
        if dir.startswith(self._pathprefix):
            return self._matcher.visitdir(dir[len(self._pathprefix) :])
        return dir in self._pathdirs

    def visitchildrenset(self, dir):
        if dir == self._path:
            return self._matcher.visitchildrenset(b'')
        if dir.startswith(self._pathprefix):
            return self._matcher.visitchildrenset(dir[len(self._pathprefix) :])
        if dir in self._pathdirs:
            return b'this'
        return set()

    def isexact(self):
        return self._matcher.isexact()

    def prefix(self):
        return self._matcher.prefix()

    @encoding.strmethod
    def __repr__(self):
        return b'<prefixdirmatcher path=%r, matcher=%r>' % (
            pycompat.bytestr(self._path),
            self._matcher,
        )


class unionmatcher(basematcher):
    """A matcher that is the union of several matchers.

    The non-matching-attributes (bad, traversedir) are taken from the first
    matcher.
    """

    def __init__(self, matchers):
        m1 = matchers[0]
        super(unionmatcher, self).__init__()
        self.traversedir = m1.traversedir
        self._matchers = matchers

    def matchfn(self, f):
        for match in self._matchers:
            if match(f):
                return True
        return False

    def visitdir(self, dir):
        r = False
        for m in self._matchers:
            v = m.visitdir(dir)
            if v == b'all':
                return v
            r |= v
        return r

    def visitchildrenset(self, dir):
        r = set()
        this = False
        for m in self._matchers:
            v = m.visitchildrenset(dir)
            if not v:
                continue
            if v == b'all':
                return v
            if this or v == b'this':
                this = True
                # don't break, we might have an 'all' in here.
                continue
            assert isinstance(v, set)
            r = r.union(v)
        if this:
            return b'this'
        return r

    @encoding.strmethod
    def __repr__(self):
        return b'<unionmatcher matchers=%r>' % self._matchers


def patkind(pattern, default=None):
    r"""If pattern is 'kind:pat' with a known kind, return kind.

    >>> patkind(br're:.*\.c$')
    're'
    >>> patkind(b'glob:*.c')
    'glob'
    >>> patkind(b'relpath:test.py')
    'relpath'
    >>> patkind(b'main.py')
    >>> patkind(b'main.py', default=b're')
    're'
    """
    return _patsplit(pattern, default)[0]


def _patsplit(pattern, default):
    """Split a string into the optional pattern kind prefix and the actual
    pattern."""
    if b':' in pattern:
        kind, pat = pattern.split(b':', 1)
        if kind in allpatternkinds:
            return kind, pat
    return default, pattern


def _globre(pat):
    r"""Convert an extended glob string to a regexp string.

    >>> from . import pycompat
    >>> def bprint(s):
    ...     print(pycompat.sysstr(s))
    >>> bprint(_globre(br'?'))
    .
    >>> bprint(_globre(br'*'))
    [^/]*
    >>> bprint(_globre(br'**'))
    .*
    >>> bprint(_globre(br'**/a'))
    (?:.*/)?a
    >>> bprint(_globre(br'a/**/b'))
    a/(?:.*/)?b
    >>> bprint(_globre(br'[a*?!^][^b][!c]'))
    [a*?!^][\^b][^c]
    >>> bprint(_globre(br'{a,b}'))
    (?:a|b)
    >>> bprint(_globre(br'.\*\?'))
    \.\*\?
    """
    i, n = 0, len(pat)
    res = b''
    group = 0
    escape = util.stringutil.regexbytesescapemap.get

    def peek():
        return i < n and pat[i : i + 1]

    while i < n:
        c = pat[i : i + 1]
        i += 1
        if c not in b'*?[{},\\':
            res += escape(c, c)
        elif c == b'*':
            if peek() == b'*':
                i += 1
                if peek() == b'/':
                    i += 1
                    res += b'(?:.*/)?'
                else:
                    res += b'.*'
            else:
                res += b'[^/]*'
        elif c == b'?':
            res += b'.'
        elif c == b'[':
            j = i
            if j < n and pat[j : j + 1] in b'!]':
                j += 1
            while j < n and pat[j : j + 1] != b']':
                j += 1
            if j >= n:
                res += b'\\['
            else:
                stuff = pat[i:j].replace(b'\\', b'\\\\')
                i = j + 1
                if stuff[0:1] == b'!':
                    stuff = b'^' + stuff[1:]
                elif stuff[0:1] == b'^':
                    stuff = b'\\' + stuff
                res = b'%s[%s]' % (res, stuff)
        elif c == b'{':
            group += 1
            res += b'(?:'
        elif c == b'}' and group:
            res += b')'
            group -= 1
        elif c == b',' and group:
            res += b'|'
        elif c == b'\\':
            p = peek()
            if p:
                i += 1
                res += escape(p, p)
            else:
                res += escape(c, c)
        else:
            res += escape(c, c)
    return res


def _regex(kind, pat, globsuffix):
    """Convert a (normalized) pattern of any kind into a
    regular expression.
    globsuffix is appended to the regexp of globs."""
    if not pat and kind in (b'glob', b'relpath'):
        return b''
    if kind == b're':
        return pat
    if kind in (b'path', b'relpath'):
        if pat == b'.':
            return b''
        return util.stringutil.reescape(pat) + b'(?:/|$)'
    if kind == b'rootfilesin':
        if pat == b'.':
            escaped = b''
        else:
            # Pattern is a directory name.
            escaped = util.stringutil.reescape(pat) + b'/'
        # Anything after the pattern must be a non-directory.
        return escaped + b'[^/]+$'
    if kind == b'relglob':
        globre = _globre(pat)
        if globre.startswith(b'[^/]*'):
            # When pat has the form *XYZ (common), make the returned regex more
            # legible by returning the regex for **XYZ instead of **/*XYZ.
            return b'.*' + globre[len(b'[^/]*') :] + globsuffix
        return b'(?:|.*/)' + globre + globsuffix
    if kind == b'relre':
        if pat.startswith(b'^'):
            return pat
        return b'.*' + pat
    if kind in (b'glob', b'rootglob'):
        return _globre(pat) + globsuffix
    raise error.ProgrammingError(b'not a regex pattern: %s:%s' % (kind, pat))


def _buildmatch(kindpats, globsuffix, root):
    """Return regexp string and a matcher function for kindpats.
    globsuffix is appended to the regexp of globs."""
    matchfuncs = []

    subincludes, kindpats = _expandsubinclude(kindpats, root)
    if subincludes:
        submatchers = {}

        def matchsubinclude(f):
            for prefix, matcherargs in subincludes:
                if f.startswith(prefix):
                    mf = submatchers.get(prefix)
                    if mf is None:
                        mf = match(*matcherargs)
                        submatchers[prefix] = mf

                    if mf(f[len(prefix) :]):
                        return True
            return False

        matchfuncs.append(matchsubinclude)

    regex = b''
    if kindpats:
        if all(k == b'rootfilesin' for k, p, s in kindpats):
            dirs = {p for k, p, s in kindpats}

            def mf(f):
                i = f.rfind(b'/')
                if i >= 0:
                    dir = f[:i]
                else:
                    dir = b'.'
                return dir in dirs

            regex = b'rootfilesin: %s' % stringutil.pprint(list(sorted(dirs)))
            matchfuncs.append(mf)
        else:
            regex, mf = _buildregexmatch(kindpats, globsuffix)
            matchfuncs.append(mf)

    if len(matchfuncs) == 1:
        return regex, matchfuncs[0]
    else:
        return regex, lambda f: any(mf(f) for mf in matchfuncs)


MAX_RE_SIZE = 20000


def _joinregexes(regexps):
    """gather multiple regular expressions into a single one"""
    return b'|'.join(regexps)


def _buildregexmatch(kindpats, globsuffix):
    """Build a match function from a list of kinds and kindpats,
    return regexp string and a matcher function.

    Test too large input
    >>> _buildregexmatch([
    ...     (b'relglob', b'?' * MAX_RE_SIZE, b'')
    ... ], b'$')
    Traceback (most recent call last):
    ...
    Abort: matcher pattern is too long (20009 bytes)
    """
    try:
        allgroups = []
        regexps = [_regex(k, p, globsuffix) for (k, p, s) in kindpats]
        fullregexp = _joinregexes(regexps)

        startidx = 0
        groupsize = 0
        for idx, r in enumerate(regexps):
            piecesize = len(r)
            if piecesize > MAX_RE_SIZE:
                msg = _(b"matcher pattern is too long (%d bytes)") % piecesize
                raise error.Abort(msg)
            elif (groupsize + piecesize) > MAX_RE_SIZE:
                group = regexps[startidx:idx]
                allgroups.append(_joinregexes(group))
                startidx = idx
                groupsize = 0
            groupsize += piecesize + 1

        if startidx == 0:
            matcher = _rematcher(fullregexp)
            func = lambda s: bool(matcher(s))
        else:
            group = regexps[startidx:]
            allgroups.append(_joinregexes(group))
            allmatchers = [_rematcher(g) for g in allgroups]
            func = lambda s: any(m(s) for m in allmatchers)
        return fullregexp, func
    except re.error:
        for k, p, s in kindpats:
            try:
                _rematcher(_regex(k, p, globsuffix))
            except re.error:
                if s:
                    raise error.Abort(
                        _(b"%s: invalid pattern (%s): %s") % (s, k, p)
                    )
                else:
                    raise error.Abort(_(b"invalid pattern (%s): %s") % (k, p))
        raise error.Abort(_(b"invalid pattern"))


def _patternrootsanddirs(kindpats):
    """Returns roots and directories corresponding to each pattern.

    This calculates the roots and directories exactly matching the patterns and
    returns a tuple of (roots, dirs) for each. It does not return other
    directories which may also need to be considered, like the parent
    directories.
    """
    r = []
    d = []
    for kind, pat, source in kindpats:
        if kind in (b'glob', b'rootglob'):  # find the non-glob prefix
            root = []
            for p in pat.split(b'/'):
                if b'[' in p or b'{' in p or b'*' in p or b'?' in p:
                    break
                root.append(p)
            r.append(b'/'.join(root))
        elif kind in (b'relpath', b'path'):
            if pat == b'.':
                pat = b''
            r.append(pat)
        elif kind in (b'rootfilesin',):
            if pat == b'.':
                pat = b''
            d.append(pat)
        else:  # relglob, re, relre
            r.append(b'')
    return r, d


def _roots(kindpats):
    '''Returns root directories to match recursively from the given patterns.'''
    roots, dirs = _patternrootsanddirs(kindpats)
    return roots


def _rootsdirsandparents(kindpats):
    """Returns roots and exact directories from patterns.

    `roots` are directories to match recursively, `dirs` should
    be matched non-recursively, and `parents` are the implicitly required
    directories to walk to items in either roots or dirs.

    Returns a tuple of (roots, dirs, parents).

    >>> r = _rootsdirsandparents(
    ...     [(b'glob', b'g/h/*', b''), (b'glob', b'g/h', b''),
    ...      (b'glob', b'g*', b'')])
    >>> print(r[0:2], sorted(r[2])) # the set has an unstable output
    (['g/h', 'g/h', ''], []) ['', 'g']
    >>> r = _rootsdirsandparents(
    ...     [(b'rootfilesin', b'g/h', b''), (b'rootfilesin', b'', b'')])
    >>> print(r[0:2], sorted(r[2])) # the set has an unstable output
    ([], ['g/h', '']) ['', 'g']
    >>> r = _rootsdirsandparents(
    ...     [(b'relpath', b'r', b''), (b'path', b'p/p', b''),
    ...      (b'path', b'', b'')])
    >>> print(r[0:2], sorted(r[2])) # the set has an unstable output
    (['r', 'p/p', ''], []) ['', 'p']
    >>> r = _rootsdirsandparents(
    ...     [(b'relglob', b'rg*', b''), (b're', b're/', b''),
    ...      (b'relre', b'rr', b'')])
    >>> print(r[0:2], sorted(r[2])) # the set has an unstable output
    (['', '', ''], []) ['']
    """
    r, d = _patternrootsanddirs(kindpats)

    p = set()
    # Add the parents as non-recursive/exact directories, since they must be
    # scanned to get to either the roots or the other exact directories.
    p.update(pathutil.dirs(d))
    p.update(pathutil.dirs(r))

    # FIXME: all uses of this function convert these to sets, do so before
    # returning.
    # FIXME: all uses of this function do not need anything in 'roots' and
    # 'dirs' to also be in 'parents', consider removing them before returning.
    return r, d, p


def _explicitfiles(kindpats):
    """Returns the potential explicit filenames from the patterns.

    >>> _explicitfiles([(b'path', b'foo/bar', b'')])
    ['foo/bar']
    >>> _explicitfiles([(b'rootfilesin', b'foo/bar', b'')])
    []
    """
    # Keep only the pattern kinds where one can specify filenames (vs only
    # directory names).
    filable = [kp for kp in kindpats if kp[0] not in (b'rootfilesin',)]
    return _roots(filable)


def _prefix(kindpats):
    '''Whether all the patterns match a prefix (i.e. recursively)'''
    for kind, pat, source in kindpats:
        if kind not in (b'path', b'relpath'):
            return False
    return True


_commentre = None


def readpatternfile(filepath, warn, sourceinfo=False):
    """parse a pattern file, returning a list of
    patterns. These patterns should be given to compile()
    to be validated and converted into a match function.

    trailing white space is dropped.
    the escape character is backslash.
    comments start with #.
    empty lines are skipped.

    lines can be of the following formats:

    syntax: regexp # defaults following lines to non-rooted regexps
    syntax: glob   # defaults following lines to non-rooted globs
    re:pattern     # non-rooted regular expression
    glob:pattern   # non-rooted glob
    rootglob:pat   # rooted glob (same root as ^ in regexps)
    pattern        # pattern of the current default type

    if sourceinfo is set, returns a list of tuples:
    (pattern, lineno, originalline).
    This is useful to debug ignore patterns.
    """

    syntaxes = {
        b're': b'relre:',
        b'regexp': b'relre:',
        b'glob': b'relglob:',
        b'rootglob': b'rootglob:',
        b'include': b'include',
        b'subinclude': b'subinclude',
    }
    syntax = b'relre:'
    patterns = []

    fp = open(filepath, b'rb')
    for lineno, line in enumerate(util.iterfile(fp), start=1):
        if b"#" in line:
            global _commentre
            if not _commentre:
                _commentre = util.re.compile(br'((?:^|[^\\])(?:\\\\)*)#.*')
            # remove comments prefixed by an even number of escapes
            m = _commentre.search(line)
            if m:
                line = line[: m.end(1)]
            # fixup properly escaped comments that survived the above
            line = line.replace(b"\\#", b"#")
        line = line.rstrip()
        if not line:
            continue

        if line.startswith(b'syntax:'):
            s = line[7:].strip()
            try:
                syntax = syntaxes[s]
            except KeyError:
                if warn:
                    warn(
                        _(b"%s: ignoring invalid syntax '%s'\n") % (filepath, s)
                    )
            continue

        linesyntax = syntax
        for s, rels in pycompat.iteritems(syntaxes):
            if line.startswith(rels):
                linesyntax = rels
                line = line[len(rels) :]
                break
            elif line.startswith(s + b':'):
                linesyntax = rels
                line = line[len(s) + 1 :]
                break
        if sourceinfo:
            patterns.append((linesyntax + line, lineno, line))
        else:
            patterns.append(linesyntax + line)
    fp.close()
    return patterns