view tests/test-fix-metadata.t @ 42743:8c9a6adec67a

rust-discovery: using the children cache in add_missing The DAG range computation often needs to get back to very old revisions, and turns out to be disproportionately long, given that the end goal is to remove the descendents of the given missing revisons from the undecided set. The fast iteration capabilities available in the Rust case make it possible to avoid the DAG range entirely, at the cost of precomputing the children cache, and to simply iterate on children of the given missing revisions. This is a case where staying on the same side of the interface between the two languages has clear benefits. On discoveries with initial undecided sets small enough to bypass sampling entirely, the total cost of computing the children cache and the subsequent iteration becomes better than the Python + C counterpart, which relies on reachableroots2. For example, on a repo with more than one million revisions with an initial undecided set of 11 elements, we get these figures: Rust version with simple iteration addcommons: 57.287us first undecided computation: 184.278334ms first children cache computation: 131.056us addmissings iteration: 42.766us first addinfo total: 185.24 ms Python + C version first addcommons: 0.29 ms addcommons 0.21 ms first undecided computation 191.35 ms addmissings 45.75 ms first addinfo total: 237.77 ms On discoveries with large undecided sets, the initial price paid makes the first addinfo slower than the Python + C version, but that's more than compensated by the gain in sampling and subsequent iterations. Here's an extreme example with an undecided set of a million revisions: Rust version: first undecided computation: 293.842629ms first children cache computation: 407.911297ms addmissings iteration: 34.312869ms first addinfo total: 776.02 ms taking initial sample query 2: sampling time: 1318.38 ms query 2; still undecided: 1005013, sample size is: 200 addmissings: 143.062us Python + C version: first undecided computation 298.13 ms addmissings 80.13 ms first addinfo total: 399.62 ms taking initial sample query 2: sampling time: 3957.23 ms query 2; still undecided: 1005013, sample size is: 200 addmissings 52.88 ms Differential Revision: https://phab.mercurial-scm.org/D6428
author Georges Racinet <georges.racinet@octobus.net>
date Tue, 16 Apr 2019 01:16:39 +0200
parents 6ed04139ed37
children 2d70b1118af2
line wrap: on
line source

A python hook for "hg fix" that prints out the number of files and revisions
that were affected, along with which fixer tools were applied. Also checks how
many times it sees a specific key generated by one of the fixer tools defined
below.

  $ cat >> $TESTTMP/postfixhook.py <<EOF
  > import collections
  > def file(ui, repo, rev=None, path=b'', metadata=None, **kwargs):
  >   ui.status(b'fixed %s in revision %d using %s\n' %
  >             (path, rev, b', '.join(metadata.keys())))
  > def summarize(ui, repo, replacements=None, wdirwritten=False,
  >               metadata=None, **kwargs):
  >     counts = collections.defaultdict(int)
  >     keys = 0
  >     for fixername, metadatalist in metadata.items():
  >         for metadata in metadatalist:
  >             if metadata is None:
  >                 continue
  >             counts[fixername] += 1
  >             if 'key' in metadata:
  >                 keys += 1
  >     ui.status(b'saw "key" %d times\n' % (keys,))
  >     for name, count in sorted(counts.items()):
  >         ui.status(b'fixed %d files with %s\n' % (count, name))
  >     if replacements:
  >         ui.status(b'fixed %d revisions\n' % (len(replacements),))
  >     if wdirwritten:
  >         ui.status(b'fixed the working copy\n')
  > EOF

Some mock output for fixer tools that demonstrate what could go wrong with
expecting the metadata output format.

  $ printf 'new content\n' > $TESTTMP/missing
  $ printf 'not valid json\0new content\n' > $TESTTMP/invalid
  $ printf '{"key": "value"}\0new content\n' > $TESTTMP/valid

Configure some fixer tools based on the output defined above, and enable the
hooks defined above. Disable parallelism to make output of the parallel file
processing phase stable.

  $ cat >> $HGRCPATH <<EOF
  > [extensions]
  > fix =
  > [fix]
  > missing:command=cat $TESTTMP/missing
  > missing:pattern=missing
  > missing:metadata=true
  > invalid:command=cat $TESTTMP/invalid
  > invalid:pattern=invalid
  > invalid:metadata=true
  > valid:command=cat $TESTTMP/valid
  > valid:pattern=valid
  > valid:metadata=true
  > [hooks]
  > postfixfile = python:$TESTTMP/postfixhook.py:file
  > postfix = python:$TESTTMP/postfixhook.py:summarize
  > [worker]
  > enabled=false
  > EOF

See what happens when we execute each of the fixer tools. Some print warnings,
some write back to the file.

  $ hg init repo
  $ cd repo

  $ printf "old content\n" > invalid
  $ printf "old content\n" > missing
  $ printf "old content\n" > valid
  $ hg add -q

  $ hg fix -w
  ignored invalid output from fixer tool: invalid
  ignored invalid output from fixer tool: missing
  fixed valid in revision 2147483647 using valid
  saw "key" 1 times
  fixed 1 files with valid
  fixed the working copy

  $ cat missing invalid valid
  old content
  old content
  new content

  $ cd ..