Mercurial > hg-stable
changeset 42021:4c700c847aa2
perf: copyedit a few documentation strings
Differential Revision: https://phab.mercurial-scm.org/D6170
author | Augie Fackler <augie@google.com> |
---|---|
date | Tue, 26 Mar 2019 10:21:17 -0400 |
parents | 00c1ee0f746a |
children | 1f567a9e25bd |
files | contrib/perf.py tests/test-contrib-perf.t |
diffstat | 2 files changed, 16 insertions(+), 15 deletions(-) [+] |
line wrap: on
line diff
--- a/contrib/perf.py Sun Mar 24 20:13:13 2019 -0400 +++ b/contrib/perf.py Tue Mar 26 10:21:17 2019 -0400 @@ -8,25 +8,26 @@ -------- ``all-timing`` - When set, additional statistic will be reported for each benchmark: best, + When set, additional statistics will be reported for each benchmark: best, worst, median average. If not set only the best timing is reported (default: off). ``presleep`` - number of second to wait before any group of run (default: 1) + number of second to wait before any group of runs (default: 1) ``run-limits`` - Control the number of run each benchmark will perform. The option value + Control the number of runs each benchmark will perform. The option value should be a list of `<time>-<numberofrun>` pairs. After each run the - condition are considered in order with the following logic: - - If benchmark have been running for <time> seconds, and we have performed + conditions are considered in order with the following logic: + + If benchmark has been running for <time> seconds, and we have performed <numberofrun> iterations, stop the benchmark, The default value is: `3.0-100, 10.0-3` ``stub`` - When set, benchmark will only be run once, useful for testing (default: off) + When set, benchmarks will only be run once, useful for testing + (default: off) ''' # "historical portability" policy of perf.py: @@ -1217,8 +1218,8 @@ def perfparents(ui, repo, **opts): """benchmark the time necessary to fetch one changeset's parents. - The fetch is done using the `node identifier`, traversing all object layer - from the repository object. The N first revision will be used for this + The fetch is done using the `node identifier`, traversing all object layers + from the repository object. The first N revisions will be used for this benchmark. N is controlled by the ``perf.parentscount`` config option (default: 1000). """
--- a/tests/test-contrib-perf.t Sun Mar 24 20:13:13 2019 -0400 +++ b/tests/test-contrib-perf.t Tue Mar 26 10:21:17 2019 -0400 @@ -48,25 +48,25 @@ ------ "all-timing" - When set, additional statistic will be reported for each benchmark: best, + When set, additional statistics will be reported for each benchmark: best, worst, median average. If not set only the best timing is reported (default: off). "presleep" - number of second to wait before any group of run (default: 1) + number of second to wait before any group of runs (default: 1) "run-limits" - Control the number of run each benchmark will perform. The option value + Control the number of runs each benchmark will perform. The option value should be a list of '<time>-<numberofrun>' pairs. After each run the - condition are considered in order with the following logic: + conditions are considered in order with the following logic: - If benchmark have been running for <time> seconds, and we have performed + If benchmark has been running for <time> seconds, and we have performed <numberofrun> iterations, stop the benchmark, The default value is: '3.0-100, 10.0-3' "stub" - When set, benchmark will only be run once, useful for testing (default: + When set, benchmarks will only be run once, useful for testing (default: off) list of commands: