Mercurial > hg
view tests/basic_test_result.py @ 42309:604c086ddde6
log: add config for making `hg log -G` always topo-sorted
I (and everyone else at Google) have an log alias that adds graph mode
and templating. I have another one that builds on the first and also
restricts the set of revisions to only show those I'm most likely to
care about. This second alias also adds topological sorting. I still
sometimes use the first one. When I do, it very often bothers me that
it's not topologically sorted (branches are interleaved). This patch
adds a config option for always using topological sorting with graph
log.
The revision set is sorted eagerly, which seems like a bad idea, but
it doesn't seem to make a big difference in the hg repo (150ms). I
initially tried to instead wrap the user's revset in sort(...,topo),
but that seemed much harder.
Differential Revision: https://phab.mercurial-scm.org/D6331
author | Martin von Zweigbergk <martinvonz@google.com> |
---|---|
date | Wed, 01 May 2019 09:34:47 -0700 |
parents | f4a214300957 |
children | 2372284d9457 |
line wrap: on
line source
from __future__ import absolute_import, print_function import unittest class TestResult(unittest._TextTestResult): def __init__(self, options, *args, **kwargs): super(TestResult, self).__init__(*args, **kwargs) self._options = options # unittest.TestResult didn't have skipped until 2.7. We need to # polyfill it. self.skipped = [] # We have a custom "ignored" result that isn't present in any Python # unittest implementation. It is very similar to skipped. It may make # sense to map it into skip some day. self.ignored = [] self.times = [] self._firststarttime = None # Data stored for the benefit of generating xunit reports. self.successes = [] self.faildata = {} def addFailure(self, test, reason): print("FAILURE!", test, reason) def addSuccess(self, test): print("SUCCESS!", test) def addError(self, test, err): print("ERR!", test, err) # Polyfill. def addSkip(self, test, reason): print("SKIP!", test, reason) def addIgnore(self, test, reason): print("IGNORE!", test, reason) def onStart(self, test): print("ON_START!", test) def onEnd(self): print("ON_END!") def addOutputMismatch(self, test, ret, got, expected): return False def stopTest(self, test, interrupted=False): super(TestResult, self).stopTest(test)