Sun, 20 Apr 2014 15:16:29 -0700 run-tests: move some functionality to Test.tearDown()
Gregory Szorc <gregory.szorc@gmail.com> [Sun, 20 Apr 2014 15:16:29 -0700] rev 21452
run-tests: move some functionality to Test.tearDown()
Sun, 20 Apr 2014 15:08:05 -0700 run-tests: store last result in Test._result
Gregory Szorc <gregory.szorc@gmail.com> [Sun, 20 Apr 2014 15:08:05 -0700] rev 21451
run-tests: store last result in Test._result Subsequent patches move post-test actions into tearDown(). This requires tearDown() to have access to the last result.
Sun, 20 Apr 2014 15:00:13 -0700 run-tests: report test duration in Test.tearDown()
Gregory Szorc <gregory.szorc@gmail.com> [Sun, 20 Apr 2014 15:00:13 -0700] rev 21450
run-tests: report test duration in Test.tearDown()
Sun, 20 Apr 2014 14:58:36 -0700 run-tests: store test return code and output in Test instance
Gregory Szorc <gregory.szorc@gmail.com> [Sun, 20 Apr 2014 14:58:36 -0700] rev 21449
run-tests: store test return code and output in Test instance A subsequent patch that will introduce tearDown() requires test execution state be stored in the instance.
Sun, 20 Apr 2014 14:55:33 -0700 run-tests: keep track of test start and stop in MercurialTest.run()
Gregory Szorc <gregory.szorc@gmail.com> [Sun, 20 Apr 2014 14:55:33 -0700] rev 21448
run-tests: keep track of test start and stop in MercurialTest.run() This brings run() more compatible with unittest.TestCase.run()
Sun, 20 Apr 2014 14:52:57 -0700 run-tests: keep track of test execution state in Test
Gregory Szorc <gregory.szorc@gmail.com> [Sun, 20 Apr 2014 14:52:57 -0700] rev 21447
run-tests: keep track of test execution state in Test This patch starts a mini series of moving functionality into the newly-established setUp() and tearDown() methods.
Sun, 20 Apr 2014 14:41:11 -0700 run-tests: support setUp() and tearDown() in TestCase wrapper
Gregory Szorc <gregory.szorc@gmail.com> [Sun, 20 Apr 2014 14:41:11 -0700] rev 21446
run-tests: support setUp() and tearDown() in TestCase wrapper unittest.TestCase.run() calls setUp() and tearDown() during run(). We emulate that implementation.
Sun, 20 Apr 2014 14:34:03 -0700 run-tests: fail tests by raising an exception
Gregory Szorc <gregory.szorc@gmail.com> [Sun, 20 Apr 2014 14:34:03 -0700] rev 21445
run-tests: fail tests by raising an exception When in unittest mode, Test.run() will now raise for all non-success cases. This makes it behave like TestCase.run().
Sun, 20 Apr 2014 14:32:03 -0700 run-tests: record warnings by raising WarnTest
Gregory Szorc <gregory.szorc@gmail.com> [Sun, 20 Apr 2014 14:32:03 -0700] rev 21444
run-tests: record warnings by raising WarnTest We continue the conversion of recording test results by raising exceptions.
Sun, 20 Apr 2014 14:28:29 -0700 run-tests: record ignored tests by raising IgnoreTest
Gregory Szorc <gregory.szorc@gmail.com> [Sun, 20 Apr 2014 14:28:29 -0700] rev 21443
run-tests: record ignored tests by raising IgnoreTest
Sun, 20 Apr 2014 14:23:50 -0700 run-tests: record skips by raising SkipTest
Gregory Szorc <gregory.szorc@gmail.com> [Sun, 20 Apr 2014 14:23:50 -0700] rev 21442
run-tests: record skips by raising SkipTest The unittest way of recording a skipped test is to raise an exception (at least with modern unittest implementation). We change Test to raise a SkipTest when operating in unittest mode. This does prevent some "tear down" activities from running in unittest mode. This will be fixed in subsequent patches. Since unittest mode is experimental, this should be OK.
Sun, 20 Apr 2014 14:19:59 -0700 run-tests: implement TestCase.run()
Gregory Szorc <gregory.szorc@gmail.com> [Sun, 20 Apr 2014 14:19:59 -0700] rev 21441
run-tests: implement TestCase.run() Simply wrapping TestCase.run() is not sufficient for robust results reporting because unittest in Python 2.4 does not know about things like skipped tests and reports them as success or failures instead of skips. We will reimplement TestCase.run() with knowledge and semantics present in modern Python releases.
Sun, 20 Apr 2014 14:04:37 -0700 run-tests: don't print progress from Test when in unittest mode
Gregory Szorc <gregory.szorc@gmail.com> [Sun, 20 Apr 2014 14:04:37 -0700] rev 21440
run-tests: don't print progress from Test when in unittest mode unittest does its own printing of progress indicators as part of TestResult. So, there is no need to print them when running in unittest mode. This will fix the double output of progress indicators that had been occurring in unittest mode.
Sun, 20 Apr 2014 13:04:19 -0700 run-tests: define a custom TestSuite that uses _executetests()
Gregory Szorc <gregory.szorc@gmail.com> [Sun, 20 Apr 2014 13:04:19 -0700] rev 21439
run-tests: define a custom TestSuite that uses _executetests() We now have a custom unittest.TestSuite implementation that uses _executetests() and thus knows how to execute tests concurrently. Running tests in --unittest mode will use this TestSuite. Since the TestSuite handles concurrency, the warnings around --jobs and --loop have been removed.
Sun, 20 Apr 2014 13:00:40 -0700 run-tests: pass an optional TestResult into _executetests()
Gregory Szorc <gregory.szorc@gmail.com> [Sun, 20 Apr 2014 13:00:40 -0700] rev 21438
run-tests: pass an optional TestResult into _executetests() If the result is passed, we execute tests in the unittest way. A subsequent patch will actually do this.
Sun, 20 Apr 2014 12:49:43 -0700 run-tests: pass Test instances into TestRunner._executetests()
Gregory Szorc <gregory.szorc@gmail.com> [Sun, 20 Apr 2014 12:49:43 -0700] rev 21437
run-tests: pass Test instances into TestRunner._executetests() Test instances (not paths) are passed into _executetests(). This means the logic for instantiating Test instances has been moved outside of _executetests(). This was done because an upcoming patch will reuse the logic in _executetests(). As part of this change, test counts are no longer managed by _executetests().
Sat, 17 May 2014 12:35:44 -0700 run-tests: define cleanup() on unittest class
Gregory Szorc <gregory.szorc@gmail.com> [Sat, 17 May 2014 12:35:44 -0700] rev 21436
run-tests: define cleanup() on unittest class We need a cleanup() on our unittest TestCase class to conform with the execution API. This will eventually be replaced by tearDown().
Sun, 20 Apr 2014 12:45:44 -0700 run-tests: rename Test._test to Test.name
Gregory Szorc <gregory.szorc@gmail.com> [Sun, 20 Apr 2014 12:45:44 -0700] rev 21435
run-tests: rename Test._test to Test.name A lot of the attributes in Test could probably be renamed. This one is needed for an upcoming patch.
Sun, 20 Apr 2014 12:19:48 -0700 run-tests: teach Test instances whether they are in unittest mode
Gregory Szorc <gregory.szorc@gmail.com> [Sun, 20 Apr 2014 12:19:48 -0700] rev 21434
run-tests: teach Test instances whether they are in unittest mode Tests executing in unittest mode behave a little differently. For example, they report their results to a TestResult rather than just printing. This patch paves the way for better integration of Test into the unittest framework.
Sun, 20 Apr 2014 11:55:02 -0700 run-tests: teach unittest about warned results
Gregory Szorc <gregory.szorc@gmail.com> [Sun, 20 Apr 2014 11:55:02 -0700] rev 21433
run-tests: teach unittest about warned results
Sun, 20 Apr 2014 11:52:41 -0700 run-tests: document how success code is handled
Gregory Szorc <gregory.szorc@gmail.com> [Sun, 20 Apr 2014 11:52:41 -0700] rev 21432
run-tests: document how success code is handled
Sun, 20 Apr 2014 11:51:11 -0700 run-tests: teach unittest about ignored tests
Gregory Szorc <gregory.szorc@gmail.com> [Sun, 20 Apr 2014 11:51:11 -0700] rev 21431
run-tests: teach unittest about ignored tests
Sun, 20 Apr 2014 11:48:19 -0700 run-tests: teach unittest about skipped tests
Gregory Szorc <gregory.szorc@gmail.com> [Sun, 20 Apr 2014 11:48:19 -0700] rev 21430
run-tests: teach unittest about skipped tests
Sun, 20 Apr 2014 11:47:40 -0700 run-tests: define custom result and runner classes for unittest
Gregory Szorc <gregory.szorc@gmail.com> [Sun, 20 Apr 2014 11:47:40 -0700] rev 21429
run-tests: define custom result and runner classes for unittest We need to teach unittest about some custom result types. This will require some custom classes. This patch creates a skeleton for them.
Sun, 20 Apr 2014 11:29:39 -0700 run-tests: start to report test results against TestResult
Gregory Szorc <gregory.szorc@gmail.com> [Sun, 20 Apr 2014 11:29:39 -0700] rev 21428
run-tests: start to report test results against TestResult Previously, our unittest wrapper didn't report results properly. We now properly report failures. We had to rename the local variable to prevent "t" from being overwritten in the local scope.
Sun, 20 Apr 2014 11:24:37 -0700 run-tests: give unittest wrapper a short description
Gregory Szorc <gregory.szorc@gmail.com> [Sun, 20 Apr 2014 11:24:37 -0700] rev 21427
run-tests: give unittest wrapper a short description This will make the output contain something useful rather than the default repr() for the class.
Sun, 20 Apr 2014 11:22:08 -0700 run-tests: initial support for running tests with unittest
Gregory Szorc <gregory.szorc@gmail.com> [Sun, 20 Apr 2014 11:22:08 -0700] rev 21426
run-tests: initial support for running tests with unittest The unittest package in Python's standard library provides an almost universal mechanism for defining and running tests. This patch starts the process of making run-tests.py talk to it. The main benefit of speaking unittest is that this will enable Mercurial's tests to be more easily consumed by other testing tools, like nose. This is useful for 3rd party extensions having their own test infrastructure, for example. Running tests in unittest mode will not result in completely sane behavior until the unittest mode is made the default execution mode. Expect things like double printing of output until support stabilizes.
Wed, 16 Apr 2014 16:34:48 -0500 localrepo: reverse contexts in status
Sean Farley <sean.michael.farley@gmail.com> [Wed, 16 Apr 2014 16:34:48 -0500] rev 21425
localrepo: reverse contexts in status This is a slight tweak to how localrepo.status calculates what files have changed. By forcing a changectx to be first operator and anything not a changectx to be the second operator, we can later exploit this to allow refactoring the status operation as a method of a context object. Furthermore, this change will speed up 'hg diff --reverse' when used with the working directory because the code will now hit a fast path without needing to calculate an unneeded second manifest.
Fri, 16 May 2014 13:18:57 -0700 test: split test-largefile.t in multiple file
Pierre-Yves David <pierre-yves.david@fb.com> [Fri, 16 May 2014 13:18:57 -0700] rev 21424
test: split test-largefile.t in multiple file The `test-largefiles.t` unified test is significantly longer (about 30%) than any other tests in the mercurial test suite. As a result, its is alway the last test my test runner is waiting for at the end of a run. In practice, this means that `test-largefile.t` is wasting half a minute of my life every times I'm running the mercurial test suites. This probably mean more a few cumulated day by now. I've finally decided to split it up in multiple smaller tests to bring it back in reasonable length. This changeset extracts independent test cases in two files. One dedicated to wire protocole testing, and another one dedicated to all other tests that could be independently extracted. No test case were haltered in the making of this changeset. Various timing available below. All timing have been done on a with 90 jobs on a 64 cores machine. Similar result are shown on firefly (20 jobs on 12 core). General timing of the whole run -------------------------------- We see a 25% real time improvement for no significant cpu time impact. Before split: real 2m1.149s user 58m4.662s sys 11m28.563s After split: real 1m31.977s user 57m45.993s sys 11m33.634s Last test to finish (using run-test.py --time) ---------------------------------------------- test-largefile.t is now finishing at the same time than other slow tests. Before split: Time Test 119.280 test-largefiles.t 93.995 test-mq.t 89.897 test-subrepo.t 86.920 test-glog.t 85.508 test-rename-merge2.t 83.594 test-revset.t 79.824 test-keyword.t 78.077 test-mq-header-date.t After split: Time Test 90.414 test-mq.t 88.594 test-largefiles.t 85.363 test-subrepo.t 81.059 test-glog.t 78.927 test-rename-merge2.t 78.021 test-revset.t 77.777 test-command-template.t Timing of largefile test themself ----------------------------------- Running only tests prefixed with "test-largefiles". No significant change in cumulated time. Before: Time Test 58.673 test-largefiles.t 2.931 test-largefiles-cache.t 0.583 test-largefiles-small-disk.t After: Time Test 31.754 test-largefiles.t 17.460 test-largefiles-misc.t 8.888 test-largefiles-wireproto.t 2.864 test-largefiles-cache.t 0.580 test-largefiles-small-disk.t
Sun, 11 May 2014 00:49:36 +0900 mq: use the editor gotten by "getcommiteditor()" instead of "ui.edit()" (qrefresh/qfold)
FUJIWARA Katsunori <foozy@lares.dti.ne.jp> [Sun, 11 May 2014 00:49:36 +0900] rev 21423
mq: use the editor gotten by "getcommiteditor()" instead of "ui.edit()" (qrefresh/qfold) This patch also replaces "editor = False" by "editor = cmdutil.getcommiteditor()", because: - it allows to hook commit message determination easily, even in the case without "--edit" - it avoids regression (or additional care) around saving "last-message.txt", even if MQ's "newcommit()" changes its implementation logic from "localrepository.commit" to "localrepository.commitctx" with "memctx" in the future to save commit message into "last-messge.txt" with "memctx", "editor" should be valid function.
(0) -10000 -3000 -1000 -300 -100 -50 -30 +30 +50 +100 +300 +1000 +3000 +10000 +30000 tip