worker: avoid potential partial write of pickled data
Previously, the code wrote the pickled data using os.write(). However,
os.write() can write less bytes than passed to it. To trigger the problem, the
pickled data had to be larger than
2147479552 bytes on my system.
Instead, open a file object and pass it to pickle.dump(). This also has the
advantage that it doesn’t buffer the whole pickled data in memory.
Note that the opened file must be buffered because pickle doesn’t support
unbuffered streams because unbuffered streams’ write() method might write less
bytes than passed to it (like os.write()) but pickle.dump() relies on that all
bytes are written (see https://github.com/python/cpython/issues/93050).
The side effect of using a file object and a with statement is that wfd is
explicitly closed now while it seems like before it was implicitly closed by
process exit.
$ hg init a
$ cd a
$ touch empty1
$ hg add empty1
$ hg commit -m 'add empty1'
$ touch empty2
$ hg add empty2
$ hg commit -m 'add empty2'
$ hg up -C 0
0 files updated, 0 files merged, 1 files removed, 0 files unresolved
$ touch empty3
$ hg add empty3
$ hg commit -m 'add empty3'
created new head
$ hg heads
changeset: 2:a1cb177e0d44
tag: tip
parent: 0:1e1d9c4e5b64
user: test
date: Thu Jan 01 00:00:00 1970 +0000
summary: add empty3
changeset: 1:097d2b0e17f6
user: test
date: Thu Jan 01 00:00:00 1970 +0000
summary: add empty2
$ hg merge 1
1 files updated, 0 files merged, 0 files removed, 0 files unresolved
(branch merge, don't forget to commit)
Before changeset 05257fd28591, we didn't notice the
empty file that came from rev 1:
$ hg status
M empty2
$ hg commit -m merge
$ hg manifest --debug tip
b80de5d138758541c5f05265ad144ab9fa86d1db 644 empty1
b80de5d138758541c5f05265ad144ab9fa86d1db 644 empty2
b80de5d138758541c5f05265ad144ab9fa86d1db 644 empty3
$ cd ..