Mercurial > hg
view tests/check-perf-code.py @ 30764:e75463e3179f
protocol: send application/mercurial-0.2 responses to capable clients
With this commit, the HTTP transport now parses the X-HgProto-<N>
header to determine what media type and compression engine to use for
responses. So far, we only compress responses that are already being
compressed with zlib today (stream response types to specific
commands). We can expand things to cover additional response types
later.
The practical side-effect of this commit is that non-zlib compression
engines will be used if both ends support them. This means if both
ends have zstd support, zstd - not zlib - will be used to compress
data!
When cloning the mozilla-unified repository between a local HTTP
server and client, the benefits of non-zlib compression are quite
noticeable:
engine server CPU (s) client CPU (s) bundle size
zlib (l=6) 174.1 283.2 1,148,547,026
zstd (l=1) 99.2 267.3 1,127,513,841
zstd (l=3) 103.1 266.9 1,018,861,363
zstd (l=7) 128.3 269.7 919,190,278
zstd (l=10) 162.0 - 894,547,179
none 95.3 277.2 4,097,566,064
The default zstd compression level is 3. So if you deploy zstd
capable Mercurial to your clients and servers and CPU time on
your server is dominated by "getbundle" requests (clients cloning
and pulling) - and my experience at Mozilla tells me this is often
the case - this commit could drastically reduce your server-side
CPU usage *and* save on bandwidth costs!
Another benefit of this change is that server operators can install
*any* compression engine. While it isn't enabled by default, the
"none" compression engine can now be used to disable wire protocol
compression completely. Previously, commands like "getbundle" always
zlib compressed output, adding considerable overhead to generating
responses. If you are on a high speed network and your server is under
high load, it might be advantageous to trade bandwidth for CPU.
Although, zstd at level 1 doesn't use that much CPU, so I'm not
convinced that disabling compression wholesale is worthwhile. And, my
data seems to indicate a slow down on the client without compression.
I suspect this is due to a lack of buffering resulting in an increase
in socket read() calls and/or the fact we're transferring an extra 3 GB
of data (parsing HTTP chunked transfer and processing extra TCP packets
can add up). This is definitely worth investigating and optimizing. But
since the "none" compressor isn't enabled by default, I'm inclined to
punt on this issue.
This commit introduces tons of tests. Some of these should arguably
have been implemented on previous commits. But it was difficult to
test without the server functionality in place.
author | Gregory Szorc <gregory.szorc@gmail.com> |
---|---|
date | Sat, 24 Dec 2016 15:29:32 -0700 |
parents | d8a2c536dd96 |
children | bd872f64a8ba |
line wrap: on
line source
#!/usr/bin/env python # # check-perf-code - (historical) portability checker for contrib/perf.py from __future__ import absolute_import import os import sys # write static check patterns here perfpypats = [ [ (r'(branchmap|repoview)\.subsettable', "use getbranchmapsubsettable() for early Mercurial"), (r'\.(vfs|svfs|opener|sopener)', "use getvfs()/getsvfs() for early Mercurial"), (r'ui\.configint', "use getint() instead of ui.configint() for early Mercurial"), ], # warnings [ ] ] def modulewhitelist(names): replacement = [('.py', ''), ('.c', ''), # trim suffix ('mercurial%s' % (os.sep), ''), # trim "mercurial/" path ] ignored = set(['__init__']) modules = {} # convert from file name to module name, and count # of appearances for name in names: name = name.strip() for old, new in replacement: name = name.replace(old, new) if name not in ignored: modules[name] = modules.get(name, 0) + 1 # list up module names, which appear multiple times whitelist = [] for name, count in modules.items(): if count > 1: whitelist.append(name) return whitelist if __name__ == "__main__": # in this case, it is assumed that result of "hg files" at # multiple revisions is given via stdin whitelist = modulewhitelist(sys.stdin) assert whitelist, "module whitelist is empty" # build up module whitelist check from file names given at runtime perfpypats[0].append( # this matching pattern assumes importing modules from # "mercurial" package in the current style below, for simplicity # # from mercurial import ( # foo, # bar, # baz # ) ((r'from mercurial import [(][a-z0-9, \n#]*\n(?! *%s,|^[ #]*\n|[)])' % ',| *'.join(whitelist)), "import newer module separately in try clause for early Mercurial" )) # import contrib/check-code.py as checkcode assert 'RUNTESTDIR' in os.environ, "use check-perf-code.py in *.t script" contribpath = os.path.join(os.environ['RUNTESTDIR'], '..', 'contrib') sys.path.insert(0, contribpath) checkcode = __import__('check-code') # register perf.py specific entry with "checks" in check-code.py checkcode.checks.append(('perf.py', r'contrib/perf.py$', '', checkcode.pyfilters, perfpypats)) sys.exit(checkcode.main())