Mercurial > hg
view mercurial/util.py @ 7604:7c4765d51c3c
tests: shrink wait interval for test-convert-bzr-merges
author | Dirkjan Ochtman <dirkjan@ochtman.nl> |
---|---|
date | Mon, 05 Jan 2009 18:45:20 +0100 |
parents | 016a7319e76b |
children | 9626819b2e3d |
line wrap: on
line source
""" util.py - Mercurial utility functions and platform specfic implementations Copyright 2005 K. Thananchayan <thananck@yahoo.com> Copyright 2005-2007 Matt Mackall <mpm@selenic.com> Copyright 2006 Vadim Gelfer <vadim.gelfer@gmail.com> This software may be used and distributed according to the terms of the GNU General Public License, incorporated herein by reference. This contains helper routines that are independent of the SCM core and hide platform-specific details from the core. """ from i18n import _ import cStringIO, errno, getpass, re, shutil, sys, tempfile, traceback import os, stat, threading, time, calendar, ConfigParser, locale, glob, osutil import imp # Python compatibility try: set = set frozenset = frozenset except NameError: from sets import Set as set, ImmutableSet as frozenset _md5 = None def md5(s): global _md5 if _md5 is None: try: import hashlib _md5 = hashlib.md5 except ImportError: import md5 _md5 = md5.md5 return _md5(s) _sha1 = None def sha1(s): global _sha1 if _sha1 is None: try: import hashlib _sha1 = hashlib.sha1 except ImportError: import sha _sha1 = sha.sha return _sha1(s) try: import subprocess subprocess.Popen # trigger ImportError early closefds = os.name == 'posix' def popen2(cmd, mode='t', bufsize=-1): p = subprocess.Popen(cmd, shell=True, bufsize=bufsize, close_fds=closefds, stdin=subprocess.PIPE, stdout=subprocess.PIPE) return p.stdin, p.stdout def popen3(cmd, mode='t', bufsize=-1): p = subprocess.Popen(cmd, shell=True, bufsize=bufsize, close_fds=closefds, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE) return p.stdin, p.stdout, p.stderr def Popen3(cmd, capturestderr=False, bufsize=-1): stderr = capturestderr and subprocess.PIPE or None p = subprocess.Popen(cmd, shell=True, bufsize=bufsize, close_fds=closefds, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=stderr) p.fromchild = p.stdout p.tochild = p.stdin p.childerr = p.stderr return p except ImportError: subprocess = None from popen2 import Popen3 popen2 = os.popen2 popen3 = os.popen3 _encodingfixup = {'646': 'ascii', 'ANSI_X3.4-1968': 'ascii'} try: _encoding = os.environ.get("HGENCODING") if sys.platform == 'darwin' and not _encoding: # On darwin, getpreferredencoding ignores the locale environment and # always returns mac-roman. We override this if the environment is # not C (has been customized by the user). locale.setlocale(locale.LC_CTYPE, '') _encoding = locale.getlocale()[1] if not _encoding: _encoding = locale.getpreferredencoding() or 'ascii' _encoding = _encodingfixup.get(_encoding, _encoding) except locale.Error: _encoding = 'ascii' _encodingmode = os.environ.get("HGENCODINGMODE", "strict") _fallbackencoding = 'ISO-8859-1' def tolocal(s): """ Convert a string from internal UTF-8 to local encoding All internal strings should be UTF-8 but some repos before the implementation of locale support may contain latin1 or possibly other character sets. We attempt to decode everything strictly using UTF-8, then Latin-1, and failing that, we use UTF-8 and replace unknown characters. """ for e in ('UTF-8', _fallbackencoding): try: u = s.decode(e) # attempt strict decoding return u.encode(_encoding, "replace") except LookupError, k: raise Abort(_("%s, please check your locale settings") % k) except UnicodeDecodeError: pass u = s.decode("utf-8", "replace") # last ditch return u.encode(_encoding, "replace") def fromlocal(s): """ Convert a string from the local character encoding to UTF-8 We attempt to decode strings using the encoding mode set by HGENCODINGMODE, which defaults to 'strict'. In this mode, unknown characters will cause an error message. Other modes include 'replace', which replaces unknown characters with a special Unicode character, and 'ignore', which drops the character. """ try: return s.decode(_encoding, _encodingmode).encode("utf-8") except UnicodeDecodeError, inst: sub = s[max(0, inst.start-10):inst.start+10] raise Abort("decoding near '%s': %s!" % (sub, inst)) except LookupError, k: raise Abort(_("%s, please check your locale settings") % k) def locallen(s): """Find the length in characters of a local string""" return len(s.decode(_encoding, "replace")) # used by parsedate defaultdateformats = ( '%Y-%m-%d %H:%M:%S', '%Y-%m-%d %I:%M:%S%p', '%Y-%m-%d %H:%M', '%Y-%m-%d %I:%M%p', '%Y-%m-%d', '%m-%d', '%m/%d', '%m/%d/%y', '%m/%d/%Y', '%a %b %d %H:%M:%S %Y', '%a %b %d %I:%M:%S%p %Y', '%a, %d %b %Y %H:%M:%S', # GNU coreutils "/bin/date --rfc-2822" '%b %d %H:%M:%S %Y', '%b %d %I:%M:%S%p %Y', '%b %d %H:%M:%S', '%b %d %I:%M:%S%p', '%b %d %H:%M', '%b %d %I:%M%p', '%b %d %Y', '%b %d', '%H:%M:%S', '%I:%M:%SP', '%H:%M', '%I:%M%p', ) extendeddateformats = defaultdateformats + ( "%Y", "%Y-%m", "%b", "%b %Y", ) class SignalInterrupt(Exception): """Exception raised on SIGTERM and SIGHUP.""" # differences from SafeConfigParser: # - case-sensitive keys # - allows values that are not strings (this means that you may not # be able to save the configuration to a file) class configparser(ConfigParser.SafeConfigParser): def optionxform(self, optionstr): return optionstr def set(self, section, option, value): return ConfigParser.ConfigParser.set(self, section, option, value) def _interpolate(self, section, option, rawval, vars): if not isinstance(rawval, basestring): return rawval return ConfigParser.SafeConfigParser._interpolate(self, section, option, rawval, vars) def cachefunc(func): '''cache the result of function calls''' # XXX doesn't handle keywords args cache = {} if func.func_code.co_argcount == 1: # we gain a small amount of time because # we don't need to pack/unpack the list def f(arg): if arg not in cache: cache[arg] = func(arg) return cache[arg] else: def f(*args): if args not in cache: cache[args] = func(*args) return cache[args] return f def pipefilter(s, cmd): '''filter string S through command CMD, returning its output''' (pin, pout) = popen2(cmd, 'b') def writer(): try: pin.write(s) pin.close() except IOError, inst: if inst.errno != errno.EPIPE: raise # we should use select instead on UNIX, but this will work on most # systems, including Windows w = threading.Thread(target=writer) w.start() f = pout.read() pout.close() w.join() return f def tempfilter(s, cmd): '''filter string S through a pair of temporary files with CMD. CMD is used as a template to create the real command to be run, with the strings INFILE and OUTFILE replaced by the real names of the temporary files generated.''' inname, outname = None, None try: infd, inname = tempfile.mkstemp(prefix='hg-filter-in-') fp = os.fdopen(infd, 'wb') fp.write(s) fp.close() outfd, outname = tempfile.mkstemp(prefix='hg-filter-out-') os.close(outfd) cmd = cmd.replace('INFILE', inname) cmd = cmd.replace('OUTFILE', outname) code = os.system(cmd) if sys.platform == 'OpenVMS' and code & 1: code = 0 if code: raise Abort(_("command '%s' failed: %s") % (cmd, explain_exit(code))) return open(outname, 'rb').read() finally: try: if inname: os.unlink(inname) except: pass try: if outname: os.unlink(outname) except: pass filtertable = { 'tempfile:': tempfilter, 'pipe:': pipefilter, } def filter(s, cmd): "filter a string through a command that transforms its input to its output" for name, fn in filtertable.iteritems(): if cmd.startswith(name): return fn(s, cmd[len(name):].lstrip()) return pipefilter(s, cmd) def binary(s): """return true if a string is binary data""" if s and '\0' in s: return True return False def unique(g): """return the uniq elements of iterable g""" return dict.fromkeys(g).keys() def sort(l): if not isinstance(l, list): l = list(l) l.sort() return l def increasingchunks(source, min=1024, max=65536): '''return no less than min bytes per chunk while data remains, doubling min after each chunk until it reaches max''' def log2(x): if not x: return 0 i = 0 while x: x >>= 1 i += 1 return i - 1 buf = [] blen = 0 for chunk in source: buf.append(chunk) blen += len(chunk) if blen >= min: if min < max: min = min << 1 nmin = 1 << log2(blen) if nmin > min: min = nmin if min > max: min = max yield ''.join(buf) blen = 0 buf = [] if buf: yield ''.join(buf) class Abort(Exception): """Raised if a command needs to print an error and exit.""" class UnexpectedOutput(Abort): """Raised to print an error with part of output and exit.""" def always(fn): return True def never(fn): return False def expand_glob(pats): '''On Windows, expand the implicit globs in a list of patterns''' if os.name != 'nt': return list(pats) ret = [] for p in pats: kind, name = patkind(p, None) if kind is None: globbed = glob.glob(name) if globbed: ret.extend(globbed) continue # if we couldn't expand the glob, just keep it around ret.append(p) return ret def patkind(name, default): """Split a string into an optional pattern kind prefix and the actual pattern.""" for prefix in 're', 'glob', 'path', 'relglob', 'relpath', 'relre': if name.startswith(prefix + ':'): return name.split(':', 1) return default, name def globre(pat, head='^', tail='$'): "convert a glob pattern into a regexp" i, n = 0, len(pat) res = '' group = 0 def peek(): return i < n and pat[i] while i < n: c = pat[i] i = i+1 if c == '*': if peek() == '*': i += 1 res += '.*' else: res += '[^/]*' elif c == '?': res += '.' elif c == '[': j = i if j < n and pat[j] in '!]': j += 1 while j < n and pat[j] != ']': j += 1 if j >= n: res += '\\[' else: stuff = pat[i:j].replace('\\','\\\\') i = j + 1 if stuff[0] == '!': stuff = '^' + stuff[1:] elif stuff[0] == '^': stuff = '\\' + stuff res = '%s[%s]' % (res, stuff) elif c == '{': group += 1 res += '(?:' elif c == '}' and group: res += ')' group -= 1 elif c == ',' and group: res += '|' elif c == '\\': p = peek() if p: i += 1 res += re.escape(p) else: res += re.escape(c) else: res += re.escape(c) return head + res + tail _globchars = {'[': 1, '{': 1, '*': 1, '?': 1} def pathto(root, n1, n2): '''return the relative path from one place to another. root should use os.sep to separate directories n1 should use os.sep to separate directories n2 should use "/" to separate directories returns an os.sep-separated path. If n1 is a relative path, it's assumed it's relative to root. n2 should always be relative to root. ''' if not n1: return localpath(n2) if os.path.isabs(n1): if os.path.splitdrive(root)[0] != os.path.splitdrive(n1)[0]: return os.path.join(root, localpath(n2)) n2 = '/'.join((pconvert(root), n2)) a, b = splitpath(n1), n2.split('/') a.reverse() b.reverse() while a and b and a[-1] == b[-1]: a.pop() b.pop() b.reverse() return os.sep.join((['..'] * len(a)) + b) or '.' def canonpath(root, cwd, myname): """return the canonical path of myname, given cwd and root""" if root == os.sep: rootsep = os.sep elif endswithsep(root): rootsep = root else: rootsep = root + os.sep name = myname if not os.path.isabs(name): name = os.path.join(root, cwd, name) name = os.path.normpath(name) audit_path = path_auditor(root) if name != rootsep and name.startswith(rootsep): name = name[len(rootsep):] audit_path(name) return pconvert(name) elif name == root: return '' else: # Determine whether `name' is in the hierarchy at or beneath `root', # by iterating name=dirname(name) until that causes no change (can't # check name == '/', because that doesn't work on windows). For each # `name', compare dev/inode numbers. If they match, the list `rel' # holds the reversed list of components making up the relative file # name we want. root_st = os.stat(root) rel = [] while True: try: name_st = os.stat(name) except OSError: break if samestat(name_st, root_st): if not rel: # name was actually the same as root (maybe a symlink) return '' rel.reverse() name = os.path.join(*rel) audit_path(name) return pconvert(name) dirname, basename = os.path.split(name) rel.append(basename) if dirname == name: break name = dirname raise Abort('%s not under root' % myname) def matcher(canonroot, cwd='', names=[], inc=[], exc=[], src=None, dflt_pat='glob'): """build a function to match a set of file patterns arguments: canonroot - the canonical root of the tree you're matching against cwd - the current working directory, if relevant names - patterns to find inc - patterns to include exc - patterns to exclude dflt_pat - if a pattern in names has no explicit type, assume this one src - where these patterns came from (e.g. .hgignore) a pattern is one of: 'glob:<glob>' - a glob relative to cwd 're:<regexp>' - a regular expression 'path:<path>' - a path relative to canonroot 'relglob:<glob>' - an unrooted glob (*.c matches C files in all dirs) 'relpath:<path>' - a path relative to cwd 'relre:<regexp>' - a regexp that doesn't have to match the start of a name '<something>' - one of the cases above, selected by the dflt_pat argument returns: a 3-tuple containing - list of roots (places where one should start a recursive walk of the fs); this often matches the explicit non-pattern names passed in, but also includes the initial part of glob: patterns that has no glob characters - a bool match(filename) function - a bool indicating if any patterns were passed in """ # a common case: no patterns at all if not names and not inc and not exc: return [], always, False def contains_glob(name): for c in name: if c in _globchars: return True return False def regex(kind, name, tail): '''convert a pattern into a regular expression''' if not name: return '' if kind == 're': return name elif kind == 'path': return '^' + re.escape(name) + '(?:/|$)' elif kind == 'relglob': return globre(name, '(?:|.*/)', tail) elif kind == 'relpath': return re.escape(name) + '(?:/|$)' elif kind == 'relre': if name.startswith('^'): return name return '.*' + name return globre(name, '', tail) def matchfn(pats, tail): """build a matching function from a set of patterns""" if not pats: return try: pat = '(?:%s)' % '|'.join([regex(k, p, tail) for (k, p) in pats]) if len(pat) > 20000: raise OverflowError() return re.compile(pat).match except OverflowError: # We're using a Python with a tiny regex engine and we # made it explode, so we'll divide the pattern list in two # until it works l = len(pats) if l < 2: raise a, b = matchfn(pats[:l//2], tail), matchfn(pats[l//2:], tail) return lambda s: a(s) or b(s) except re.error: for k, p in pats: try: re.compile('(?:%s)' % regex(k, p, tail)) except re.error: if src: raise Abort("%s: invalid pattern (%s): %s" % (src, k, p)) else: raise Abort("invalid pattern (%s): %s" % (k, p)) raise Abort("invalid pattern") def globprefix(pat): '''return the non-glob prefix of a path, e.g. foo/* -> foo''' root = [] for p in pat.split('/'): if contains_glob(p): break root.append(p) return '/'.join(root) or '.' def normalizepats(names, default): pats = [] roots = [] anypats = False for kind, name in [patkind(p, default) for p in names]: if kind in ('glob', 'relpath'): name = canonpath(canonroot, cwd, name) elif kind in ('relglob', 'path'): name = normpath(name) pats.append((kind, name)) if kind in ('glob', 're', 'relglob', 'relre'): anypats = True if kind == 'glob': root = globprefix(name) roots.append(root) elif kind in ('relpath', 'path'): roots.append(name or '.') elif kind == 'relglob': roots.append('.') return roots, pats, anypats roots, pats, anypats = normalizepats(names, dflt_pat) patmatch = matchfn(pats, '$') or always incmatch = always if inc: dummy, inckinds, dummy = normalizepats(inc, 'glob') incmatch = matchfn(inckinds, '(?:/|$)') excmatch = never if exc: dummy, exckinds, dummy = normalizepats(exc, 'glob') excmatch = matchfn(exckinds, '(?:/|$)') if not names and inc and not exc: # common case: hgignore patterns match = incmatch else: match = lambda fn: incmatch(fn) and not excmatch(fn) and patmatch(fn) return (roots, match, (inc or exc or anypats) and True) _hgexecutable = None def main_is_frozen(): """return True if we are a frozen executable. The code supports py2exe (most common, Windows only) and tools/freeze (portable, not much used). """ return (hasattr(sys, "frozen") or # new py2exe hasattr(sys, "importers") or # old py2exe imp.is_frozen("__main__")) # tools/freeze def hgexecutable(): """return location of the 'hg' executable. Defaults to $HG or 'hg' in the search path. """ if _hgexecutable is None: hg = os.environ.get('HG') if hg: set_hgexecutable(hg) elif main_is_frozen(): set_hgexecutable(sys.executable) else: set_hgexecutable(find_exe('hg', 'hg')) return _hgexecutable def set_hgexecutable(path): """set location of the 'hg' executable""" global _hgexecutable _hgexecutable = path def system(cmd, environ={}, cwd=None, onerr=None, errprefix=None): '''enhanced shell command execution. run with environment maybe modified, maybe in different dir. if command fails and onerr is None, return status. if ui object, print error message and return status, else raise onerr object as exception.''' def py2shell(val): 'convert python object into string that is useful to shell' if val in (None, False): return '0' if val == True: return '1' return str(val) oldenv = {} for k in environ: oldenv[k] = os.environ.get(k) if cwd is not None: oldcwd = os.getcwd() origcmd = cmd if os.name == 'nt': cmd = '"%s"' % cmd try: for k, v in environ.iteritems(): os.environ[k] = py2shell(v) os.environ['HG'] = hgexecutable() if cwd is not None and oldcwd != cwd: os.chdir(cwd) rc = os.system(cmd) if sys.platform == 'OpenVMS' and rc & 1: rc = 0 if rc and onerr: errmsg = '%s %s' % (os.path.basename(origcmd.split(None, 1)[0]), explain_exit(rc)[0]) if errprefix: errmsg = '%s: %s' % (errprefix, errmsg) try: onerr.warn(errmsg + '\n') except AttributeError: raise onerr(errmsg) return rc finally: for k, v in oldenv.iteritems(): if v is None: del os.environ[k] else: os.environ[k] = v if cwd is not None and oldcwd != cwd: os.chdir(oldcwd) class SignatureError(Exception): pass def checksignature(func): '''wrap a function with code to check for calling errors''' def check(*args, **kwargs): try: return func(*args, **kwargs) except TypeError: if len(traceback.extract_tb(sys.exc_info()[2])) == 1: raise SignatureError raise return check # os.path.lexists is not available on python2.3 def lexists(filename): "test whether a file with this name exists. does not follow symlinks" try: os.lstat(filename) except: return False return True def rename(src, dst): """forcibly rename a file""" try: os.rename(src, dst) except OSError, err: # FIXME: check err (EEXIST ?) # on windows, rename to existing file is not allowed, so we # must delete destination first. but if file is open, unlink # schedules it for delete but does not delete it. rename # happens immediately even for open files, so we create # temporary file, delete it, rename destination to that name, # then delete that. then rename is safe to do. fd, temp = tempfile.mkstemp(dir=os.path.dirname(dst) or '.') os.close(fd) os.unlink(temp) os.rename(dst, temp) os.unlink(temp) os.rename(src, dst) def unlink(f): """unlink and remove the directory if it is empty""" os.unlink(f) # try removing directories that might now be empty try: os.removedirs(os.path.dirname(f)) except OSError: pass def copyfile(src, dest): "copy a file, preserving mode" if os.path.islink(src): try: os.unlink(dest) except: pass os.symlink(os.readlink(src), dest) else: try: shutil.copyfile(src, dest) shutil.copymode(src, dest) except shutil.Error, inst: raise Abort(str(inst)) def copyfiles(src, dst, hardlink=None): """Copy a directory tree using hardlinks if possible""" if hardlink is None: hardlink = (os.stat(src).st_dev == os.stat(os.path.dirname(dst)).st_dev) if os.path.isdir(src): os.mkdir(dst) for name, kind in osutil.listdir(src): srcname = os.path.join(src, name) dstname = os.path.join(dst, name) copyfiles(srcname, dstname, hardlink) else: if hardlink: try: os_link(src, dst) except (IOError, OSError): hardlink = False shutil.copy(src, dst) else: shutil.copy(src, dst) class path_auditor(object): '''ensure that a filesystem path contains no banned components. the following properties of a path are checked: - under top-level .hg - starts at the root of a windows drive - contains ".." - traverses a symlink (e.g. a/symlink_here/b) - inside a nested repository''' def __init__(self, root): self.audited = set() self.auditeddir = set() self.root = root def __call__(self, path): if path in self.audited: return normpath = os.path.normcase(path) parts = splitpath(normpath) if (os.path.splitdrive(path)[0] or parts[0] in ('.hg', '.hg.', '') or os.pardir in parts): raise Abort(_("path contains illegal component: %s") % path) if '.hg' in path: for p in '.hg', '.hg.': if p in parts[1:-1]: pos = parts.index(p) base = os.path.join(*parts[:pos]) raise Abort(_('path %r is inside repo %r') % (path, base)) def check(prefix): curpath = os.path.join(self.root, prefix) try: st = os.lstat(curpath) except OSError, err: # EINVAL can be raised as invalid path syntax under win32. # They must be ignored for patterns can be checked too. if err.errno not in (errno.ENOENT, errno.ENOTDIR, errno.EINVAL): raise else: if stat.S_ISLNK(st.st_mode): raise Abort(_('path %r traverses symbolic link %r') % (path, prefix)) elif (stat.S_ISDIR(st.st_mode) and os.path.isdir(os.path.join(curpath, '.hg'))): raise Abort(_('path %r is inside repo %r') % (path, prefix)) parts.pop() prefixes = [] for n in range(len(parts)): prefix = os.sep.join(parts) if prefix in self.auditeddir: break check(prefix) prefixes.append(prefix) parts.pop() self.audited.add(path) # only add prefixes to the cache after checking everything: we don't # want to add "foo/bar/baz" before checking if there's a "foo/.hg" self.auditeddir.update(prefixes) def _makelock_file(info, pathname): ld = os.open(pathname, os.O_CREAT | os.O_WRONLY | os.O_EXCL) os.write(ld, info) os.close(ld) def _readlock_file(pathname): return posixfile(pathname).read() def nlinks(pathname): """Return number of hardlinks for the given file.""" return os.lstat(pathname).st_nlink if hasattr(os, 'link'): os_link = os.link else: def os_link(src, dst): raise OSError(0, _("Hardlinks not supported")) def fstat(fp): '''stat file object that may not have fileno method.''' try: return os.fstat(fp.fileno()) except AttributeError: return os.stat(fp.name) posixfile = file def openhardlinks(): '''return true if it is safe to hold open file handles to hardlinks''' return True def _statfiles(files): 'Stat each file in files and yield stat or None if file does not exist.' lstat = os.lstat for nf in files: try: st = lstat(nf) except OSError, err: if err.errno not in (errno.ENOENT, errno.ENOTDIR): raise st = None yield st def _statfiles_clustered(files): '''Stat each file in files and yield stat or None if file does not exist. Cluster and cache stat per directory to minimize number of OS stat calls.''' lstat = os.lstat ncase = os.path.normcase sep = os.sep dircache = {} # dirname -> filename -> status | None if file does not exist for nf in files: nf = ncase(nf) pos = nf.rfind(sep) if pos == -1: dir, base = '.', nf else: dir, base = nf[:pos+1], nf[pos+1:] cache = dircache.get(dir, None) if cache is None: try: dmap = dict([(ncase(n), s) for n, k, s in osutil.listdir(dir, True)]) except OSError, err: # handle directory not found in Python version prior to 2.5 # Python <= 2.4 returns native Windows code 3 in errno # Python >= 2.5 returns ENOENT and adds winerror field # EINVAL is raised if dir is not a directory. if err.errno not in (3, errno.ENOENT, errno.EINVAL, errno.ENOTDIR): raise dmap = {} cache = dircache.setdefault(dir, dmap) yield cache.get(base, None) if sys.platform == 'win32': statfiles = _statfiles_clustered else: statfiles = _statfiles getuser_fallback = None def getuser(): '''return name of current user''' try: return getpass.getuser() except ImportError: # import of pwd will fail on windows - try fallback if getuser_fallback: return getuser_fallback() # raised if win32api not available raise Abort(_('user name not available - set USERNAME ' 'environment variable')) def username(uid=None): """Return the name of the user with the given uid. If uid is None, return the name of the current user.""" try: import pwd if uid is None: uid = os.getuid() try: return pwd.getpwuid(uid)[0] except KeyError: return str(uid) except ImportError: return None def groupname(gid=None): """Return the name of the group with the given gid. If gid is None, return the name of the current group.""" try: import grp if gid is None: gid = os.getgid() try: return grp.getgrgid(gid)[0] except KeyError: return str(gid) except ImportError: return None # File system features def checkcase(path): """ Check whether the given path is on a case-sensitive filesystem Requires a path (like /foo/.hg) ending with a foldable final directory component. """ s1 = os.stat(path) d, b = os.path.split(path) p2 = os.path.join(d, b.upper()) if path == p2: p2 = os.path.join(d, b.lower()) try: s2 = os.stat(p2) if s2 == s1: return False return True except: return True _fspathcache = {} def fspath(name, root): '''Get name in the case stored in the filesystem The name is either relative to root, or it is an absolute path starting with root. Note that this function is unnecessary, and should not be called, for case-sensitive filesystems (simply because it's expensive). ''' # If name is absolute, make it relative if name.lower().startswith(root.lower()): l = len(root) if name[l] == os.sep or name[l] == os.altsep: l = l + 1 name = name[l:] if not os.path.exists(os.path.join(root, name)): return None seps = os.sep if os.altsep: seps = seps + os.altsep # Protect backslashes. This gets silly very quickly. seps.replace('\\','\\\\') pattern = re.compile(r'([^%s]+)|([%s]+)' % (seps, seps)) dir = os.path.normcase(os.path.normpath(root)) result = [] for part, sep in pattern.findall(name): if sep: result.append(sep) continue if dir not in _fspathcache: _fspathcache[dir] = os.listdir(dir) contents = _fspathcache[dir] lpart = part.lower() for n in contents: if n.lower() == lpart: result.append(n) break else: # Cannot happen, as the file exists! result.append(part) dir = os.path.join(dir, lpart) return ''.join(result) def checkexec(path): """ Check whether the given path is on a filesystem with UNIX-like exec flags Requires a directory (like /foo/.hg) """ # VFAT on some Linux versions can flip mode but it doesn't persist # a FS remount. Frequently we can detect it if files are created # with exec bit on. try: EXECFLAGS = stat.S_IXUSR | stat.S_IXGRP | stat.S_IXOTH fh, fn = tempfile.mkstemp("", "", path) try: os.close(fh) m = os.stat(fn).st_mode & 0777 new_file_has_exec = m & EXECFLAGS os.chmod(fn, m ^ EXECFLAGS) exec_flags_cannot_flip = ((os.stat(fn).st_mode & 0777) == m) finally: os.unlink(fn) except (IOError, OSError): # we don't care, the user probably won't be able to commit anyway return False return not (new_file_has_exec or exec_flags_cannot_flip) def checklink(path): """check whether the given path is on a symlink-capable filesystem""" # mktemp is not racy because symlink creation will fail if the # file already exists name = tempfile.mktemp(dir=path) try: os.symlink(".", name) os.unlink(name) return True except (OSError, AttributeError): return False _umask = os.umask(0) os.umask(_umask) def needbinarypatch(): """return True if patches should be applied in binary mode by default.""" return os.name == 'nt' def endswithsep(path): '''Check path ends with os.sep or os.altsep.''' return path.endswith(os.sep) or os.altsep and path.endswith(os.altsep) def splitpath(path): '''Split path by os.sep. Note that this function does not use os.altsep because this is an alternative of simple "xxx.split(os.sep)". It is recommended to use os.path.normpath() before using this function if need.''' return path.split(os.sep) def gui(): '''Are we running in a GUI?''' return os.name == "nt" or os.name == "mac" or os.environ.get("DISPLAY") def lookup_reg(key, name=None, scope=None): return None # Platform specific variants if os.name == 'nt': import msvcrt nulldev = 'NUL:' class winstdout: '''stdout on windows misbehaves if sent through a pipe''' def __init__(self, fp): self.fp = fp def __getattr__(self, key): return getattr(self.fp, key) def close(self): try: self.fp.close() except: pass def write(self, s): try: # This is workaround for "Not enough space" error on # writing large size of data to console. limit = 16000 l = len(s) start = 0 while start < l: end = start + limit self.fp.write(s[start:end]) start = end except IOError, inst: if inst.errno != 0: raise self.close() raise IOError(errno.EPIPE, 'Broken pipe') def flush(self): try: return self.fp.flush() except IOError, inst: if inst.errno != errno.EINVAL: raise self.close() raise IOError(errno.EPIPE, 'Broken pipe') sys.stdout = winstdout(sys.stdout) def _is_win_9x(): '''return true if run on windows 95, 98 or me.''' try: return sys.getwindowsversion()[3] == 1 except AttributeError: return 'command' in os.environ.get('comspec', '') def openhardlinks(): return not _is_win_9x and "win32api" in locals() def system_rcpath(): try: return system_rcpath_win32() except: return [r'c:\mercurial\mercurial.ini'] def user_rcpath(): '''return os-specific hgrc search path to the user dir''' try: path = user_rcpath_win32() except: home = os.path.expanduser('~') path = [os.path.join(home, 'mercurial.ini'), os.path.join(home, '.hgrc')] userprofile = os.environ.get('USERPROFILE') if userprofile: path.append(os.path.join(userprofile, 'mercurial.ini')) path.append(os.path.join(userprofile, '.hgrc')) return path def parse_patch_output(output_line): """parses the output produced by patch and returns the file name""" pf = output_line[14:] if pf[0] == '`': pf = pf[1:-1] # Remove the quotes return pf def sshargs(sshcmd, host, user, port): '''Build argument list for ssh or Plink''' pflag = 'plink' in sshcmd.lower() and '-P' or '-p' args = user and ("%s@%s" % (user, host)) or host return port and ("%s %s %s" % (args, pflag, port)) or args def testpid(pid): '''return False if pid dead, True if running or not known''' return True def set_flags(f, l, x): pass def set_binary(fd): # When run without console, pipes may expose invalid # fileno(), usually set to -1. if hasattr(fd, 'fileno') and fd.fileno() >= 0: msvcrt.setmode(fd.fileno(), os.O_BINARY) def pconvert(path): return '/'.join(splitpath(path)) def localpath(path): return path.replace('/', '\\') def normpath(path): return pconvert(os.path.normpath(path)) makelock = _makelock_file readlock = _readlock_file def samestat(s1, s2): return False # A sequence of backslashes is special iff it precedes a double quote: # - if there's an even number of backslashes, the double quote is not # quoted (i.e. it ends the quoted region) # - if there's an odd number of backslashes, the double quote is quoted # - in both cases, every pair of backslashes is unquoted into a single # backslash # (See http://msdn2.microsoft.com/en-us/library/a1y7w461.aspx ) # So, to quote a string, we must surround it in double quotes, double # the number of backslashes that preceed double quotes and add another # backslash before every double quote (being careful with the double # quote we've appended to the end) _quotere = None def shellquote(s): global _quotere if _quotere is None: _quotere = re.compile(r'(\\*)("|\\$)') return '"%s"' % _quotere.sub(r'\1\1\\\2', s) def quotecommand(cmd): """Build a command string suitable for os.popen* calls.""" # The extra quotes are needed because popen* runs the command # through the current COMSPEC. cmd.exe suppress enclosing quotes. return '"' + cmd + '"' def popen(command, mode='r'): # Work around "popen spawned process may not write to stdout # under windows" # http://bugs.python.org/issue1366 command += " 2> %s" % nulldev return os.popen(quotecommand(command), mode) def explain_exit(code): return _("exited with status %d") % code, code # if you change this stub into a real check, please try to implement the # username and groupname functions above, too. def isowner(fp, st=None): return True def find_in_path(name, path, default=None): '''find name in search path. path can be string (will be split with os.pathsep), or iterable thing that returns strings. if name found, return path to name. else return default. name is looked up using cmd.exe rules, using PATHEXT.''' if isinstance(path, str): path = path.split(os.pathsep) pathext = os.environ.get('PATHEXT', '.COM;.EXE;.BAT;.CMD') pathext = pathext.lower().split(os.pathsep) isexec = os.path.splitext(name)[1].lower() in pathext for p in path: p_name = os.path.join(p, name) if isexec and os.path.exists(p_name): return p_name for ext in pathext: p_name_ext = p_name + ext if os.path.exists(p_name_ext): return p_name_ext return default def set_signal_handler(): try: set_signal_handler_win32() except NameError: pass try: # override functions with win32 versions if possible from util_win32 import * if not _is_win_9x(): posixfile = posixfile_nt except ImportError: pass else: nulldev = '/dev/null' def rcfiles(path): rcs = [os.path.join(path, 'hgrc')] rcdir = os.path.join(path, 'hgrc.d') try: rcs.extend([os.path.join(rcdir, f) for f, kind in osutil.listdir(rcdir) if f.endswith(".rc")]) except OSError: pass return rcs def system_rcpath(): path = [] # old mod_python does not set sys.argv if len(getattr(sys, 'argv', [])) > 0: path.extend(rcfiles(os.path.dirname(sys.argv[0]) + '/../etc/mercurial')) path.extend(rcfiles('/etc/mercurial')) return path def user_rcpath(): return [os.path.expanduser('~/.hgrc')] def parse_patch_output(output_line): """parses the output produced by patch and returns the file name""" pf = output_line[14:] if os.sys.platform == 'OpenVMS': if pf[0] == '`': pf = pf[1:-1] # Remove the quotes else: if pf.startswith("'") and pf.endswith("'") and " " in pf: pf = pf[1:-1] # Remove the quotes return pf def sshargs(sshcmd, host, user, port): '''Build argument list for ssh''' args = user and ("%s@%s" % (user, host)) or host return port and ("%s -p %s" % (args, port)) or args def is_exec(f): """check whether a file is executable""" return (os.lstat(f).st_mode & 0100 != 0) def set_flags(f, l, x): s = os.lstat(f).st_mode if l: if not stat.S_ISLNK(s): # switch file to link data = file(f).read() os.unlink(f) try: os.symlink(data, f) except: # failed to make a link, rewrite file file(f, "w").write(data) # no chmod needed at this point return if stat.S_ISLNK(s): # switch link to file data = os.readlink(f) os.unlink(f) file(f, "w").write(data) s = 0666 & ~_umask # avoid restatting for chmod sx = s & 0100 if x and not sx: # Turn on +x for every +r bit when making a file executable # and obey umask. os.chmod(f, s | (s & 0444) >> 2 & ~_umask) elif not x and sx: # Turn off all +x bits os.chmod(f, s & 0666) def set_binary(fd): pass def pconvert(path): return path def localpath(path): return path normpath = os.path.normpath samestat = os.path.samestat def makelock(info, pathname): try: os.symlink(info, pathname) except OSError, why: if why.errno == errno.EEXIST: raise else: _makelock_file(info, pathname) def readlock(pathname): try: return os.readlink(pathname) except OSError, why: if why.errno in (errno.EINVAL, errno.ENOSYS): return _readlock_file(pathname) else: raise def shellquote(s): if os.sys.platform == 'OpenVMS': return '"%s"' % s else: return "'%s'" % s.replace("'", "'\\''") def quotecommand(cmd): return cmd def popen(command, mode='r'): return os.popen(command, mode) def testpid(pid): '''return False if pid dead, True if running or not sure''' if os.sys.platform == 'OpenVMS': return True try: os.kill(pid, 0) return True except OSError, inst: return inst.errno != errno.ESRCH def explain_exit(code): """return a 2-tuple (desc, code) describing a process's status""" if os.WIFEXITED(code): val = os.WEXITSTATUS(code) return _("exited with status %d") % val, val elif os.WIFSIGNALED(code): val = os.WTERMSIG(code) return _("killed by signal %d") % val, val elif os.WIFSTOPPED(code): val = os.WSTOPSIG(code) return _("stopped by signal %d") % val, val raise ValueError(_("invalid exit code")) def isowner(fp, st=None): """Return True if the file object f belongs to the current user. The return value of a util.fstat(f) may be passed as the st argument. """ if st is None: st = fstat(fp) return st.st_uid == os.getuid() def find_in_path(name, path, default=None): '''find name in search path. path can be string (will be split with os.pathsep), or iterable thing that returns strings. if name found, return path to name. else return default.''' if isinstance(path, str): path = path.split(os.pathsep) for p in path: p_name = os.path.join(p, name) if os.path.exists(p_name): return p_name return default def set_signal_handler(): pass def find_exe(name, default=None): '''find path of an executable. if name contains a path component, return it as is. otherwise, use normal executable search path.''' if os.sep in name or sys.platform == 'OpenVMS': # don't check the executable bit. if the file isn't # executable, whoever tries to actually run it will give a # much more useful error message. return name return find_in_path(name, os.environ.get('PATH', ''), default=default) def mktempcopy(name, emptyok=False, createmode=None): """Create a temporary file with the same contents from name The permission bits are copied from the original file. If the temporary file is going to be truncated immediately, you can use emptyok=True as an optimization. Returns the name of the temporary file. """ d, fn = os.path.split(name) fd, temp = tempfile.mkstemp(prefix='.%s-' % fn, dir=d) os.close(fd) # Temporary files are created with mode 0600, which is usually not # what we want. If the original file already exists, just copy # its mode. Otherwise, manually obey umask. try: st_mode = os.lstat(name).st_mode & 0777 except OSError, inst: if inst.errno != errno.ENOENT: raise st_mode = createmode if st_mode is None: st_mode = ~_umask st_mode &= 0666 os.chmod(temp, st_mode) if emptyok: return temp try: try: ifp = posixfile(name, "rb") except IOError, inst: if inst.errno == errno.ENOENT: return temp if not getattr(inst, 'filename', None): inst.filename = name raise ofp = posixfile(temp, "wb") for chunk in filechunkiter(ifp): ofp.write(chunk) ifp.close() ofp.close() except: try: os.unlink(temp) except: pass raise return temp class atomictempfile(posixfile): """file-like object that atomically updates a file All writes will be redirected to a temporary copy of the original file. When rename is called, the copy is renamed to the original name, making the changes visible. """ def __init__(self, name, mode, createmode): self.__name = name self.temp = mktempcopy(name, emptyok=('w' in mode), createmode=createmode) posixfile.__init__(self, self.temp, mode) def rename(self): if not self.closed: posixfile.close(self) rename(self.temp, localpath(self.__name)) def __del__(self): if not self.closed: try: os.unlink(self.temp) except: pass posixfile.close(self) def makedirs(name, mode=None): """recursive directory creation with parent mode inheritance""" try: os.mkdir(name) if mode is not None: os.chmod(name, mode) return except OSError, err: if err.errno == errno.EEXIST: return if err.errno != errno.ENOENT: raise parent = os.path.abspath(os.path.dirname(name)) makedirs(parent, mode) makedirs(name, mode) class opener(object): """Open files relative to a base directory This class is used to hide the details of COW semantics and remote file access from higher level code. """ def __init__(self, base, audit=True): self.base = base if audit: self.audit_path = path_auditor(base) else: self.audit_path = always self.createmode = None def __getattr__(self, name): if name == '_can_symlink': self._can_symlink = checklink(self.base) return self._can_symlink raise AttributeError(name) def _fixfilemode(self, name): if self.createmode is None: return os.chmod(name, self.createmode & 0666) def __call__(self, path, mode="r", text=False, atomictemp=False): self.audit_path(path) f = os.path.join(self.base, path) if not text and "b" not in mode: mode += "b" # for that other OS nlink = -1 if mode not in ("r", "rb"): try: nlink = nlinks(f) except OSError: nlink = 0 d = os.path.dirname(f) if not os.path.isdir(d): makedirs(d, self.createmode) if atomictemp: return atomictempfile(f, mode, self.createmode) if nlink > 1: rename(mktempcopy(f), f) fp = posixfile(f, mode) if nlink == 0: self._fixfilemode(f) return fp def symlink(self, src, dst): self.audit_path(dst) linkname = os.path.join(self.base, dst) try: os.unlink(linkname) except OSError: pass dirname = os.path.dirname(linkname) if not os.path.exists(dirname): makedirs(dirname, self.createmode) if self._can_symlink: try: os.symlink(src, linkname) except OSError, err: raise OSError(err.errno, _('could not symlink to %r: %s') % (src, err.strerror), linkname) else: f = self(dst, "w") f.write(src) f.close() self._fixfilemode(dst) class chunkbuffer(object): """Allow arbitrary sized chunks of data to be efficiently read from an iterator over chunks of arbitrary size.""" def __init__(self, in_iter): """in_iter is the iterator that's iterating over the input chunks. targetsize is how big a buffer to try to maintain.""" self.iter = iter(in_iter) self.buf = '' self.targetsize = 2**16 def read(self, l): """Read L bytes of data from the iterator of chunks of data. Returns less than L bytes if the iterator runs dry.""" if l > len(self.buf) and self.iter: # Clamp to a multiple of self.targetsize targetsize = max(l, self.targetsize) collector = cStringIO.StringIO() collector.write(self.buf) collected = len(self.buf) for chunk in self.iter: collector.write(chunk) collected += len(chunk) if collected >= targetsize: break if collected < targetsize: self.iter = False self.buf = collector.getvalue() if len(self.buf) == l: s, self.buf = str(self.buf), '' else: s, self.buf = self.buf[:l], buffer(self.buf, l) return s def filechunkiter(f, size=65536, limit=None): """Create a generator that produces the data in the file size (default 65536) bytes at a time, up to optional limit (default is to read all data). Chunks may be less than size bytes if the chunk is the last chunk in the file, or the file is a socket or some other type of file that sometimes reads less data than is requested.""" assert size >= 0 assert limit is None or limit >= 0 while True: if limit is None: nbytes = size else: nbytes = min(limit, size) s = nbytes and f.read(nbytes) if not s: break if limit: limit -= len(s) yield s def makedate(): lt = time.localtime() if lt[8] == 1 and time.daylight: tz = time.altzone else: tz = time.timezone return time.mktime(lt), tz def datestr(date=None, format='%a %b %d %H:%M:%S %Y %1%2'): """represent a (unixtime, offset) tuple as a localized time. unixtime is seconds since the epoch, and offset is the time zone's number of seconds away from UTC. if timezone is false, do not append time zone to string.""" t, tz = date or makedate() if "%1" in format or "%2" in format: sign = (tz > 0) and "-" or "+" minutes = abs(tz) / 60 format = format.replace("%1", "%c%02d" % (sign, minutes / 60)) format = format.replace("%2", "%02d" % (minutes % 60)) s = time.strftime(format, time.gmtime(float(t) - tz)) return s def shortdate(date=None): """turn (timestamp, tzoff) tuple into iso 8631 date.""" return datestr(date, format='%Y-%m-%d') def strdate(string, format, defaults=[]): """parse a localized time string and return a (unixtime, offset) tuple. if the string cannot be parsed, ValueError is raised.""" def timezone(string): tz = string.split()[-1] if tz[0] in "+-" and len(tz) == 5 and tz[1:].isdigit(): sign = (tz[0] == "+") and 1 or -1 hours = int(tz[1:3]) minutes = int(tz[3:5]) return -sign * (hours * 60 + minutes) * 60 if tz == "GMT" or tz == "UTC": return 0 return None # NOTE: unixtime = localunixtime + offset offset, date = timezone(string), string if offset != None: date = " ".join(string.split()[:-1]) # add missing elements from defaults for part in defaults: found = [True for p in part if ("%"+p) in format] if not found: date += "@" + defaults[part] format += "@%" + part[0] timetuple = time.strptime(date, format) localunixtime = int(calendar.timegm(timetuple)) if offset is None: # local timezone unixtime = int(time.mktime(timetuple)) offset = unixtime - localunixtime else: unixtime = localunixtime + offset return unixtime, offset def parsedate(date, formats=None, defaults=None): """parse a localized date/time string and return a (unixtime, offset) tuple. The date may be a "unixtime offset" string or in one of the specified formats. If the date already is a (unixtime, offset) tuple, it is returned. """ if not date: return 0, 0 if isinstance(date, tuple) and len(date) == 2: return date if not formats: formats = defaultdateformats date = date.strip() try: when, offset = map(int, date.split(' ')) except ValueError: # fill out defaults if not defaults: defaults = {} now = makedate() for part in "d mb yY HI M S".split(): if part not in defaults: if part[0] in "HMS": defaults[part] = "00" else: defaults[part] = datestr(now, "%" + part[0]) for format in formats: try: when, offset = strdate(date, format, defaults) except (ValueError, OverflowError): pass else: break else: raise Abort(_('invalid date: %r ') % date) # validate explicit (probably user-specified) date and # time zone offset. values must fit in signed 32 bits for # current 32-bit linux runtimes. timezones go from UTC-12 # to UTC+14 if abs(when) > 0x7fffffff: raise Abort(_('date exceeds 32 bits: %d') % when) if offset < -50400 or offset > 43200: raise Abort(_('impossible time zone offset: %d') % offset) return when, offset def matchdate(date): """Return a function that matches a given date match specifier Formats include: '{date}' match a given date to the accuracy provided '<{date}' on or before a given date '>{date}' on or after a given date """ def lower(date): d = dict(mb="1", d="1") return parsedate(date, extendeddateformats, d)[0] def upper(date): d = dict(mb="12", HI="23", M="59", S="59") for days in "31 30 29".split(): try: d["d"] = days return parsedate(date, extendeddateformats, d)[0] except: pass d["d"] = "28" return parsedate(date, extendeddateformats, d)[0] if date[0] == "<": when = upper(date[1:]) return lambda x: x <= when elif date[0] == ">": when = lower(date[1:]) return lambda x: x >= when elif date[0] == "-": try: days = int(date[1:]) except ValueError: raise Abort(_("invalid day spec: %s") % date[1:]) when = makedate()[0] - days * 3600 * 24 return lambda x: x >= when elif " to " in date: a, b = date.split(" to ") start, stop = lower(a), upper(b) return lambda x: x >= start and x <= stop else: start, stop = lower(date), upper(date) return lambda x: x >= start and x <= stop def shortuser(user): """Return a short representation of a user name or email address.""" f = user.find('@') if f >= 0: user = user[:f] f = user.find('<') if f >= 0: user = user[f+1:] f = user.find(' ') if f >= 0: user = user[:f] f = user.find('.') if f >= 0: user = user[:f] return user def email(author): '''get email of author.''' r = author.find('>') if r == -1: r = None return author[author.find('<')+1:r] def ellipsis(text, maxlength=400): """Trim string to at most maxlength (default: 400) characters.""" if len(text) <= maxlength: return text else: return "%s..." % (text[:maxlength-3]) def walkrepos(path, followsym=False, seen_dirs=None, recurse=False): '''yield every hg repository under path, recursively.''' def errhandler(err): if err.filename == path: raise err if followsym and hasattr(os.path, 'samestat'): def _add_dir_if_not_there(dirlst, dirname): match = False samestat = os.path.samestat dirstat = os.stat(dirname) for lstdirstat in dirlst: if samestat(dirstat, lstdirstat): match = True break if not match: dirlst.append(dirstat) return not match else: followsym = False if (seen_dirs is None) and followsym: seen_dirs = [] _add_dir_if_not_there(seen_dirs, path) for root, dirs, files in os.walk(path, topdown=True, onerror=errhandler): if '.hg' in dirs: yield root # found a repository qroot = os.path.join(root, '.hg', 'patches') if os.path.isdir(os.path.join(qroot, '.hg')): yield qroot # we have a patch queue repo here if recurse: # avoid recursing inside the .hg directory dirs.remove('.hg') else: dirs[:] = [] # don't descend further elif followsym: newdirs = [] for d in dirs: fname = os.path.join(root, d) if _add_dir_if_not_there(seen_dirs, fname): if os.path.islink(fname): for hgname in walkrepos(fname, True, seen_dirs): yield hgname else: newdirs.append(d) dirs[:] = newdirs _rcpath = None def os_rcpath(): '''return default os-specific hgrc search path''' path = system_rcpath() path.extend(user_rcpath()) path = [os.path.normpath(f) for f in path] return path def rcpath(): '''return hgrc search path. if env var HGRCPATH is set, use it. for each item in path, if directory, use files ending in .rc, else use item. make HGRCPATH empty to only look in .hg/hgrc of current repo. if no HGRCPATH, use default os-specific path.''' global _rcpath if _rcpath is None: if 'HGRCPATH' in os.environ: _rcpath = [] for p in os.environ['HGRCPATH'].split(os.pathsep): if not p: continue if os.path.isdir(p): for f, kind in osutil.listdir(p): if f.endswith('.rc'): _rcpath.append(os.path.join(p, f)) else: _rcpath.append(p) else: _rcpath = os_rcpath() return _rcpath def bytecount(nbytes): '''return byte count formatted as readable string, with units''' units = ( (100, 1<<30, _('%.0f GB')), (10, 1<<30, _('%.1f GB')), (1, 1<<30, _('%.2f GB')), (100, 1<<20, _('%.0f MB')), (10, 1<<20, _('%.1f MB')), (1, 1<<20, _('%.2f MB')), (100, 1<<10, _('%.0f KB')), (10, 1<<10, _('%.1f KB')), (1, 1<<10, _('%.2f KB')), (1, 1, _('%.0f bytes')), ) for multiplier, divisor, format in units: if nbytes >= divisor * multiplier: return format % (nbytes / float(divisor)) return units[-1][2] % nbytes def drop_scheme(scheme, path): sc = scheme + ':' if path.startswith(sc): path = path[len(sc):] if path.startswith('//'): path = path[2:] return path def uirepr(s): # Avoid double backslash in Windows path repr() return repr(s).replace('\\\\', '\\') def termwidth(): if 'COLUMNS' in os.environ: try: return int(os.environ['COLUMNS']) except ValueError: pass try: import termios, array, fcntl for dev in (sys.stdout, sys.stdin): try: fd = dev.fileno() if not os.isatty(fd): continue arri = fcntl.ioctl(fd, termios.TIOCGWINSZ, '\0' * 8) return array.array('h', arri)[1] except ValueError: pass except ImportError: pass return 80