view mercurial/archival.py @ 45377:da3b7c80aa34

hgweb: handle None from templatedir() equally bad in webcommands.py The following paragraph is based just on my reading of the code; I have not tried to test it. Before my recent work on templates in frozen binaries, it seems both `hgwebdir_mod.py` and `webcommands.py` would pass in an empty list into `staticfile()` when running in a frozen binary. That would then result in a variable in that function (`path`) not getting bound before its first use. I then changed that without thinking in D8786 so we passed a `None` value into the function, which made it break in another way (trying to iterate over `None`). Then I tried to fix it up in D8810, but I only changed `hgwebdir_mod.py` for some reason, and it still doesn't actually work in frozen binaries (which seems fair, since was broken before my changes too). This patch just replicates the half-assed "fix" from D8810 in `webcommands.py`, so they look more similar so I can start refactoring them in the same way. Differential Revision: https://phab.mercurial-scm.org/D8933
author Martin von Zweigbergk <martinvonz@google.com>
date Mon, 03 Aug 2020 22:40:05 -0700
parents 3a6ec080b521
children 89a2afe31e82
line wrap: on
line source

# archival.py - revision archival for mercurial
#
# Copyright 2006 Vadim Gelfer <vadim.gelfer@gmail.com>
#
# This software may be used and distributed according to the terms of the
# GNU General Public License version 2 or any later version.

from __future__ import absolute_import

import gzip
import os
import struct
import tarfile
import time
import zipfile
import zlib

from .i18n import _
from .node import nullrev
from .pycompat import open

from . import (
    error,
    formatter,
    match as matchmod,
    pycompat,
    scmutil,
    util,
    vfs as vfsmod,
)

stringio = util.stringio

# from unzip source code:
_UNX_IFREG = 0x8000
_UNX_IFLNK = 0xA000


def tidyprefix(dest, kind, prefix):
    '''choose prefix to use for names in archive.  make sure prefix is
    safe for consumers.'''

    if prefix:
        prefix = util.normpath(prefix)
    else:
        if not isinstance(dest, bytes):
            raise ValueError(b'dest must be string if no prefix')
        prefix = os.path.basename(dest)
        lower = prefix.lower()
        for sfx in exts.get(kind, []):
            if lower.endswith(sfx):
                prefix = prefix[: -len(sfx)]
                break
    lpfx = os.path.normpath(util.localpath(prefix))
    prefix = util.pconvert(lpfx)
    if not prefix.endswith(b'/'):
        prefix += b'/'
    # Drop the leading '.' path component if present, so Windows can read the
    # zip files (issue4634)
    if prefix.startswith(b'./'):
        prefix = prefix[2:]
    if prefix.startswith(b'../') or os.path.isabs(lpfx) or b'/../' in prefix:
        raise error.Abort(_(b'archive prefix contains illegal components'))
    return prefix


exts = {
    b'tar': [b'.tar'],
    b'tbz2': [b'.tbz2', b'.tar.bz2'],
    b'tgz': [b'.tgz', b'.tar.gz'],
    b'zip': [b'.zip'],
    b'txz': [b'.txz', b'.tar.xz'],
}


def guesskind(dest):
    for kind, extensions in pycompat.iteritems(exts):
        if any(dest.endswith(ext) for ext in extensions):
            return kind
    return None


def _rootctx(repo):
    # repo[0] may be hidden
    for rev in repo:
        return repo[rev]
    return repo[nullrev]


# {tags} on ctx includes local tags and 'tip', with no current way to limit
# that to global tags.  Therefore, use {latesttag} as a substitute when
# the distance is 0, since that will be the list of global tags on ctx.
_defaultmetatemplate = br'''
repo: {root}
node: {ifcontains(rev, revset("wdir()"), "{p1node}{dirty}", "{node}")}
branch: {branch|utf8}
{ifeq(latesttagdistance, 0, join(latesttag % "tag: {tag}", "\n"),
      separate("\n",
               join(latesttag % "latesttag: {tag}", "\n"),
               "latesttagdistance: {latesttagdistance}",
               "changessincelatesttag: {changessincelatesttag}"))}
'''[
    1:
]  # drop leading '\n'


def buildmetadata(ctx):
    '''build content of .hg_archival.txt'''
    repo = ctx.repo()

    opts = {
        b'template': repo.ui.config(
            b'experimental', b'archivemetatemplate', _defaultmetatemplate
        )
    }

    out = util.stringio()

    fm = formatter.formatter(repo.ui, out, b'archive', opts)
    fm.startitem()
    fm.context(ctx=ctx)
    fm.data(root=_rootctx(repo).hex())

    if ctx.rev() is None:
        dirty = b''
        if ctx.dirty(missing=True):
            dirty = b'+'
        fm.data(dirty=dirty)
    fm.end()

    return out.getvalue()


class tarit(object):
    '''write archive to tar file or stream.  can write uncompressed,
    or compress with gzip or bzip2.'''

    if pycompat.ispy3:
        GzipFileWithTime = gzip.GzipFile  # camelcase-required
    else:

        class GzipFileWithTime(gzip.GzipFile):
            def __init__(self, *args, **kw):
                timestamp = None
                if 'mtime' in kw:
                    timestamp = kw.pop('mtime')
                if timestamp is None:
                    self.timestamp = time.time()
                else:
                    self.timestamp = timestamp
                gzip.GzipFile.__init__(self, *args, **kw)

            def _write_gzip_header(self):
                self.fileobj.write(b'\037\213')  # magic header
                self.fileobj.write(b'\010')  # compression method
                fname = self.name
                if fname and fname.endswith(b'.gz'):
                    fname = fname[:-3]
                flags = 0
                if fname:
                    flags = gzip.FNAME  # pytype: disable=module-attr
                self.fileobj.write(pycompat.bytechr(flags))
                gzip.write32u(  # pytype: disable=module-attr
                    self.fileobj, int(self.timestamp)
                )
                self.fileobj.write(b'\002')
                self.fileobj.write(b'\377')
                if fname:
                    self.fileobj.write(fname + b'\000')

    def __init__(self, dest, mtime, kind=b''):
        self.mtime = mtime
        self.fileobj = None

        def taropen(mode, name=b'', fileobj=None):
            if kind == b'gz':
                mode = mode[0:1]
                if not fileobj:
                    fileobj = open(name, mode + b'b')
                gzfileobj = self.GzipFileWithTime(
                    name,
                    pycompat.sysstr(mode + b'b'),
                    zlib.Z_BEST_COMPRESSION,
                    fileobj,
                    mtime=mtime,
                )
                self.fileobj = gzfileobj
                return tarfile.TarFile.taropen(  # pytype: disable=attribute-error
                    name, pycompat.sysstr(mode), gzfileobj
                )
            else:
                try:
                    return tarfile.open(
                        name, pycompat.sysstr(mode + kind), fileobj
                    )
                except tarfile.CompressionError as e:
                    raise error.Abort(pycompat.bytestr(e))

        if isinstance(dest, bytes):
            self.z = taropen(b'w:', name=dest)
        else:
            self.z = taropen(b'w|', fileobj=dest)

    def addfile(self, name, mode, islink, data):
        name = pycompat.fsdecode(name)
        i = tarfile.TarInfo(name)
        i.mtime = self.mtime
        i.size = len(data)
        if islink:
            i.type = tarfile.SYMTYPE
            i.mode = 0o777
            i.linkname = pycompat.fsdecode(data)
            data = None
            i.size = 0
        else:
            i.mode = mode
            data = stringio(data)
        self.z.addfile(i, data)

    def done(self):
        self.z.close()
        if self.fileobj:
            self.fileobj.close()


class zipit(object):
    '''write archive to zip file or stream.  can write uncompressed,
    or compressed with deflate.'''

    def __init__(self, dest, mtime, compress=True):
        if isinstance(dest, bytes):
            dest = pycompat.fsdecode(dest)
        self.z = zipfile.ZipFile(
            dest, 'w', compress and zipfile.ZIP_DEFLATED or zipfile.ZIP_STORED
        )

        # Python's zipfile module emits deprecation warnings if we try
        # to store files with a date before 1980.
        epoch = 315532800  # calendar.timegm((1980, 1, 1, 0, 0, 0, 1, 1, 0))
        if mtime < epoch:
            mtime = epoch

        self.mtime = mtime
        self.date_time = time.gmtime(mtime)[:6]

    def addfile(self, name, mode, islink, data):
        i = zipfile.ZipInfo(pycompat.fsdecode(name), self.date_time)
        i.compress_type = self.z.compression  # pytype: disable=attribute-error
        # unzip will not honor unix file modes unless file creator is
        # set to unix (id 3).
        i.create_system = 3
        ftype = _UNX_IFREG
        if islink:
            mode = 0o777
            ftype = _UNX_IFLNK
        i.external_attr = (mode | ftype) << 16
        # add "extended-timestamp" extra block, because zip archives
        # without this will be extracted with unexpected timestamp,
        # if TZ is not configured as GMT
        i.extra += struct.pack(
            b'<hhBl',
            0x5455,  # block type: "extended-timestamp"
            1 + 4,  # size of this block
            1,  # "modification time is present"
            int(self.mtime),
        )  # last modification (UTC)
        self.z.writestr(i, data)

    def done(self):
        self.z.close()


class fileit(object):
    '''write archive as files in directory.'''

    def __init__(self, name, mtime):
        self.basedir = name
        self.opener = vfsmod.vfs(self.basedir)
        self.mtime = mtime

    def addfile(self, name, mode, islink, data):
        if islink:
            self.opener.symlink(data, name)
            return
        f = self.opener(name, b"w", atomictemp=False)
        f.write(data)
        f.close()
        destfile = os.path.join(self.basedir, name)
        os.chmod(destfile, mode)
        if self.mtime is not None:
            os.utime(destfile, (self.mtime, self.mtime))

    def done(self):
        pass


archivers = {
    b'files': fileit,
    b'tar': tarit,
    b'tbz2': lambda name, mtime: tarit(name, mtime, b'bz2'),
    b'tgz': lambda name, mtime: tarit(name, mtime, b'gz'),
    b'txz': lambda name, mtime: tarit(name, mtime, b'xz'),
    b'uzip': lambda name, mtime: zipit(name, mtime, False),
    b'zip': zipit,
}


def archive(
    repo,
    dest,
    node,
    kind,
    decode=True,
    match=None,
    prefix=b'',
    mtime=None,
    subrepos=False,
):
    '''create archive of repo as it was at node.

    dest can be name of directory, name of archive file, or file
    object to write archive to.

    kind is type of archive to create.

    decode tells whether to put files through decode filters from
    hgrc.

    match is a matcher to filter names of files to write to archive.

    prefix is name of path to put before every archive member.

    mtime is the modified time, in seconds, or None to use the changeset time.

    subrepos tells whether to include subrepos.
    '''

    if kind == b'txz' and not pycompat.ispy3:
        raise error.Abort(_(b'xz compression is only available in Python 3'))

    if kind == b'files':
        if prefix:
            raise error.Abort(_(b'cannot give prefix when archiving to files'))
    else:
        prefix = tidyprefix(dest, kind, prefix)

    def write(name, mode, islink, getdata):
        data = getdata()
        if decode:
            data = repo.wwritedata(name, data)
        archiver.addfile(prefix + name, mode, islink, data)

    if kind not in archivers:
        raise error.Abort(_(b"unknown archive type '%s'") % kind)

    ctx = repo[node]
    archiver = archivers[kind](dest, mtime or ctx.date()[0])

    if not match:
        match = scmutil.matchall(repo)

    if repo.ui.configbool(b"ui", b"archivemeta"):
        name = b'.hg_archival.txt'
        if match(name):
            write(name, 0o644, False, lambda: buildmetadata(ctx))

    files = list(ctx.manifest().walk(match))
    total = len(files)
    if total:
        files.sort()
        scmutil.prefetchfiles(
            repo, [(ctx.rev(), scmutil.matchfiles(repo, files))]
        )
        progress = repo.ui.makeprogress(
            _(b'archiving'), unit=_(b'files'), total=total
        )
        progress.update(0)
        for f in files:
            ff = ctx.flags(f)
            write(f, b'x' in ff and 0o755 or 0o644, b'l' in ff, ctx[f].data)
            progress.increment(item=f)
        progress.complete()

    if subrepos:
        for subpath in sorted(ctx.substate):
            sub = ctx.workingsub(subpath)
            submatch = matchmod.subdirmatcher(subpath, match)
            subprefix = prefix + subpath + b'/'
            total += sub.archive(archiver, subprefix, submatch, decode)

    if total == 0:
        raise error.Abort(_(b'no files match the archive pattern'))

    archiver.done()
    return total