view contrib/vagrant/Vagrantfile @ 45095:8e04607023e5

procutil: ensure that procutil.std{out,err}.write() writes all bytes Python 3 offers different kind of streams and it’s not guaranteed for all of them that calling write() writes all bytes. When Python is started in unbuffered mode, sys.std{out,err}.buffer are instances of io.FileIO, whose write() can write less bytes for platform-specific reasons (e.g. Linux has a 0x7ffff000 bytes maximum and could write less if interrupted by a signal; when writing to Windows consoles, it’s limited to 32767 bytes to avoid the "not enough space" error). This can lead to silent loss of data, both when using sys.std{out,err}.buffer (which may in fact not be a buffered stream) and when using the text streams sys.std{out,err} (I’ve created a CPython bug report for that: https://bugs.python.org/issue41221). Python may fix the problem at some point. For now, we implement our own wrapper for procutil.std{out,err} that calls the raw stream’s write() method until all bytes have been written. We don’t use sys.std{out,err} for larger writes, so I think it’s not worth the effort to patch them.
author Manuel Jacob <me@manueljacob.de>
date Fri, 10 Jul 2020 12:27:58 +0200
parents 7fd16ddabaa0
children
line wrap: on
line source

# -*- mode: ruby -*-

Vagrant.configure('2') do |config|
  # Debian 10.1 x86_64 without configuration management software
  config.vm.box = "debian/buster64"
  config.vm.hostname = "tests"

  config.vm.define "tests" do |conf|
    conf.vm.provision :file, source: "run-tests.sh", destination:"run-tests.sh"
    conf.vm.provision :shell, path: "provision.sh"
    conf.vm.synced_folder "../..", "/hgshared"
  end
end