Borg Documentation

BorgBackup Basic Usage

More screencasts: installation, advanced usage

What is BorgBackup?

BorgBackup (short: Borg) is a deduplicating backup program. Optionally, it supports compression and authenticated encryption.

The main goal of Borg is to provide an efficient and secure way to backup data. The data deduplication technique used makes Borg suitable for daily backups since only changes are stored. The authenticated encryption technique makes it suitable for backups to not fully trusted targets.

See the installation manual or, if you have already downloaded Borg, docs/installation.rst to get started with Borg. There is also an offline documentation available, in multiple formats.

Main features

Space efficient storage

Deduplication based on content-defined chunking is used to reduce the number of bytes stored: each file is split into a number of variable length chunks and only chunks that have never been seen before are added to the repository.

A chunk is considered duplicate if its id_hash value is identical. A cryptographically strong hash or MAC function is used as id_hash, e.g. (hmac-)sha256.

To deduplicate, all the chunks in the same repository are considered, no matter whether they come from different machines, from previous backups, from the same backup or even from the same single file.

Compared to other deduplication approaches, this method does NOT depend on:

  • file/directory names staying the same: So you can move your stuff around without killing the deduplication, even between machines sharing a repo.

  • complete files or time stamps staying the same: If a big file changes a little, only a few new chunks need to be stored - this is great for VMs or raw disks.

  • The absolute position of a data chunk inside a file: Stuff may get shifted and will still be found by the deduplication algorithm.

Speed
  • performance-critical code (chunking, compression, encryption) is implemented in C/Cython

  • local caching of files/chunks index data

  • quick detection of unmodified files

Data encryption

All data can be protected using 256-bit AES encryption, data integrity and authenticity is verified using HMAC-SHA256. Data is encrypted clientside.

Obfuscation

Optionally, borg can actively obfuscate e.g. the size of files / chunks to make fingerprinting attacks more difficult.

Compression

All data can be optionally compressed:

  • lz4 (super fast, low compression)

  • zstd (wide range from high speed and low compression to high compression and lower speed)

  • zlib (medium speed and compression)

  • lzma (low speed, high compression)

Off-site backups

Borg can store data on any remote host accessible over SSH. If Borg is installed on the remote host, big performance gains can be achieved compared to using a network filesystem (sshfs, nfs, ...).

Backups mountable as filesystems

Backup archives are mountable as userspace filesystems for easy interactive backup examination and restores (e.g. by using a regular file manager).

Easy installation on multiple platforms

We offer single-file binaries that do not require installing anything - you can just run them on these platforms:

  • Linux

  • Mac OS X

  • FreeBSD

  • OpenBSD and NetBSD (no xattrs/ACLs support or binaries yet)

  • Cygwin (experimental, no binaries yet)

  • Linux Subsystem of Windows 10 (experimental)

Free and Open Source Software
  • security and functionality can be audited independently

  • licensed under the BSD (3-clause) license, see License for the complete license

Easy to use

Initialize a new backup repository (see borg init --help for encryption options):

$ borg init -e repokey /path/to/repo

Create a backup archive:

$ borg create /path/to/repo::Saturday1 ~/Documents

Now doing another backup, just to show off the great deduplication:

$ borg create -v --stats /path/to/repo::Saturday2 ~/Documents
-----------------------------------------------------------------------------
Archive name: Saturday2
Archive fingerprint: 622b7c53c...
Time (start): Sat, 2016-02-27 14:48:13
Time (end):   Sat, 2016-02-27 14:48:14
Duration: 0.88 seconds
Number of files: 163
-----------------------------------------------------------------------------
               Original size      Compressed size    Deduplicated size
This archive:        6.85 MB              6.85 MB             30.79 kB  <-- !
All archives:       13.69 MB             13.71 MB              6.88 MB

               Unique chunks         Total chunks
Chunk index:             167                  330
-----------------------------------------------------------------------------

For a graphical frontend refer to our complementary project BorgWeb.

Helping, Donations and Bounties, becoming a Patron

Your help is always welcome!

Spread the word, give feedback, help with documentation, testing or development.

You can also give monetary support to the project, see there for details:

https://www.borgbackup.org/support/fund.html

Compatibility notes

EXPECT THAT WE WILL BREAK COMPATIBILITY REPEATEDLY WHEN MAJOR RELEASE NUMBER CHANGES (like when going from 0.x.y to 1.0.0 or from 1.x.y to 2.0.0).

NOT RELEASED DEVELOPMENT VERSIONS HAVE UNKNOWN COMPATIBILITY PROPERTIES.

THIS IS SOFTWARE IN DEVELOPMENT, DECIDE YOURSELF WHETHER IT FITS YOUR NEEDS.

Security issues should be reported to the Security contact (or see docs/support.rst in the source distribution).

Documentation Build Status (master) Test Coverage Best Practices Score

Installation

There are different ways to install Borg:

  • Distribution Package - easy and fast if a package is available from your distribution.

  • Standalone Binary - easy and fast, we provide a ready-to-use binary file that comes bundled with all dependencies.

  • From Source, either:

    • Using pip - installing a source package with pip needs more installation steps and requires all dependencies with development headers and a compiler.

    • Using git - for developers and power users who want to have the latest code or use revision control (each release is tagged).

Distribution Package

Some distributions might offer a ready-to-use borgbackup package which can be installed with the package manager.

Important

Those packages may not be up to date with the latest Borg releases. Before submitting a bug report, check the package version and compare that to our latest release then review Important notes to see if the bug has been fixed. Report bugs to the package maintainer rather than directly to Borg if the package is out of date in the distribution.

Distribution

Source

Command

Alpine Linux

Alpine repository

apk add borgbackup

Arch Linux

[community]

pacman -S borg

Debian

Debian packages

apt install borgbackup

Gentoo

ebuild

emerge borgbackup

GNU Guix

GNU Guix

guix package --install borg

Fedora/RHEL

Fedora official repository

dnf install borgbackup

FreeBSD

FreeBSD ports

cd /usr/ports/archivers/py-borgbackup && make install clean

macOS

Homebrew

brew install borgbackup (official formula, no FUSE support)
or
brew install --cask macfuse (private Tap, FUSE support)
brew install borgbackup/tap/borgbackup-fuse

Mageia

cauldron

urpmi borgbackup

NetBSD

pkgsrc

pkg_add py-borgbackup

NixOS

.nix file

nix-env -i borgbackup

OpenBSD

OpenBSD ports

pkg_add borgbackup

OpenIndiana

OpenIndiana hipster repository

pkg install borg

openSUSE

openSUSE official repository

zypper in borgbackup

Raspbian

Raspbian testing

apt install borgbackup

Ubuntu

Ubuntu packages, Ubuntu PPA

apt install borgbackup

Please ask package maintainers to build a package or, if you can package / submit it yourself, please help us with that! See #105 on github to followup on packaging efforts.

Current status of package in the repositories

Packaging status

Standalone Binary

Note

Releases are signed with an OpenPGP key, see Security for more instructions.

Borg x86/x64 amd/intel compatible binaries (generated with pyinstaller) are available on the releases page for the following platforms:

  • Linux: glibc >= 2.28 (ok for most supported Linux releases). Older glibc releases are untested and may not work.

  • MacOS: 10.12 or newer (To avoid signing issues download the file via command line or remove the quarantine attribute after downloading: $ xattr -dr com.apple.quarantine borg-macosx64.tgz)

  • FreeBSD: 12.1 (unknown whether it works for older releases)

ARM binaries are built by Johann Bauer, see: https://borg.bauerj.eu/

To install such a binary, just drop it into a directory in your PATH, make borg readable and executable for its users and then you can run borg:

sudo cp borg-linux64 /usr/local/bin/borg
sudo chown root:root /usr/local/bin/borg
sudo chmod 755 /usr/local/bin/borg

Optionally you can create a symlink to have borgfs available, which is an alias for borg mount:

ln -s /usr/local/bin/borg /usr/local/bin/borgfs

Note that the binary uses /tmp to unpack Borg with all dependencies. It will fail if /tmp has not enough free space or is mounted with the noexec option. You can change the temporary directory by setting the TEMP environment variable before running Borg.

If a new version is released, you will have to manually download it and replace the old version using the same steps as shown above.

From Source

Note

Some older Linux systems (like RHEL/CentOS 5) and Python interpreter binaries compiled to be able to run on such systems (like Python installed via Anaconda) might miss functions required by Borg.

This issue will be detected early and Borg will abort with a fatal error.

Dependencies

To install Borg from a source package (including pip), you have to install the following dependencies first:

  • Python 3 >= 3.8.0, plus development headers.

  • OpenSSL >= 1.0.0, plus development headers.

  • libacl (which depends on libattr), both plus development headers.

  • We have bundled code of the following packages, but borg by default (see setup.py if you want to change that) prefers a shared library if it can be found on the system (lib + dev headers) at build time:

    • liblz4 >= 1.7.0 (r129)

    • libzstd >= 1.3.0

    • libxxhash >= 0.8.1 (0.8.0 might work also)

  • pkg-config (cli tool) and pkgconfig python package (borg uses these to discover header and library location - if it can't import pkgconfig and is not pointed to header/library locations via env vars [see setup.py], it will fall back to using the bundled code, see above). These must be present before invoking setup.py!

  • some other Python dependencies, pip will automatically install them for you.

  • optionally, if you wish to mount an archive as a FUSE filesystem, you need a FUSE implementation for Python:

    • Either pyfuse3 (preferably, newer and maintained) or llfuse (older, unmaintained now). See also the BORG_FUSE_IMPL env variable.

    • See setup.py about the version requirements.

If you have troubles finding the right package names, have a look at the distribution specific sections below or the Vagrantfile in the git repository, which contains installation scripts for a number of operating systems.

In the following, the steps needed to install the dependencies are listed for a selection of platforms. If your distribution is not covered by these instructions, try to use your package manager to install the dependencies. On FreeBSD, you may need to get a recent enough OpenSSL version from FreeBSD ports.

After you have installed the dependencies, you can proceed with steps outlined under Using pip.

Debian / Ubuntu

Install the dependencies with development headers:

sudo apt-get install python3 python3-dev python3-pip python3-virtualenv \
libacl1-dev libacl1 \
libssl-dev \
liblz4-dev libzstd-dev libxxhash-dev \
build-essential \
pkg-config python3-pkgconfig
sudo apt-get install libfuse-dev fuse    # needed for llfuse
sudo apt-get install libfuse3-dev fuse3  # needed for pyfuse3

In case you get complaints about permission denied on /etc/fuse.conf: on Ubuntu this means your user is not in the fuse group. Add yourself to that group, log out and log in again.

Fedora

Install the dependencies with development headers:

sudo dnf install python3 python3-devel python3-pip python3-virtualenv \
libacl-devel libacl \
openssl-devel \
lz4-devel libzstd-devel xxhash-devel \
pkgconf python3-pkgconfig
sudo dnf install gcc gcc-c++ redhat-rpm-config
sudo dnf install fuse-devel fuse         # needed for llfuse
sudo dnf install fuse3-devel fuse3       # needed for pyfuse3
openSUSE Tumbleweed / Leap

Install the dependencies automatically using zypper:

sudo zypper source-install --build-deps-only borgbackup

Alternatively, you can enumerate all build dependencies in the command line:

sudo zypper install python3 python3-devel \
libacl-devel openssl-devel \
libxxhash-devel \
python3-Cython python3-Sphinx python3-msgpack-python python3-pkgconfig pkgconf \
python3-pytest python3-setuptools python3-setuptools_scm \
python3-sphinx_rtd_theme gcc gcc-c++
sudo zypper install python3-llfuse  # llfuse
macOS

When installing via Homebrew, dependencies are installed automatically. To install dependencies manually:

brew install python3 openssl zstd lz4 xxhash
brew install pkg-config
pip3 install virtualenv pkgconfig

For FUSE support to mount the backup archives, you need at least version 3.0 of macFUSE, which is available via github, or Homebrew:

brew install --cask macfuse

When installing Borg via pip, be sure to install the llfuse extra, since macFUSE only supports FUSE API v2. Also, since Homebrew won't link the installed openssl formula, point pkg-config to the correct path:

PKG_CONFIG_PATH="/usr/local/opt/openssl@1.1/lib/pkgconfig" pip install borgbackup[llfuse]

Be aware that for all recent macOS releases you must authorize full disk access. It is no longer sufficient to run borg backups as root. If you have not yet granted full disk access, and you run Borg backup from cron, you will see messages such as:

/Users/you/Pictures/Photos Library.photoslibrary: scandir: [Errno 1] Operation not permitted:

To fix this problem, you should grant full disk access to cron, and to your Terminal application. More information can be found here.

FreeBSD

Listed below are packages you will need to install Borg, its dependencies, and commands to make FUSE work for using the mount command.

pkg install -y python3 pkgconf
pkg install openssl
pkg install liblz4 zstd xxhash
pkg install fusefs-libs  # needed for llfuse
pkg install -y git
python3 -m ensurepip # to install pip for Python3
To use the mount command:
echo 'fuse_load="YES"' >> /boot/loader.conf
echo 'vfs.usermount=1' >> /etc/sysctl.conf
kldload fuse
sysctl vfs.usermount=1
Windows 10's Linux Subsystem

Note

Running under Windows 10's Linux Subsystem is experimental and has not been tested much yet.

Just follow the Ubuntu Linux installation steps. You can omit the FUSE stuff, it won't work anyway.

Cygwin

Note

Running under Cygwin is experimental and has not been tested much yet.

Use the Cygwin installer to install the dependencies:

python38 python38-devel python38-pkgconfig
python38-setuptools python38-pip python38-wheel python38-virtualenv
libssl-devel libxxhash-devel liblz4-devel libzstd-devel
binutils gcc-g++ git make openssh
Using pip

Ensure to install the dependencies as described within From Source.

Virtualenv can be used to build and install Borg without affecting the system Python or requiring root access. Using a virtual environment is optional, but recommended except for the most simple use cases.

Note

If you install into a virtual environment, you need to activate it first (source borg-env/bin/activate), before running borg. Alternatively, symlink borg-env/bin/borg into some directory that is in your PATH so you can just run borg.

This will use pip to install the latest release from PyPi:

virtualenv --python=python3 borg-env
source borg-env/bin/activate

# might be required if your tools are outdated
pip install -U pip setuptools wheel

# pkgconfig MUST be available before borg is installed!
pip install pkgconfig

# install Borg + Python dependencies into virtualenv
pip install borgbackup
# or alternatively (if you want FUSE support):
pip install borgbackup[llfuse]  # to use llfuse
pip install borgbackup[pyfuse3]  # to use pyfuse3

To upgrade Borg to a new version later, run the following after activating your virtual environment:

pip install -U borgbackup  # or ... borgbackup[llfuse/pyfuse3]

When doing manual pip installation, man pages are not automatically installed. You can run these commands to install the man pages locally:

# get borg from github
git clone https://github.com/borgbackup/borg.git borg

# Install the files with proper permissions
install -D -m 0644 borg/docs/man/borg*.1* $HOME/.local/share/man/man1/borg.1

# Update the man page cache
mandb
Using git

Ensure to install the dependencies as described within From Source.

This uses latest, unreleased development code from git. While we try not to break master, there are no guarantees on anything.

# get borg from github
git clone https://github.com/borgbackup/borg.git

# create a virtual environment
virtualenv --python=$(which python3) borg-env
source borg-env/bin/activate   # always before using!

# install borg + dependencies into virtualenv
cd borg
pip install -r requirements.d/development.txt
pip install -r requirements.d/docs.txt  # optional, to build the docs

pip install -e .           # in-place editable mode
or
pip install -e .[pyfuse3]  # in-place editable mode, use pyfuse3
or
pip install -e .[llfuse]   # in-place editable mode, use llfuse

# optional: run all the tests, on all installed Python versions
# requires fakeroot, available through your package manager
fakeroot -u tox --skip-missing-interpreters

By default the system installation of python will be used. If you need to use a different version of Python you can install this using pyenv:

...
# create a virtual environment
pyenv install 3.8.0  # minimum, preferably use something more recent!
pyenv global 3.8.0
pyenv local 3.8.0
virtualenv --python=${pyenv which python} borg-env
source borg-env/bin/activate   # always before using!
...

Note

As a developer or power user, you always want to use a virtual environment.

Quick Start

This chapter will get you started with Borg and covers various use cases.

A step by step example

  1. Before a backup can be made a repository has to be initialized:

    $ borg init --encryption=repokey /path/to/repo
    
  2. Backup the ~/src and ~/Documents directories into an archive called Monday:

    $ borg create /path/to/repo::Monday ~/src ~/Documents
    
  3. The next day create a new archive called Tuesday:

    $ borg create --stats /path/to/repo::Tuesday ~/src ~/Documents
    

    This backup will be a lot quicker and a lot smaller since only new never before seen data is stored. The --stats option causes Borg to output statistics about the newly created archive such as the amount of unique data (not shared with other archives):

    ------------------------------------------------------------------------------
    Archive name: Tuesday
    Archive fingerprint: bd31004d58f51ea06ff735d2e5ac49376901b21d58035f8fb05dbf866566e3c2
    Time (start): Tue, 2016-02-16 18:15:11
    Time (end):   Tue, 2016-02-16 18:15:11
    
    Duration: 0.19 seconds
    Number of files: 127
    ------------------------------------------------------------------------------
                          Original size      Compressed size    Deduplicated size
    This archive:                4.16 MB              4.17 MB             26.78 kB
    All archives:                8.33 MB              8.34 MB              4.19 MB
    
                          Unique chunks         Total chunks
    Chunk index:                     132                  261
    ------------------------------------------------------------------------------
    
  4. List all archives in the repository:

    $ borg list /path/to/repo
    Monday                               Mon, 2016-02-15 19:14:44
    Tuesday                              Tue, 2016-02-16 19:15:11
    
  5. List the contents of the Monday archive:

    $ borg list /path/to/repo::Monday
    drwxr-xr-x user   group          0 Mon, 2016-02-15 18:22:30 home/user/Documents
    -rw-r--r-- user   group       7961 Mon, 2016-02-15 18:22:30 home/user/Documents/Important.doc
    ...
    
  6. Restore the Monday archive by extracting the files relative to the current directory:

    $ borg extract /path/to/repo::Monday
    
  7. Delete the Monday archive (please note that this does not free repo disk space):

    $ borg delete /path/to/repo::Monday
    
  8. Recover disk space by compacting the segment files in the repo:

    $ borg compact /path/to/repo
    

Note

Borg is quiet by default (it works on WARNING log level). You can use options like --progress or --list to get specific reports during command execution. You can also add the -v (or --verbose or --info) option to adjust the log level to INFO to get other informational messages.

Archives and repositories

A Borg archive is the result of a single backup (borg create). An archive stores a snapshot of the data of the files "inside" it. One can later extract or mount an archive to restore from a backup.

Repositories are filesystem directories acting as self-contained stores of archives. Repositories can be accessed locally via path or remotely via ssh. Under the hood, repositories contain data blocks and a manifest tracking which blocks are in each archive. If some data hasn't changed from one backup to another, Borg can simply reference an already uploaded data chunk (deduplication).

Important note about free space

Before you start creating backups, please make sure that there is always a good amount of free space on the filesystem that has your backup repository (and also on ~/.cache). A few GB should suffice for most hard-drive sized repositories. See also Indexes / Caches memory usage.

Borg doesn't use space reserved for root on repository disks (even when run as root), on file systems which do not support this mechanism (e.g. XFS) we recommend to reserve some space in Borg itself just to be safe by adjusting the additional_free_space setting (a good starting point is 2G):

borg config /path/to/repo additional_free_space 2G

If Borg runs out of disk space, it tries to free as much space as it can while aborting the current operation safely, which allows the user to free more space by deleting/pruning archives. This mechanism is not bullet-proof in some circumstances [1].

If you really run out of disk space, it can be hard or impossible to free space, because Borg needs free space to operate - even to delete backup archives.

You can use some monitoring process or just include the free space information in your backup log files (you check them regularly anyway, right?).

Also helpful:

  • create a big file as a "space reserve", that you can delete to free space

  • if you use LVM: use a LV + a filesystem that you can resize later and have some unallocated PEs you can add to the LV.

  • consider using quotas

  • use prune and compact regularly

Important note about permissions

To avoid permissions issues (in your borg repository or borg cache), always access the repository using the same user account.

If you want to backup files of other users or the operating system, running borg as root likely will be required (otherwise you'ld get Permission denied errors). If you only back up your own files, you neither need nor want to run borg as root, just run it as your normal user.

For a local repository just always use the same user to invoke borg.

For a remote repository: always use e.g. borg@remote_host. You can use this from different local users, the remote user running borg and accessing the repo will always be borg.

If you need to access a local repository from different users, you can use the same method by using ssh to borg@localhost.

Important note about files changing during the backup process

Borg does not do anything about the internal consistency of the data it backs up. It just reads and backs up each file in whatever state that file is when Borg gets to it. On an active system, this can lead to two kinds of inconsistency:

  • By the time Borg backs up a file, it might have changed since the backup process was initiated

  • A file could change while Borg is backing it up, making the file internally inconsistent

If you have a set of files and want to ensure that they are backed up in a specific or consistent state, you must take steps to prevent changes to those files during the backup process. There are a few common techniques to achieve this.

  • Avoid running any programs that might change the files.

  • Snapshot files, filesystems, container storage volumes, or logical volumes. LVM or ZFS might be useful here.

  • Dump databases or stop the database servers.

  • Shut down virtual machines before backing up their images.

  • Shut down containers before backing up their storage volumes.

For some systems Borg might work well enough without these precautions. If you are simply backing up the files on a system that isn't very active (e.g. in a typical home directory), Borg usually works well enough without further care for consistency. Log files and caches might not be in a perfect state, but this is rarely a problem.

For databases, virtual machines, and containers, there are specific techniques for backing them up that do not simply use Borg to backup the underlying filesystem. For databases, check your database documentation for techniques that will save the database state between transactions. For virtual machines, consider running the backup on the VM itself or mounting the filesystem while the VM is shut down. For Docker containers, perhaps docker's "save" command can help.

Automating backups

The following example script is meant to be run daily by the root user on different local machines. It backs up a machine's important files (but not the complete operating system) to a repository ~/backup/main on a remote server. Some files which aren't necessarily needed in this backup are excluded. See borg help patterns on how to add more exclude options.

After the backup this script also uses the borg prune subcommand to keep only a certain number of old archives and deletes the others.

Finally, it uses the borg compact subcommand to remove deleted objects from the segment files in the repository to preserve disk space.

Before running, make sure that the repository is initialized as documented in Remote repositories and that the script has the correct permissions to be executable by the root user, but not executable or readable by anyone else, i.e. root:root 0700.

You can use this script as a starting point and modify it where it's necessary to fit your setup.

Do not forget to test your created backups to make sure everything you need is being backed up and that the prune command is keeping and deleting the correct backups.

#!/bin/sh

# Setting this, so the repo does not need to be given on the commandline:
export BORG_REPO=ssh://username@example.com:2022/~/backup/main

# See the section "Passphrase notes" for more infos.
export BORG_PASSPHRASE='XYZl0ngandsecurepa_55_phrasea&&123'

# some helpers and error handling:
info() { printf "\n%s %s\n\n" "$( date )" "$*" >&2; }
trap 'echo $( date ) Backup interrupted >&2; exit 2' INT TERM

info "Starting backup"

# Backup the most important directories into an archive named after
# the machine this script is currently running on:

borg create                         \
    --verbose                       \
    --filter AME                    \
    --list                          \
    --stats                         \
    --show-rc                       \
    --compression lz4               \
    --exclude-caches                \
    --exclude 'home/*/.cache/*'     \
    --exclude 'var/tmp/*'           \
                                    \
    ::'{hostname}-{now}'            \
    /etc                            \
    /home                           \
    /root                           \
    /var

backup_exit=$?

info "Pruning repository"

# Use the `prune` subcommand to maintain 7 daily, 4 weekly and 6 monthly
# archives of THIS machine. The '{hostname}-*' matching is very important to
# limit prune's operation to this machine's archives and not apply to
# other machines' archives also:

borg prune                          \
    --list                          \
    --glob-archives '{hostname}-*'  \
    --show-rc                       \
    --keep-daily    7               \
    --keep-weekly   4               \
    --keep-monthly  6

prune_exit=$?

# actually free repo disk space by compacting segments

info "Compacting repository"

borg compact

compact_exit=$?

# use highest exit code as global exit code
global_exit=$(( backup_exit > prune_exit ? backup_exit : prune_exit ))
global_exit=$(( compact_exit > global_exit ? compact_exit : global_exit ))

if [ ${global_exit} -eq 0 ]; then
    info "Backup, Prune, and Compact finished successfully"
elif [ ${global_exit} -eq 1 ]; then
    info "Backup, Prune, and/or Compact finished with warnings"
else
    info "Backup, Prune, and/or Compact finished with errors"
fi

exit ${global_exit}

Pitfalls with shell variables and environment variables

This applies to all environment variables you want Borg to see, not just BORG_PASSPHRASE. The short explanation is: always export your variable, and use single quotes if you're unsure of the details of your shell's expansion behavior. E.g.:

export BORG_PASSPHRASE='complicated & long'

This is because export exposes variables to subprocesses, which Borg may be one of. More on export can be found in the "ENVIRONMENT" section of the bash(1) man page.

Beware of how sudo interacts with environment variables. For example, you may be surprised that the following export has no effect on your command:

export BORG_PASSPHRASE='complicated & long'
sudo ./yourborgwrapper.sh  # still prompts for password

For more information, refer to the sudo(8) man page and env_keep in the sudoers(5) man page.

Tip

To debug what your borg process is actually seeing, find its PID (ps aux|grep borg) and then look into /proc/<PID>/environ.

Passphrase notes

If you use encryption (or authentication), Borg will interactively ask you for a passphrase to encrypt/decrypt the keyfile / repokey.

A passphrase should be a single line of text, a trailing linefeed will be stripped.

For your own safety, you maybe want to avoid empty passphrases as well extremely long passphrase (much more than 256 bits of entropy).

Also avoid passphrases containing non-ASCII characters. Borg is technically able to process all unicode text, but you might get into trouble reproducing the same encoded utf-8 bytes or with keyboard layouts, so better just avoid non-ASCII stuff.

If you want to automate, you can alternatively supply the passphrase directly or indirectly using some environment variables.

You can directly give a passphrase:

# use this passphrase (use safe permissions on the script!):
export BORG_PASSPHRASE='my super secret passphrase'

Or ask an external program to supply the passphrase:

# use the "pass" password manager to get the passphrase:
export BORG_PASSCOMMAND='pass show backup'

# use GPG to get the passphrase contained in a gpg-encrypted file:
export BORG_PASSCOMMAND='gpg --decrypt borg-passphrase.gpg'

Or read the passphrase from an open file descriptor:

export BORG_PASSPHRASE_FD=42

Using hardware crypto devices (like Nitrokey, Yubikey and others) is not directly supported by borg, but you can use these indirectly. E.g. if your crypto device supports GPG and borg calls gpg via BORG_PASSCOMMAND, it should just work.

Backup compression

The default is lz4 (very fast, but low compression ratio), but other methods are supported for different situations.

You can use zstd for a wide range from high speed (and relatively low compression) using N=1 to high compression (and lower speed) using N=22.

zstd is a modern compression algorithm and might be preferable over zlib and lzma, except if you need compatibility to older borg versions (< 1.1.4) that did not yet offer zstd.:

$ borg create --compression zstd,N /path/to/repo::arch ~

Other options are:

If you have a fast repo storage and you want minimum CPU usage, no compression:

$ borg create --compression none /path/to/repo::arch ~

If you have a less fast repo storage and you want a bit more compression (N=0..9, 0 means no compression, 9 means high compression):

$ borg create --compression zlib,N /path/to/repo::arch ~

If you have a very slow repo storage and you want high compression (N=0..9, 0 means low compression, 9 means high compression):

$ borg create --compression lzma,N /path/to/repo::arch ~

You'll need to experiment a bit to find the best compression for your use case. Keep an eye on CPU load and throughput.

Repository encryption

You can choose the repository encryption mode at repository creation time:

$ borg init --encryption=MODE PATH

For a list of available encryption MODEs and their descriptions, please refer to borg init.

If you use encryption, all data is encrypted on the client before being written to the repository. This means that an attacker who manages to compromise the host containing an encrypted repository will not be able to access any of the data, even while the backup is being made.

Key material is stored in encrypted form and can be only decrypted by providing the correct passphrase.

For automated backups the passphrase can be specified using the BORG_PASSPHRASE environment variable.

Note

Be careful about how you set that environment, see this note about password environments for more information.

Warning

The repository data is totally inaccessible without the key and the key passphrase.

Make a backup copy of the key file (keyfile mode) or repo config file (repokey mode) and keep it at a safe place, so you still have the key in case it gets corrupted or lost. Also keep your passphrase at a safe place. You can make backups using borg key export subcommand.

If you want to print a backup of your key to paper use the --paper option of this command and print the result, or print this template if you need a version with QR-Code.

A backup inside of the backup that is encrypted with that key/passphrase won't help you with that, of course.

In case you lose your repository and the security information, but have an older copy of it to restore from, don't use that later for creating new backups – you would run into security issues (reuse of nonce counter values). It is better to initialize a new Borg repository. See also: My repository is corrupt, how can I restore from an older copy of it?

Remote repositories

Borg can initialize and access repositories on remote hosts if the host is accessible using SSH. This is fastest and easiest when Borg is installed on the remote host, in which case the following syntax is used:

$ borg init user@hostname:/path/to/repo

Note: please see the usage chapter for a full documentation of repo URLs.

Remote operations over SSH can be automated with SSH keys. You can restrict the use of the SSH keypair by prepending a forced command to the SSH public key in the remote server's authorized_keys file. This example will start Borg in server mode and limit it to a specific filesystem path:

command="borg serve --restrict-to-path /path/to/repo",restrict ssh-rsa AAAAB3[...]

If it is not possible to install Borg on the remote host, it is still possible to use the remote host to store a repository by mounting the remote filesystem, for example, using sshfs:

$ sshfs user@hostname:/path/to /path/to
$ borg init /path/to/repo
$ fusermount -u /path/to

You can also use other remote filesystems in a similar way. Just be careful, not all filesystems out there are really stable and working good enough to be acceptable for backup usage.

Restoring a backup

Please note that we are only describing the most basic commands and options here - please refer to the command reference to see more.

For restoring, you usually want to work on the same machine as the same user that was also used to create the backups of the wanted files. Doing it like that avoids quite some issues:

  • no confusion relating to paths

  • same mapping of user/group names to user/group IDs

  • no permission issues

  • you likely already have a working borg setup there,

    • maybe including a environment variable for the key passphrase (for encrypted repos),

    • maybe including a keyfile for the repo (not needed for repokey mode),

    • maybe including a ssh key for the repo server (not needed for locally mounted repos),

    • maybe including a valid borg cache for that repo (quicker than cache rebuild).

The user might be:

  • root (if full backups, backups including system stuff or multiple users' files were made)

  • some specific user using sudo to execute borg as root

  • some specific user (if backups of that user's files were made)

A borg backup repository can be either:

  • in a local directory (like e.g. a locally mounted USB disk)

  • on a remote backup server machine that is reachable via ssh (client/server)

If the repository is encrypted, you will also need the key and the passphrase (which is protecting the key).

The key can be located:

  • in the repository (repokey mode).

    Easy, this will usually "just work".

  • in the home directory of the user who did the backup (keyfile mode).

    This may cause a bit more effort:

    • if you have just lost that home directory and you first need to restore the borg key (e.g. from the separate backup you have made of it or from another user or machine accessing the same repository).

    • if you first must find out the correct machine / user / home directory (where the borg client was run to make the backups).

The passphrase for the key has been either:

  • entered interactively at backup time (not practical if backup is automated / unattended).

  • acquired via some environment variable driven mechanism in the backup script (look there for BORG_PASSPHRASE, BORG_PASSCOMMAND, etc. and just do it like that).

There are 2 ways to restore files from a borg backup repository:

  • borg mount - use this if:

    • you don't precisely know what files you want to restore

    • you don't know which archive contains the files (in the state) you want

    • you need to look into files / directories before deciding what you want

    • you need a relatively low volume of data restored

    • you don't care for restoring stuff that the FUSE mount is not implementing yet (like special fs flags, ACLs)

    • you have a client with good resources (RAM, CPU, temp. disk space)

    • you want to rather use some filemanager to restore (copy) files than borg extract shell commands

  • borg extract - use this if:

    • you precisely know what you want (repo, archive, path)

    • you need a high volume of files restored (best speed)

    • you want a as-complete-as-it-gets reproduction of file metadata (like special fs flags, ACLs)

    • you have a client with low resources (RAM, CPU, temp. disk space)

Example with borg mount:

# open a new, separate terminal (this terminal will be blocked until umount)

# now we find out the archive names we have in the repo:
borg list /mnt/backup/borg_repo

# mount one archive from a borg repo:
borg mount /mnt/backup/borg_repo::myserver-system-2019-08-11 /mnt/borg

# alternatively, mount all archives from a borg repo (slower):
borg mount /mnt/backup/borg_repo /mnt/borg

# it may take a while until you will see stuff in /mnt/borg.

# now use another terminal or file browser and look into /mnt/borg.
# when finished, umount to unlock the repo and unblock the terminal:
borg umount /mnt/borg

Example with borg extract:

# borg extract always extracts into current directory and that directory
# should be empty (borg does not support transforming a non-empty dir to
# the state as present in your backup archive).
mkdir borg_restore
cd borg_restore

# now we find out the archive names we have in the repo:
borg list /mnt/backup/borg_repo

# we could find out the archive contents, esp. the path layout:
borg list /mnt/backup/borg_repo::myserver-system-2019-08-11

# we extract only some specific path (note: no leading / !):
borg extract /mnt/backup/borg_repo::myserver-system-2019-08-11 path/to/extract

# alternatively, we could fully extract the archive:
borg extract /mnt/backup/borg_repo::myserver-system-2019-08-11

# now move the files to the correct place...

Difference when using a remote borg backup server:

It is basically all the same as with the local repository, but you need to refer to the repo using a ssh:// URL.

In the given example, borg is the user name used to log into the machine backup.example.org which runs ssh on port 2222 and has the borg repo in /path/to/repo.

Instead of giving a FQDN or a hostname, you can also give an IP address.

As usual, you either need a password to log in or the backup server might have authentication set up via ssh authorized_keys (which is likely the case if unattended, automated backups were done).

borg mount ssh://borg@backup.example.org:2222/path/to/repo /mnt/borg
# or
borg extract ssh://borg@backup.example.org:2222/path/to/repo

Usage

General

Borg consists of a number of commands. Each command accepts a number of arguments and options and interprets various environment variables. The following sections will describe each command in detail.

Commands, options, parameters, paths and such are set in fixed-width. Option values are underlined. Borg has few options accepting a fixed set of values (e.g. --encryption of borg init).

Experimental features are marked with red stripes on the sides, like this paragraph.

Experimental features are not stable, which means that they may be changed in incompatible ways or even removed entirely without prior notice in following releases.

Positional Arguments and Options: Order matters

Borg only supports taking options (-s and --progress in the example) to the left or right of all positional arguments (repo::archive and path in the example), but not in between them:

borg create -s --progress repo::archive path  # good and preferred
borg create repo::archive path -s --progress  # also works
borg create -s repo::archive path --progress  # works, but ugly
borg create repo::archive -s --progress path  # BAD

This is due to a problem in the argparse module: https://bugs.python.org/issue15112

Repository URLs

Local filesystem (or locally mounted network filesystem):

/path/to/repo - filesystem path to repo directory, absolute path

path/to/repo - filesystem path to repo directory, relative path

Also, stuff like ~/path/to/repo or ~other/path/to/repo works (this is expanded by your shell).

Note: you may also prepend a file:// to a filesystem path to get URL style.

Remote repositories accessed via ssh user@host:

user@host:/path/to/repo - remote repo, absolute path

ssh://user@host:port/path/to/repo - same, alternative syntax, port can be given

Remote repositories with relative paths can be given using this syntax:

user@host:path/to/repo - path relative to current directory

user@host:~/path/to/repo - path relative to user's home directory

user@host:~other/path/to/repo - path relative to other's home directory

Note: giving user@host:/./path/to/repo or user@host:/~/path/to/repo or user@host:/~other/path/to/repo is also supported, but not required here.

Remote repositories with relative paths, alternative syntax with port:

ssh://user@host:port/./path/to/repo - path relative to current directory

ssh://user@host:port/~/path/to/repo - path relative to user's home directory

ssh://user@host:port/~other/path/to/repo - path relative to other's home directory

If you frequently need the same repo URL, it is a good idea to set the BORG_REPO environment variable to set a default for the repo URL:

export BORG_REPO='ssh://user@host:port/path/to/repo'

Then just leave away the repo URL if only a repo URL is needed and you want to use the default - it will be read from BORG_REPO then.

Use :: syntax to give the repo URL when syntax requires giving a positional argument for the repo (e.g. borg mount :: /mnt).

Repository / Archive Locations

Many commands want either a repository (just give the repo URL, see above) or an archive location, which is a repo URL followed by ::archive_name.

Archive names must not contain the / (slash) character. For simplicity, maybe also avoid blanks or other characters that have special meaning on the shell or in a filesystem (borg mount will use the archive name as directory name).

If you have set BORG_REPO (see above) and an archive location is needed, use ::archive_name - the repo URL part is then read from BORG_REPO.

Logging

Borg writes all log output to stderr by default. But please note that something showing up on stderr does not indicate an error condition just because it is on stderr. Please check the log levels of the messages and the return code of borg for determining error, warning or success conditions.

If you want to capture the log output to a file, just redirect it:

borg create repo::archive myfiles 2>> logfile

Custom logging configurations can be implemented via BORG_LOGGING_CONF.

The log level of the builtin logging configuration defaults to WARNING. This is because we want Borg to be mostly silent and only output warnings, errors and critical messages, unless output has been requested by supplying an option that implies output (e.g. --list or --progress).

Log levels: DEBUG < INFO < WARNING < ERROR < CRITICAL

Use --debug to set DEBUG log level - to get debug, info, warning, error and critical level output.

Use --info (or -v or --verbose) to set INFO log level - to get info, warning, error and critical level output.

Use --warning (default) to set WARNING log level - to get warning, error and critical level output.

Use --error to set ERROR log level - to get error and critical level output.

Use --critical to set CRITICAL log level - to get critical level output.

While you can set misc. log levels, do not expect that every command will give different output on different log levels - it's just a possibility.

Warning

Options --critical and --error are provided for completeness, their usage is not recommended as you might miss important information.

Return codes

Borg can exit with the following return codes (rc):

Return code

Meaning

0

success (logged as INFO)

1

warning (operation reached its normal end, but there were warnings -- you should check the log, logged as WARNING)

2

error (like a fatal error, a local or remote exception, the operation did not reach its normal end, logged as ERROR)

128+N

killed by signal N (e.g. 137 == kill -9)

If you use --show-rc, the return code is also logged at the indicated level as the last log entry.

Environment Variables

Borg uses some environment variables for automation:

General:
BORG_REPO

When set, use the value to give the default repository location. If a command needs an archive parameter, you can abbreviate as ::archive. If a command needs a repository parameter, you can either leave it away or abbreviate as ::, if a positional parameter is required.

BORG_PASSPHRASE

When set, use the value to answer the passphrase question for encrypted repositories. It is used when a passphrase is needed to access an encrypted repo as well as when a new passphrase should be initially set when initializing an encrypted repo. See also BORG_NEW_PASSPHRASE.

BORG_PASSCOMMAND

When set, use the standard output of the command (trailing newlines are stripped) to answer the passphrase question for encrypted repositories. It is used when a passphrase is needed to access an encrypted repo as well as when a new passphrase should be initially set when initializing an encrypted repo. Note that the command is executed without a shell. So variables, like $HOME will work, but ~ won't. If BORG_PASSPHRASE is also set, it takes precedence. See also BORG_NEW_PASSPHRASE.

BORG_PASSPHRASE_FD

When set, specifies a file descriptor to read a passphrase from. Programs starting borg may choose to open an anonymous pipe and use it to pass a passphrase. This is safer than passing via BORG_PASSPHRASE, because on some systems (e.g. Linux) environment can be examined by other processes. If BORG_PASSPHRASE or BORG_PASSCOMMAND are also set, they take precedence.

BORG_NEW_PASSPHRASE

When set, use the value to answer the passphrase question when a new passphrase is asked for. This variable is checked first. If it is not set, BORG_PASSPHRASE and BORG_PASSCOMMAND will also be checked. Main usecase for this is to fully automate borg change-passphrase.

BORG_DISPLAY_PASSPHRASE

When set, use the value to answer the "display the passphrase for verification" question when defining a new passphrase for encrypted repositories.

BORG_HOST_ID

Borg usually computes a host id from the FQDN plus the results of uuid.getnode() (which usually returns a unique id based on the MAC address of the network interface. Except if that MAC happens to be all-zero - in that case it returns a random value, which is not what we want (because it kills automatic stale lock removal). So, if you have a all-zero MAC address or other reasons to better externally control the host id, just set this environment variable to a unique value. If all your FQDNs are unique, you can just use the FQDN. If not, use fqdn@uniqueid.

BORG_LOGGING_CONF

When set, use the given filename as INI-style logging configuration. A basic example conf can be found at docs/misc/logging.conf.

BORG_RSH

When set, use this command instead of ssh. This can be used to specify ssh options, such as a custom identity file ssh -i /path/to/private/key. See man ssh for other options. Using the --rsh CMD commandline option overrides the environment variable.

BORG_REMOTE_PATH

When set, use the given path as borg executable on the remote (defaults to "borg" if unset). Using --remote-path PATH commandline option overrides the environment variable.

BORG_FILES_CACHE_SUFFIX

When set to a value at least one character long, instructs borg to use a specifically named (based on the suffix) alternative files cache. This can be used to avoid loading and saving cache entries for backup sources other than the current sources.

BORG_FILES_CACHE_TTL

When set to a numeric value, this determines the maximum "time to live" for the files cache entries (default: 20). The files cache is used to quickly determine whether a file is unchanged. The FAQ explains this more detailed in: It always chunks all my files, even unchanged ones!

BORG_SHOW_SYSINFO

When set to no (default: yes), system information (like OS, Python version, ...) in exceptions is not shown. Please only use for good reasons as it makes issues harder to analyze.

BORG_FUSE_IMPL

Choose the lowlevel FUSE implementation borg shall use for borg mount. This is a comma-separated list of implementation names, they are tried in the given order, e.g.:

  • pyfuse3,llfuse: default, first try to load pyfuse3, then try to load llfuse.

  • llfuse,pyfuse3: first try to load llfuse, then try to load pyfuse3.

  • pyfuse3: only try to load pyfuse3

  • llfuse: only try to load llfuse

  • none: do not try to load an implementation

BORG_SELFTEST

This can be used to influence borg's builtin self-tests. The default is to execute the tests at the beginning of each borg command invocation.

BORG_SELFTEST=disabled can be used to switch off the tests and rather save some time. Disabling is not recommended for normal borg users, but large scale borg storage providers can use this to optimize production servers after at least doing a one-time test borg (with selftests not disabled) when installing or upgrading machines / OS / borg.

BORG_WORKAROUNDS

A list of comma separated strings that trigger workarounds in borg, e.g. to work around bugs in other software.

Currently known strings are:

basesyncfile

Use the more simple BaseSyncFile code to avoid issues with sync_file_range. You might need this to run borg on WSL (Windows Subsystem for Linux) or in systemd.nspawn containers on some architectures (e.g. ARM). Using this does not affect data safety, but might result in a more bursty write to disk behaviour (not continuously streaming to disk).

retry_erofs

Retry opening a file without O_NOATIME if opening a file with O_NOATIME caused EROFS. You will need this to make archives from volume shadow copies in WSL1 (Windows Subsystem for Linux 1).

authenticated_no_key

Work around a lost passphrase or key for an authenticated mode repository (these are only authenticated, but not encrypted). If the key is missing in the repository config, add key = anything there.

This workaround is only for emergencies and only to extract data from an affected repository (read-only access):

BORG_WORKAROUNDS=authenticated_no_key borg extract repo::archive

After you have extracted all data you need, you MUST delete the repository:

BORG_WORKAROUNDS=authenticated_no_key borg delete repo

Now you can init a fresh repo. Make sure you do not use the workaround any more.

ignore_invalid_archive_tam

Work around invalid archive TAMs created by borg < 1.2.5, see #7791.

This workaround likely needs to get used only once when following the upgrade instructions for CVE-2023-36811, see Pre-1.2.5 archives spoofing vulnerability (CVE-2023-36811).

In normal production operations, this workaround should never be used.

Some automatic "answerers" (if set, they automatically answer confirmation questions):
BORG_UNKNOWN_UNENCRYPTED_REPO_ACCESS_IS_OK=no (or =yes)

For "Warning: Attempting to access a previously unknown unencrypted repository"

BORG_RELOCATED_REPO_ACCESS_IS_OK=no (or =yes)

For "Warning: The repository at location ... was previously located at ..."

BORG_CHECK_I_KNOW_WHAT_I_AM_DOING=NO (or =YES)

For "This is a potentially dangerous function..." (check --repair)

BORG_DELETE_I_KNOW_WHAT_I_AM_DOING=NO (or =YES)

For "You requested to completely DELETE the repository including all archives it contains:"

Note: answers are case sensitive. setting an invalid answer value might either give the default answer or ask you interactively, depending on whether retries are allowed (they by default are allowed). So please test your scripts interactively before making them a non-interactive script.

Directories and files:
BORG_BASE_DIR

Defaults to $HOME or ~$USER or ~ (in that order). If you want to move all borg-specific folders to a custom path at once, all you need to do is to modify BORG_BASE_DIR: the other paths for cache, config etc. will adapt accordingly (assuming you didn't set them to a different custom value).

BORG_CACHE_DIR

Defaults to $BORG_BASE_DIR/.cache/borg. If BORG_BASE_DIR is not explicitly set while XDG env var XDG_CACHE_HOME is set, then $XDG_CACHE_HOME/borg is being used instead. This directory contains the local cache and might need a lot of space for dealing with big repositories. Make sure you're aware of the associated security aspects of the cache location: Do I need to take security precautions regarding the cache?

BORG_CONFIG_DIR

Defaults to $BORG_BASE_DIR/.config/borg. If BORG_BASE_DIR is not explicitly set while XDG env var XDG_CONFIG_HOME is set, then $XDG_CONFIG_HOME/borg is being used instead. This directory contains all borg configuration directories, see the FAQ for a security advisory about the data in this directory: How important is the $HOME/.config/borg directory?

BORG_SECURITY_DIR

Defaults to $BORG_CONFIG_DIR/security. This directory contains information borg uses to track its usage of NONCES ("numbers used once" - usually in encryption context) and other security relevant data.

BORG_KEYS_DIR

Defaults to $BORG_CONFIG_DIR/keys. This directory contains keys for encrypted repositories.

BORG_KEY_FILE

When set, use the given path as repository key file. Please note that this is only for rather special applications that externally fully manage the key files:

  • this setting only applies to the keyfile modes (not to the repokey modes).

  • using a full, absolute path to the key file is recommended.

  • all directories in the given path must exist.

  • this setting forces borg to use the key file at the given location.

  • the key file must either exist (for most commands) or will be created (borg init).

  • you need to give a different path for different repositories.

  • you need to point to the correct key file matching the repository the command will operate on.

TMPDIR

This is where temporary files are stored (might need a lot of temporary space for some operations), see tempfile for details.

Building:
BORG_OPENSSL_PREFIX

Adds given OpenSSL header file directory to the default locations (setup.py).

BORG_LIBLZ4_PREFIX

Adds given prefix directory to the default locations. If a 'include/lz4.h' is found Borg will be linked against the system liblz4 instead of a bundled implementation. (setup.py)

BORG_LIBZSTD_PREFIX

Adds given prefix directory to the default locations. If a 'include/zstd.h' is found Borg will be linked against the system libzstd instead of a bundled implementation. (setup.py)

Please note:

  • Be very careful when using the "yes" sayers, the warnings with prompt exist for your / your data's security/safety.

  • Also be very careful when putting your passphrase into a script, make sure it has appropriate file permissions (e.g. mode 600, root:root).

File systems

We strongly recommend against using Borg (or any other database-like software) on non-journaling file systems like FAT, since it is not possible to assume any consistency in case of power failures (or a sudden disconnect of an external drive or similar failures).

While Borg uses a data store that is resilient against these failures when used on journaling file systems, it is not possible to guarantee this with some hardware -- independent of the software used. We don't know a list of affected hardware.

If you are suspicious whether your Borg repository is still consistent and readable after one of the failures mentioned above occurred, run borg check --verify-data to make sure it is consistent.

Requirements for Borg repository file systems

  • Long file names

  • At least three directory levels with short names

  • Typically, file sizes up to a few hundred MB. Large repositories may require large files (>2 GB).

  • Up to 1000 files per directory.

  • rename(2) / MoveFile(Ex) should work as specified, i.e. on the same file system it should be a move (not a copy) operation, and in case of a directory it should fail if the destination exists and is not an empty directory, since this is used for locking.

  • Hardlinks are needed for borg upgrade (if --inplace option is not used). Also hardlinks are used for more safe and secure file updating (e.g. of the repo config file), but the code tries to work also if hardlinks are not supported.

Units

To display quantities, Borg takes care of respecting the usual conventions of scale. Disk sizes are displayed in decimal, using powers of ten (so kB means 1000 bytes). For memory usage, binary prefixes are used, and are indicated using the IEC binary prefixes, using powers of two (so KiB means 1024 bytes).

Date and Time

We format date and time conforming to ISO-8601, that is: YYYY-MM-DD and HH:MM:SS (24h clock).

For more information about that, see: https://xkcd.com/1179/

Unless otherwise noted, we display local date and time. Internally, we store and process date and time as UTC.

Resource Usage

Borg might use a lot of resources depending on the size of the data set it is dealing with.

If one uses Borg in a client/server way (with a ssh: repository), the resource usage occurs in part on the client and in another part on the server.

If one uses Borg as a single process (with a filesystem repo), all the resource usage occurs in that one process, so just add up client + server to get the approximate resource usage.

CPU client:
  • borg create: does chunking, hashing, compression, crypto (high CPU usage)

  • chunks cache sync: quite heavy on CPU, doing lots of hashtable operations.

  • borg extract: crypto, decompression (medium to high CPU usage)

  • borg check: similar to extract, but depends on options given.

  • borg prune / borg delete archive: low to medium CPU usage

  • borg delete repo: done on the server

It won't go beyond 100% of 1 core as the code is currently single-threaded. Especially higher zlib and lzma compression levels use significant amounts of CPU cycles. Crypto might be cheap on the CPU (if hardware accelerated) or expensive (if not).

CPU server:

It usually doesn't need much CPU, it just deals with the key/value store (repository) and uses the repository index for that.

borg check: the repository check computes the checksums of all chunks (medium CPU usage) borg delete repo: low CPU usage

CPU (only for client/server operation):

When using borg in a client/server way with a ssh:-type repo, the ssh processes used for the transport layer will need some CPU on the client and on the server due to the crypto they are doing - esp. if you are pumping big amounts of data.

Memory (RAM) client:

The chunks index and the files index are read into memory for performance reasons. Might need big amounts of memory (see below). Compression, esp. lzma compression with high levels might need substantial amounts of memory.

Memory (RAM) server:

The server process will load the repository index into memory. Might need considerable amounts of memory, but less than on the client (see below).

Chunks index (client only):

Proportional to the amount of data chunks in your repo. Lots of chunks in your repo imply a big chunks index. It is possible to tweak the chunker params (see create options).

Files index (client only):

Proportional to the amount of files in your last backups. Can be switched off (see create options), but next backup might be much slower if you do. The speed benefit of using the files cache is proportional to file size.

Repository index (server only):

Proportional to the amount of data chunks in your repo. Lots of chunks in your repo imply a big repository index. It is possible to tweak the chunker params (see create options) to influence the amount of chunks being created.

Temporary files (client):

Reading data and metadata from a FUSE mounted repository will consume up to the size of all deduplicated, small chunks in the repository. Big chunks won't be locally cached.

Temporary files (server):

A non-trivial amount of data will be stored on the remote temp directory for each client that connects to it. For some remotes, this can fill the default temporary directory at /tmp. This can be remediated by ensuring the $TMPDIR, $TEMP, or $TMP environment variable is properly set for the sshd process. For some OSes, this can be done just by setting the correct value in the .bashrc (or equivalent login config file for other shells), however in other cases it may be necessary to first enable PermitUserEnvironment yes in your sshd_config file, then add environment="TMPDIR=/my/big/tmpdir" at the start of the public key to be used in the authorized_hosts file.

Cache files (client only):

Contains the chunks index and files index (plus a collection of single- archive chunk indexes which might need huge amounts of disk space, depending on archive count and size - see FAQ about how to reduce).

Network (only for client/server operation):

If your repository is remote, all deduplicated (and optionally compressed/ encrypted) data of course has to go over the connection (ssh:// repo url). If you use a locally mounted network filesystem, additionally some copy operations used for transaction support also go over the connection. If you backup multiple sources to one target repository, additional traffic happens for cache resynchronization.

Support for file metadata

Besides regular file and directory structures, Borg can preserve

  • symlinks (stored as symlink, the symlink is not followed)

  • special files:

    • character and block device files (restored via mknod)

    • FIFOs ("named pipes")

    • special file contents can be backed up in --read-special mode. By default the metadata to create them with mknod(2), mkfifo(2) etc. is stored.

  • hardlinked regular files, devices, FIFOs (considering all items in the same archive)

  • timestamps in nanosecond precision: mtime, atime, ctime

  • other timestamps: birthtime (on platforms supporting it)

  • permissions:

    • IDs of owning user and owning group

    • names of owning user and owning group (if the IDs can be resolved)

    • Unix Mode/Permissions (u/g/o permissions, suid, sgid, sticky)

On some platforms additional features are supported:

Platform

ACLs [5]

xattr [6]

Flags [7]

Linux

Yes

Yes

Yes [1]

macOS

Yes

Yes

Yes (all)

FreeBSD

Yes

Yes

Yes (all)

OpenBSD

n/a

n/a

Yes (all)

NetBSD

n/a

No [2]

Yes (all)

Solaris and derivatives

No [3]

No [3]

n/a

Windows (cygwin)

No [4]

No

No

Other Unix-like operating systems may work as well, but have not been tested at all.

Note that most of the platform-dependent features also depend on the file system. For example, ntfs-3g on Linux isn't able to convey NTFS ACLs.

In case you are interested in more details (like formulas), please see Internals. For details on the available JSON output, refer to All about JSON: How to develop frontends.

Common options

All Borg commands share these options:

-h, --help

show this help message and exit

--critical

work on log level CRITICAL

--error

work on log level ERROR

--warning

work on log level WARNING (default)

--info, -v, --verbose

work on log level INFO

--debug

enable debug output, work on log level DEBUG

--debug-topic TOPIC

enable TOPIC debugging (can be specified multiple times). The logger path is borg.debug.<TOPIC> if TOPIC is not fully qualified.

-p, --progress

show progress information

--iec

format using IEC units (1KiB = 1024B)

--log-json

Output one JSON object per log line instead of formatted text.

--lock-wait SECONDS

wait at most SECONDS for acquiring a repository/cache lock (default: 1).

--bypass-lock

Bypass locking mechanism

--show-version

show/log the borg version

--show-rc

show/log the return code (rc)

--umask M

set umask to M (local only, default: 0077)

--remote-path PATH

use PATH as borg executable on the remote (default: "borg")

--remote-ratelimit RATE

deprecated, use --upload-ratelimit instead

--upload-ratelimit RATE

set network upload rate limit in kiByte/s (default: 0=unlimited)

--remote-buffer UPLOAD_BUFFER

deprecated, use --upload-buffer instead

--upload-buffer UPLOAD_BUFFER

set network upload buffer size in MiB. (default: 0=no buffer)

--consider-part-files

treat part files like normal files (e.g. to list/extract them)

--debug-profile FILE

Write execution profile in Borg format into FILE. For local use a Python-compatible file can be generated by suffixing FILE with ".pyprof".

--rsh RSH

Use this command to connect to the 'borg serve' process (default: 'ssh')

Option --bypass-lock allows you to access the repository while bypassing borg's locking mechanism. This is necessary if your repository is on a read-only storage where you don't have write permissions or capabilities and therefore cannot create a lock. Examples are repositories stored on a Bluray disc or a read-only network storage. Avoid this option if you are able to use locks as that is the safer way; see the warning below.

Warning

If you do use --bypass-lock, you are responsible to ensure that no other borg instances have write access to the repository. Otherwise, you might experience errors and read broken data if changes to that repository are being made at the same time.

Examples
# Create an archive and log: borg version, files list, return code
$ borg create --show-version --list --show-rc /path/to/repo::my-files files

borg init

borg [common options] init [options] [REPOSITORY]

positional arguments

REPOSITORY

repository to create

options

-e MODE, --encryption MODE

select encryption key mode (required)

--append-only

create an append-only mode repository. Note that this only affects the low level structure of the repository, and running delete or prune will still be allowed. See Append-only mode (forbid compaction) in Additional Notes for more details.

--storage-quota QUOTA

Set storage quota of the new repository (e.g. 5G, 1.5T). Default: no quota.

--make-parent-dirs

create the parent directories of the repository directory, if they are missing.

Common options

Description

This command initializes an empty repository. A repository is a filesystem directory containing the deduplicated data from zero or more archives.

Encryption mode TLDR

The encryption mode can only be configured when creating a new repository - you can neither configure it on a per-archive basis nor change the encryption mode of an existing repository.

Use repokey:

borg init --encryption repokey /path/to/repo

Or repokey-blake2 depending on which is faster on your client machines (see below):

borg init --encryption repokey-blake2 /path/to/repo

Borg will:

  1. Ask you to come up with a passphrase.

  2. Create a borg key (which contains 3 random secrets. See Key files).

  3. Encrypt the key with your passphrase.

  4. Store the encrypted borg key inside the repository directory (in the repo config). This is why it is essential to use a secure passphrase.

  5. Encrypt and sign your backups to prevent anyone from reading or forging them unless they have the key and know the passphrase. Make sure to keep a backup of your key outside the repository - do not lock yourself out by "leaving your keys inside your car" (see borg key export). For remote backups the encryption is done locally - the remote machine never sees your passphrase, your unencrypted key or your unencrypted files. Chunking and id generation are also based on your key to improve your privacy.

  6. Use the key when extracting files to decrypt them and to verify that the contents of the backups have not been accidentally or maliciously altered.

Picking a passphrase

Make sure you use a good passphrase. Not too short, not too simple. The real encryption / decryption key is encrypted with / locked by your passphrase. If an attacker gets your key, he can't unlock and use it without knowing the passphrase.

Be careful with special or non-ascii characters in your passphrase:

  • Borg processes the passphrase as unicode (and encodes it as utf-8), so it does not have problems dealing with even the strangest characters.

  • BUT: that does not necessarily apply to your OS / VM / keyboard configuration.

So better use a long passphrase made from simple ascii chars than one that includes non-ascii stuff or characters that are hard/impossible to enter on a different keyboard layout.

You can change your passphrase for existing repos at any time, it won't affect the encryption/decryption key or other secrets.

More encryption modes

Only use --encryption none if you are OK with anyone who has access to your repository being able to read your backups and tamper with their contents without you noticing.

If you want "passphrase and having-the-key" security, use --encryption keyfile. The key will be stored in your home directory (in ~/.config/borg/keys).

If you do not want to encrypt the contents of your backups, but still want to detect malicious tampering use --encryption authenticated. To normally work with authenticated repos, you will need the passphrase, but there is an emergency workaround, see BORG_WORKAROUNDS=authenticated_no_key docs.

If BLAKE2b is faster than SHA-256 on your hardware, use --encryption authenticated-blake2, --encryption repokey-blake2 or --encryption keyfile-blake2. Note: for remote backups the hashing is done on your local machine.

Hash/MAC

Not encrypted no auth

Not encrypted, but authenticated

Encrypted (AEAD w/ AES) and authenticated

SHA-256

none

authenticated

repokey keyfile

BLAKE2b

n/a

authenticated-blake2

repokey-blake2 keyfile-blake2

Modes marked like this in the above table are new in Borg 1.1 and are not backwards-compatible with Borg 1.0.x.

On modern Intel/AMD CPUs (except very cheap ones), AES is usually hardware-accelerated. BLAKE2b is faster than SHA256 on Intel/AMD 64-bit CPUs (except AMD Ryzen and future CPUs with SHA extensions), which makes authenticated-blake2 faster than none and authenticated.

On modern ARM CPUs, NEON provides hardware acceleration for SHA256 making it faster than BLAKE2b-256 there. NEON accelerates AES as well.

Hardware acceleration is always used automatically when available.

repokey and keyfile use AES-CTR-256 for encryption and HMAC-SHA256 for authentication in an encrypt-then-MAC (EtM) construction. The chunk ID hash is HMAC-SHA256 as well (with a separate key). These modes are compatible with Borg 1.0.x.

repokey-blake2 and keyfile-blake2 are also authenticated encryption modes, but use BLAKE2b-256 instead of HMAC-SHA256 for authentication. The chunk ID hash is a keyed BLAKE2b-256 hash. These modes are new and not compatible with Borg 1.0.x.

authenticated mode uses no encryption, but authenticates repository contents through the same HMAC-SHA256 hash as the repokey and keyfile modes (it uses it as the chunk ID hash). The key is stored like repokey. This mode is new and not compatible with Borg 1.0.x.

authenticated-blake2 is like authenticated, but uses the keyed BLAKE2b-256 hash from the other blake2 modes. This mode is new and not compatible with Borg 1.0.x.

none mode uses no encryption and no authentication. It uses SHA256 as chunk ID hash. This mode is not recommended, you should rather consider using an authenticated or authenticated/encrypted mode. This mode has possible denial-of-service issues when running borg create on contents controlled by an attacker. Use it only for new repositories where no encryption is wanted and when compatibility with 1.0.x is important. If compatibility with 1.0.x is not important, use authenticated-blake2 or authenticated instead. This mode is compatible with Borg 1.0.x.

Examples
# Local repository, repokey encryption, BLAKE2b (often faster, since Borg 1.1)
$ borg init --encryption=repokey-blake2 /path/to/repo

# Local repository (no encryption)
$ borg init --encryption=none /path/to/repo

# Remote repository (accesses a remote borg via ssh)
# repokey: stores the (encrypted) key into <REPO_DIR>/config
$ borg init --encryption=repokey-blake2 user@hostname:backup

# Remote repository (accesses a remote borg via ssh)
# keyfile: stores the (encrypted) key into ~/.config/borg/keys/
$ borg init --encryption=keyfile user@hostname:backup

borg create

borg [common options] create [options] ARCHIVE [PATH...]

positional arguments

ARCHIVE

name of archive to create (must be also a valid directory name)

PATH

paths to archive

options

-n, --dry-run

do not create a backup archive

-s, --stats

print statistics for the created archive

--list

output verbose list of items (files, dirs, ...)

--filter STATUSCHARS

only display items with the given status characters (see description)

--json

output stats as JSON. Implies --stats.

--no-cache-sync

experimental: do not synchronize the cache. Implies not using the files cache.

--stdin-name NAME

use NAME in archive for stdin data (default: 'stdin')

--stdin-user USER

set user USER in archive for stdin data (default: 'root')

--stdin-group GROUP

set group GROUP in archive for stdin data (default: 'wheel')

--stdin-mode M

set mode to M in archive for stdin data (default: 0660)

--content-from-command

interpret PATH as command and store its stdout. See also section Reading from stdin below.

--paths-from-stdin

read DELIM-separated list of paths to backup from stdin. All control is external: it will back up all files given - no more, no less.

--paths-from-command

interpret PATH as command and treat its output as --paths-from-stdin

--paths-delimiter DELIM

set path delimiter for --paths-from-stdin and --paths-from-command (default: \n)

Common options

Include/Exclude options

-e PATTERN, --exclude PATTERN

exclude paths matching PATTERN

--exclude-from EXCLUDEFILE

read exclude patterns from EXCLUDEFILE, one per line

--pattern PATTERN

include/exclude paths matching PATTERN

--patterns-from PATTERNFILE

read include/exclude patterns from PATTERNFILE, one per line

--exclude-caches

exclude directories that contain a CACHEDIR.TAG file (http://www.bford.info/cachedir/spec.html)

--exclude-if-present NAME

exclude directories that are tagged by containing a filesystem object with the given NAME

--keep-exclude-tags

if tag objects are specified with --exclude-if-present, don't omit the tag objects themselves from the backup archive

--exclude-nodump

exclude files flagged NODUMP

Filesystem options

-x, --one-file-system

stay in the same file system and do not store mount points of other file systems - this might behave different from your expectations, see the description below.

--numeric-owner

deprecated, use --numeric-ids instead

--numeric-ids

only store numeric user and group identifiers

--noatime

do not store atime into archive

--atime

do store atime into archive

--noctime

do not store ctime into archive

--nobirthtime

do not store birthtime (creation date) into archive

--nobsdflags

deprecated, use --noflags instead

--noflags

do not read and store flags (e.g. NODUMP, IMMUTABLE) into archive

--noacls

do not read and store ACLs into archive

--noxattrs

do not read and store xattrs into archive

--sparse

detect sparse holes in input (supported only by fixed chunker)

--files-cache MODE

operate files cache in MODE. default: ctime,size,inode

--read-special

open and read block and char device files as well as FIFOs as if they were regular files. Also follows symlinks pointing to these kinds of files.

Archive options

--comment COMMENT

add a comment text to the archive

--timestamp TIMESTAMP

manually specify the archive creation date/time (UTC, yyyy-mm-ddThh:mm:ss format). Alternatively, give a reference file/directory.

-c SECONDS, --checkpoint-interval SECONDS

write checkpoint every SECONDS seconds (Default: 1800)

--chunker-params PARAMS

specify the chunker parameters (ALGO, CHUNK_MIN_EXP, CHUNK_MAX_EXP, HASH_MASK_BITS, HASH_WINDOW_SIZE). default: buzhash,19,23,21,4095

-C COMPRESSION, --compression COMPRESSION

select compression algorithm, see the output of the "borg help compression" command for details.

Description

This command creates a backup archive containing all files found while recursively traversing all paths specified. Paths are added to the archive as they are given, that means if relative paths are desired, the command has to be run from the correct directory.

When giving '-' as path, borg will read data from standard input and create a file 'stdin' in the created archive from that data. In some cases it's more appropriate to use --content-from-command, however. See section Reading from stdin below for details.

The archive will consume almost no disk space for files or parts of files that have already been stored in other archives.

The archive name needs to be unique. It must not end in '.checkpoint' or '.checkpoint.N' (with N being a number), because these names are used for checkpoints and treated in special ways.

In the archive name, you may use the following placeholders: {now}, {utcnow}, {fqdn}, {hostname}, {user} and some others.

Backup speed is increased by not reprocessing files that are already part of existing archives and weren't modified. The detection of unmodified files is done by comparing multiple file metadata values with previous values kept in the files cache.

This comparison can operate in different modes as given by --files-cache:

  • ctime,size,inode (default)

  • mtime,size,inode (default behaviour of borg versions older than 1.1.0rc4)

  • ctime,size (ignore the inode number)

  • mtime,size (ignore the inode number)

  • rechunk,ctime (all files are considered modified - rechunk, cache ctime)

  • rechunk,mtime (all files are considered modified - rechunk, cache mtime)

  • disabled (disable the files cache, all files considered modified - rechunk)

inode number: better safety, but often unstable on network filesystems

Normally, detecting file modifications will take inode information into consideration to improve the reliability of file change detection. This is problematic for files located on sshfs and similar network file systems which do not provide stable inode numbers, such files will always be considered modified. You can use modes without inode in this case to improve performance, but reliability of change detection might be reduced.

ctime vs. mtime: safety vs. speed

  • ctime is a rather safe way to detect changes to a file (metadata and contents) as it can not be set from userspace. But, a metadata-only change will already update the ctime, so there might be some unnecessary chunking/hashing even without content changes. Some filesystems do not support ctime (change time). E.g. doing a chown or chmod to a file will change its ctime.

  • mtime usually works and only updates if file contents were changed. But mtime can be arbitrarily set from userspace, e.g. to set mtime back to the same value it had before a content change happened. This can be used maliciously as well as well-meant, but in both cases mtime based cache modes can be problematic.

The mount points of filesystems or filesystem snapshots should be the same for every creation of a new archive to ensure fast operation. This is because the file cache that is used to determine changed files quickly uses absolute filenames. If this is not possible, consider creating a bind mount to a stable location.

The --progress option shows (from left to right) Original, Compressed and Deduplicated (O, C and D, respectively), then the Number of files (N) processed so far, followed by the currently processed path.

When using --stats, you will get some statistics about how much data was added - the "This Archive" deduplicated size there is most interesting as that is how much your repository will grow. Please note that the "All archives" stats refer to the state after creation. Also, the --stats and --dry-run options are mutually exclusive because the data is not actually compressed and deduplicated during a dry run.

For more help on include/exclude patterns, see the borg help patterns command output.

For more help on placeholders, see the borg help placeholders command output.

The --exclude patterns are not like tar. In tar --exclude .bundler/gems will exclude foo/.bundler/gems. In borg it will not, you need to use --exclude '*/.bundler/gems' to get the same effect.

In addition to using --exclude patterns, it is possible to use --exclude-if-present to specify the name of a filesystem object (e.g. a file or folder name) which, when contained within another folder, will prevent the containing folder from being backed up. By default, the containing folder and all of its contents will be omitted from the backup. If, however, you wish to only include the objects specified by --exclude-if-present in your backup, and not include any other contents of the containing folder, this can be enabled through using the --keep-exclude-tags option.

The -x or --one-file-system option excludes directories, that are mountpoints (and everything in them). It detects mountpoints by comparing the device number from the output of stat() of the directory and its parent directory. Specifically, it excludes directories for which stat() reports a device number different from the device number of their parent. In general: be aware that there are directories with device number different from their parent, which the kernel does not consider a mountpoint and also the other way around. Linux examples for this are bind mounts (possibly same device number, but always a mountpoint) and ALL subvolumes of a btrfs (different device number from parent but not necessarily a mountpoint). macOS examples are the apfs mounts of a typical macOS installation. Therefore, when using --one-file-system, you should double-check that the backup works as intended.

Item flags

--list outputs a list of all files, directories and other file system items it considered (no matter whether they had content changes or not). For each item, it prefixes a single-letter flag that indicates type and/or status of the item.

If you are interested only in a subset of that output, you can give e.g. --filter=AME and it will only show regular files with A, M or E status (see below).

A uppercase character represents the status of a regular file relative to the "files" cache (not relative to the repo -- this is an issue if the files cache is not used). Metadata is stored in any case and for 'A' and 'M' also new data chunks are stored. For 'U' all data chunks refer to already existing chunks.

  • 'A' = regular file, added (see also I am seeing 'A' (added) status for an unchanged file!? in the FAQ)

  • 'M' = regular file, modified

  • 'U' = regular file, unchanged

  • 'C' = regular file, it changed while we backed it up

  • 'E' = regular file, an error happened while accessing/reading this file

A lowercase character means a file type other than a regular file, borg usually just stores their metadata:

  • 'd' = directory

  • 'b' = block device

  • 'c' = char device

  • 'h' = regular file, hardlink (to already seen inodes)

  • 's' = symlink

  • 'f' = fifo

Other flags used include:

  • 'i' = backup data was read from standard input (stdin)

  • '-' = dry run, item was not backed up

  • 'x' = excluded, item was not backed up

  • '?' = missing status code (if you see this, please file a bug report!)

Reading from stdin

There are two methods to read from stdin. Either specify - as path and pipe directly to borg:

backup-vm --id myvm --stdout | borg create REPO::ARCHIVE -

Or use --content-from-command to have Borg manage the execution of the command and piping. If you do so, the first PATH argument is interpreted as command to execute and any further arguments are treated as arguments to the command:

borg create --content-from-command REPO::ARCHIVE -- backup-vm --id myvm --stdout

-- is used to ensure --id and --stdout are not considered arguments to borg but rather backup-vm.

The difference between the two approaches is that piping to borg creates an archive even if the command piping to borg exits with a failure. In this case, one can end up with truncated output being backed up. Using --content-from-command, in contrast, borg is guaranteed to fail without creating an archive should the command fail. The command is considered failed when it returned a non-zero exit code.

Reading from stdin yields just a stream of data without file metadata associated with it, and the files cache is not needed at all. So it is safe to disable it via --files-cache disabled and speed up backup creation a bit.

By default, the content read from stdin is stored in a file called 'stdin'. Use --stdin-name to change the name.

Examples
# Backup ~/Documents into an archive named "my-documents"
$ borg create /path/to/repo::my-documents ~/Documents

# same, but list all files as we process them
$ borg create --list /path/to/repo::my-documents ~/Documents

# Backup ~/Documents and ~/src but exclude pyc files
$ borg create /path/to/repo::my-files \
    ~/Documents                       \
    ~/src                             \
    --exclude '*.pyc'

# Backup home directories excluding image thumbnails (i.e. only
# /home/<one directory>/.thumbnails is excluded, not /home/*/*/.thumbnails etc.)
$ borg create /path/to/repo::my-files /home \
    --exclude 'sh:home/*/.thumbnails'

# Backup the root filesystem into an archive named "root-YYYY-MM-DD"
# use zlib compression (good, but slow) - default is lz4 (fast, low compression ratio)
$ borg create -C zlib,6 --one-file-system /path/to/repo::root-{now:%Y-%m-%d} /

# Backup onto a remote host ("push" style) via ssh to port 2222,
# logging in as user "borg" and storing into /path/to/repo
$ borg create ssh://borg@backup.example.org:2222/path/to/repo::{fqdn}-root-{now} /

# Backup a remote host locally ("pull" style) using sshfs
$ mkdir sshfs-mount
$ sshfs root@example.com:/ sshfs-mount
$ cd sshfs-mount
$ borg create /path/to/repo::example.com-root-{now:%Y-%m-%d} .
$ cd ..
$ fusermount -u sshfs-mount

# Make a big effort in fine granular deduplication (big chunk management
# overhead, needs a lot of RAM and disk space, see formula in internals
# docs - same parameters as borg < 1.0 or attic):
$ borg create --chunker-params buzhash,10,23,16,4095 /path/to/repo::small /smallstuff

# Backup a raw device (must not be active/in use/mounted at that time)
$ borg create --read-special --chunker-params fixed,4194304 /path/to/repo::my-sdx /dev/sdX

# Backup a sparse disk image (must not be active/in use/mounted at that time)
$ borg create --sparse --chunker-params fixed,4194304 /path/to/repo::my-disk my-disk.raw

# No compression (none)
$ borg create --compression none /path/to/repo::arch ~

# Super fast, low compression (lz4, default)
$ borg create /path/to/repo::arch ~

# Less fast, higher compression (zlib, N = 0..9)
$ borg create --compression zlib,N /path/to/repo::arch ~

# Even slower, even higher compression (lzma, N = 0..9)
$ borg create --compression lzma,N /path/to/repo::arch ~

# Only compress compressible data with lzma,N (N = 0..9)
$ borg create --compression auto,lzma,N /path/to/repo::arch ~

# Use short hostname, user name and current time in archive name
$ borg create /path/to/repo::{hostname}-{user}-{now} ~
# Similar, use the same datetime format that is default as of borg 1.1
$ borg create /path/to/repo::{hostname}-{user}-{now:%Y-%m-%dT%H:%M:%S} ~
# As above, but add nanoseconds
$ borg create /path/to/repo::{hostname}-{user}-{now:%Y-%m-%dT%H:%M:%S.%f} ~

# Backing up relative paths by moving into the correct directory first
$ cd /home/user/Documents
# The root directory of the archive will be "projectA"
$ borg create /path/to/repo::daily-projectA-{now:%Y-%m-%d} projectA

# Use external command to determine files to archive
# Use --paths-from-stdin with find to only backup files less than 1MB in size
$ find ~ -size -1000k | borg create --paths-from-stdin /path/to/repo::small-files-only
# Use --paths-from-command with find to only backup files from a given user
$ borg create --paths-from-command /path/to/repo::joes-files -- find /srv/samba/shared -user joe
# Use --paths-from-stdin with --paths-delimiter (for example, for filenames with newlines in them)
$ find ~ -size -1000k -print0 | borg create \
    --paths-from-stdin \
    --paths-delimiter "\0" \
    /path/to/repo::smallfiles-handle-newline

borg extract

borg [common options] extract [options] ARCHIVE [PATH...]

positional arguments

ARCHIVE

archive to extract

PATH

paths to extract; patterns are supported

options

--list

output verbose list of items (files, dirs, ...)

-n, --dry-run

do not actually change any files

--numeric-owner

deprecated, use --numeric-ids instead

--numeric-ids

only obey numeric user and group identifiers

--nobsdflags

deprecated, use --noflags instead

--noflags

do not extract/set flags (e.g. NODUMP, IMMUTABLE)

--noacls

do not extract/set ACLs

--noxattrs

do not extract/set xattrs

--stdout

write all extracted data to stdout

--sparse

create holes in output sparse file from all-zero chunks

Common options

Include/Exclude options

-e PATTERN, --exclude PATTERN

exclude paths matching PATTERN

--exclude-from EXCLUDEFILE

read exclude patterns from EXCLUDEFILE, one per line

--pattern PATTERN

include/exclude paths matching PATTERN

--patterns-from PATTERNFILE

read include/exclude patterns from PATTERNFILE, one per line

--strip-components NUMBER

Remove the specified number of leading path elements. Paths with fewer elements will be silently skipped.

Description

This command extracts the contents of an archive. By default the entire archive is extracted but a subset of files and directories can be selected by passing a list of PATHs as arguments. The file selection can further be restricted by using the --exclude option.

For more help on include/exclude patterns, see the borg help patterns command output.

By using --dry-run, you can do all extraction steps except actually writing the output data: reading metadata and data chunks from the repo, checking the hash/hmac, decrypting, decompressing.

--progress can be slower than no progress display, since it makes one additional pass over the archive metadata.

Note

Currently, extract always writes into the current working directory ("."), so make sure you cd to the right place before calling borg extract.

When parent directories are not extracted (because of using file/directory selection or any other reason), borg can not restore parent directories' metadata, e.g. owner, group, permission, etc.

Examples
# Extract entire archive
$ borg extract /path/to/repo::my-files

# Extract entire archive and list files while processing
$ borg extract --list /path/to/repo::my-files

# Verify whether an archive could be successfully extracted, but do not write files to disk
$ borg extract --dry-run /path/to/repo::my-files

# Extract the "src" directory
$ borg extract /path/to/repo::my-files home/USERNAME/src

# Extract the "src" directory but exclude object files
$ borg extract /path/to/repo::my-files home/USERNAME/src --exclude '*.o'

# Restore a raw device (must not be active/in use/mounted at that time)
$ borg extract --stdout /path/to/repo::my-sdx | dd of=/dev/sdx bs=10M

borg check

borg [common options] check [options] [REPOSITORY_OR_ARCHIVE]

positional arguments

REPOSITORY_OR_ARCHIVE

repository or archive to check consistency of

options

--repository-only

only perform repository checks

--archives-only

only perform archives checks

--verify-data

perform cryptographic archive data integrity verification (conflicts with --repository-only)

--repair

attempt to repair any inconsistencies found

--save-space

work slower, but using less space

--max-duration SECONDS

do only a partial repo check for max. SECONDS seconds (Default: unlimited)

Common options

Archive filters — Archive filters can be applied to repository targets.

-P PREFIX, --prefix PREFIX

only consider archive names starting with this prefix. (deprecated)

-a GLOB, --glob-archives GLOB

only consider archive names matching the glob. sh: rules apply (without actually using the sh: prefix), see "borg help patterns".

--sort-by KEYS

Comma-separated list of sorting keys; valid keys are: timestamp, archive, name, id; default is: timestamp

--first N

consider first N archives after other filters were applied

--last N

consider last N archives after other filters were applied

Description

The check command verifies the consistency of a repository and its archives. It consists of two major steps:

  1. Checking the consistency of the repository itself. This includes checking the segment magic headers, and both the metadata and data of all objects in the segments. The read data is checked by size and CRC. Bit rot and other types of accidental damage can be detected this way. Running the repository check can be split into multiple partial checks using --max-duration. When checking a remote repository, please note that the checks run on the server and do not cause significant network traffic.

  2. Checking consistency and correctness of the archive metadata and optionally archive data (requires --verify-data). This includes ensuring that the repository manifest exists, the archive metadata chunk is present, and that all chunks referencing files (items) in the archive exist. This requires reading archive and file metadata, but not data. To cryptographically verify the file (content) data integrity pass --verify-data, but keep in mind that this requires reading all data and is hence very time consuming. When checking archives of a remote repository, archive checks run on the client machine because they require decrypting data and therefore the encryption key.

Both steps can also be run independently. Pass --repository-only to run the repository checks only, or pass --archives-only to run the archive checks only.

The --max-duration option can be used to split a long-running repository check into multiple partial checks. After the given number of seconds the check is interrupted. The next partial check will continue where the previous one stopped, until the full repository has been checked. Assuming a complete check would take 7 hours, then running a daily check with --max-duration=3600 (1 hour) would result in one full repository check per week. Doing a full repository check aborts any previous partial check; the next partial check will restart from the beginning. With partial repository checks you can run neither archive checks, nor enable repair mode. Consequently, if you want to use --max-duration you must also pass --repository-only, and must not pass --archives-only, nor --repair.

Warning: Please note that partial repository checks (i.e. running it with --max-duration) can only perform non-cryptographic checksum checks on the segment files. A full repository check (i.e. without --max-duration) can also do a repository index check. Enabling partial repository checks excepts archive checks for the same reason. Therefore partial checks may be useful with very large repositories only where a full check would take too long.

The --verify-data option will perform a full integrity verification (as opposed to checking the CRC32 of the segment) of data, which means reading the data from the repository, decrypting and decompressing it. It is a complete cryptographic verification and hence very time consuming, but will detect any accidental and malicious corruption. Tamper-resistance is only guaranteed for encrypted repositories against attackers without access to the keys. You can not use --verify-data with --repository-only.

About repair mode

The check command is a readonly task by default. If any corruption is found, Borg will report the issue and proceed with checking. To actually repair the issues found, pass --repair.

Note

--repair is a POTENTIALLY DANGEROUS FEATURE and might lead to data loss! This does not just include data that was previously lost anyway, but might include more data for kinds of corruption it is not capable of dealing with. BE VERY CAREFUL!

Pursuant to the previous warning it is also highly recommended to test the reliability of the hardware running Borg with stress testing software. This especially includes storage and memory testers. Unreliable hardware might lead to additional data loss.

It is highly recommended to create a backup of your repository before running in repair mode (i.e. running it with --repair).

Repair mode will attempt to fix any corruptions found. Fixing corruptions does not mean recovering lost data: Borg can not magically restore data lost due to e.g. a hardware failure. Repairing a repository means sacrificing some data for the sake of the repository as a whole and the remaining data. Hence it is, by definition, a potentially lossy task.

In practice, repair mode hooks into both the repository and archive checks:

  1. When checking the repository's consistency, repair mode will try to recover as many objects from segments with integrity errors as possible, and ensure that the index is consistent with the data stored in the segments.

  2. When checking the consistency and correctness of archives, repair mode might remove whole archives from the manifest if their archive metadata chunk is corrupt or lost. On a chunk level (i.e. the contents of files), repair mode will replace corrupt or lost chunks with a same-size replacement chunk of zeroes. If a previously zeroed chunk reappears, repair mode will restore this lost chunk using the new chunk. Lastly, repair mode will also delete orphaned chunks (e.g. caused by read errors while creating the archive).

Most steps taken by repair mode have a one-time effect on the repository, like removing a lost archive from the repository. However, replacing a corrupt or lost chunk with an all-zero replacement will have an ongoing effect on the repository: When attempting to extract a file referencing an all-zero chunk, the extract command will distinctly warn about it. The FUSE filesystem created by the mount command will reject reading such a "zero-patched" file unless a special mount option is given.

As mentioned earlier, Borg might be able to "heal" a "zero-patched" file in repair mode, if all its previously lost chunks reappear (e.g. via a later backup). This is achieved by Borg not only keeping track of the all-zero replacement chunks, but also by keeping metadata about the lost chunks. In repair mode Borg will check whether a previously lost chunk reappeared and will replace the all-zero replacement chunk by the reappeared chunk. If all lost chunks of a "zero-patched" file reappear, this effectively "heals" the file. Consequently, if lost chunks were repaired earlier, it is advised to run --repair a second time after creating some new backups.

borg rename

borg [common options] rename [options] ARCHIVE NEWNAME

positional arguments

ARCHIVE

archive to rename

NEWNAME

the new archive name to use

Common options

Description

This command renames an archive in the repository.

This results in a different archive ID.

Examples
$ borg create /path/to/repo::archivename ~
$ borg list /path/to/repo
archivename                          Mon, 2016-02-15 19:50:19

$ borg rename /path/to/repo::archivename newname
$ borg list /path/to/repo
newname                              Mon, 2016-02-15 19:50:19

borg list

borg [common options] list [options] [REPOSITORY_OR_ARCHIVE] [PATH...]

positional arguments

REPOSITORY_OR_ARCHIVE

repository or archive to list contents of

PATH

paths to list; patterns are supported

options

--consider-checkpoints

Show checkpoint archives in the repository contents list (default: hidden).

--short

only print file/directory names, nothing else

--format FORMAT

specify format for file or archive listing (default for files: "{mode} {user:6} {group:6} {size:8} {mtime} {path}{extra}{NL}"; for archives: "{archive:<36} {time} [{id}]{NL}")

--json

Only valid for listing repository contents. Format output as JSON. The form of --format is ignored, but keys used in it are added to the JSON output. Some keys are always present. Note: JSON can only represent text. A "barchive" key is therefore not available.

--json-lines

Only valid for listing archive contents. Format output as JSON Lines. The form of --format is ignored, but keys used in it are added to the JSON output. Some keys are always present. Note: JSON can only represent text. A "bpath" key is therefore not available.

Common options

Archive filters — Archive filters can be applied to repository targets.

-P PREFIX, --prefix PREFIX

only consider archive names starting with this prefix. (deprecated)

-a GLOB, --glob-archives GLOB

only consider archive names matching the glob. sh: rules apply (without actually using the sh: prefix), see "borg help patterns".

--sort-by KEYS

Comma-separated list of sorting keys; valid keys are: timestamp, archive, name, id; default is: timestamp

--first N

consider first N archives after other filters were applied

--last N

consider last N archives after other filters were applied

Include/Exclude options

-e PATTERN, --exclude PATTERN

exclude paths matching PATTERN

--exclude-from EXCLUDEFILE

read exclude patterns from EXCLUDEFILE, one per line

--pattern PATTERN

include/exclude paths matching PATTERN

--patterns-from PATTERNFILE

read include/exclude patterns from PATTERNFILE, one per line

Description

This command lists the contents of a repository or an archive.

For more help on include/exclude patterns, see the borg help patterns command output.

The FORMAT specifier syntax

The --format option uses python's format string syntax.

Examples:

$ borg list --format '{archive}{NL}' /path/to/repo
ArchiveFoo
ArchiveBar
...

# {VAR:NUMBER} - pad to NUMBER columns.
# Strings are left-aligned, numbers are right-aligned.
# Note: time columns except ``isomtime``, ``isoctime`` and ``isoatime`` cannot be padded.
$ borg list --format '{archive:36} {time} [{id}]{NL}' /path/to/repo
ArchiveFoo                           Thu, 2021-12-09 10:22:28 [0b8e9a312bef3f2f6e2d0fc110c196827786c15eba0188738e81697a7fa3b274]
$ borg list --format '{mode} {user:6} {group:6} {size:8} {mtime} {path}{extra}{NL}' /path/to/repo::ArchiveFoo
-rw-rw-r-- user   user       1024 Thu, 2021-12-09 10:22:17 file-foo
...

# {VAR:<NUMBER} - pad to NUMBER columns left-aligned.
# {VAR:>NUMBER} - pad to NUMBER columns right-aligned.
$ borg list --format '{mode} {user:>6} {group:>6} {size:<8} {mtime} {path}{extra}{NL}' /path/to/repo::ArchiveFoo
-rw-rw-r--   user   user 1024     Thu, 2021-12-09 10:22:17 file-foo
...

The following keys are always available:

  • NEWLINE: OS dependent line separator

  • NL: alias of NEWLINE

  • NUL: NUL character for creating print0 / xargs -0 like output, see barchive and bpath keys below

  • SPACE

  • TAB

  • CR

  • LF

Keys available only when listing archives in a repository:

  • archive: archive name interpreted as text (might be missing non-text characters, see barchive)

  • name: alias of "archive"

  • barchive: verbatim archive name, can contain any character except NUL

  • comment: archive comment interpreted as text (might be missing non-text characters, see bcomment)

  • bcomment: verbatim archive comment, can contain any character except NUL

  • id: internal ID of the archive

  • tam: TAM authentication state of this archive

  • start: time (start) of creation of the archive

  • time: alias of "start"

  • end: time (end) of creation of the archive

  • command_line: command line which was used to create the archive

  • hostname: hostname of host on which this archive was created

  • username: username of user who created this archive

Keys available only when listing files in an archive:

  • type

  • mode

  • uid

  • gid

  • user

  • group

  • path: path interpreted as text (might be missing non-text characters, see bpath)

  • bpath: verbatim POSIX path, can contain any character except NUL

  • source: link target for links (identical to linktarget)

  • linktarget

  • flags

  • size

  • csize: compressed size

  • dsize: deduplicated size

  • dcsize: deduplicated compressed size

  • num_chunks: number of chunks in this file

  • unique_chunks: number of unique chunks in this file

  • mtime

  • ctime

  • atime

  • isomtime

  • isoctime

  • isoatime

  • blake2b

  • blake2s

  • md5

  • sha1

  • sha224

  • sha256

  • sha384

  • sha3_224

  • sha3_256

  • sha3_384

  • sha3_512

  • sha512

  • xxh64: XXH64 checksum of this file (note: this is NOT a cryptographic hash!)

  • archiveid

  • archivename

  • extra: prepends {source} with " -> " for soft links and " link to " for hard links

  • health: either "healthy" (file ok) or "broken" (if file has all-zero replacement chunks)

Examples
$ borg list /path/to/repo
Monday                               Mon, 2016-02-15 19:15:11
repo                                 Mon, 2016-02-15 19:26:54
root-2016-02-15                      Mon, 2016-02-15 19:36:29
newname                              Mon, 2016-02-15 19:50:19
...

$ borg list /path/to/repo::root-2016-02-15
drwxr-xr-x root   root          0 Mon, 2016-02-15 17:44:27 .
drwxrwxr-x root   root          0 Mon, 2016-02-15 19:04:49 bin
-rwxr-xr-x root   root    1029624 Thu, 2014-11-13 00:08:51 bin/bash
lrwxrwxrwx root   root          0 Fri, 2015-03-27 20:24:26 bin/bzcmp -> bzdiff
-rwxr-xr-x root   root       2140 Fri, 2015-03-27 20:24:22 bin/bzdiff
...

$ borg list /path/to/repo::root-2016-02-15 --pattern "- bin/ba*"
drwxr-xr-x root   root          0 Mon, 2016-02-15 17:44:27 .
drwxrwxr-x root   root          0 Mon, 2016-02-15 19:04:49 bin
lrwxrwxrwx root   root          0 Fri, 2015-03-27 20:24:26 bin/bzcmp -> bzdiff
-rwxr-xr-x root   root       2140 Fri, 2015-03-27 20:24:22 bin/bzdiff
...

$ borg list /path/to/repo::archiveA --format="{mode} {user:6} {group:6} {size:8d} {isomtime} {path}{extra}{NEWLINE}"
drwxrwxr-x user   user          0 Sun, 2015-02-01 11:00:00 .
drwxrwxr-x user   user          0 Sun, 2015-02-01 11:00:00 code
drwxrwxr-x user   user          0 Sun, 2015-02-01 11:00:00 code/myproject
-rw-rw-r-- user   user    1416192 Sun, 2015-02-01 11:00:00 code/myproject/file.ext
-rw-rw-r-- user   user    1416192 Sun, 2015-02-01 11:00:00 code/myproject/file.text
...

$ borg list /path/to/repo/::archiveA --pattern '+ re:\.ext$' --pattern '- re:^.*$'
-rw-rw-r-- user   user    1416192 Sun, 2015-02-01 11:00:00 code/myproject/file.ext
...

$ borg list /path/to/repo/::archiveA --pattern '+ re:.ext$' --pattern '- re:^.*$'
-rw-rw-r-- user   user    1416192 Sun, 2015-02-01 11:00:00 code/myproject/file.ext
-rw-rw-r-- user   user    1416192 Sun, 2015-02-01 11:00:00 code/myproject/file.text
...

borg diff

borg [common options] diff [options] REPO::ARCHIVE1 ARCHIVE2 [PATH...]

positional arguments

REPO::ARCHIVE1

repository location and ARCHIVE1 name

ARCHIVE2

ARCHIVE2 name (no repository location allowed)

PATH

paths of items inside the archives to compare; patterns are supported

options

--numeric-owner

deprecated, use --numeric-ids instead

--numeric-ids

only consider numeric user and group identifiers

--same-chunker-params

Override check of chunker parameters.

--sort

Sort the output lines by file path.

--content-only

Only compare differences in content (exclude metadata differences)

--json-lines

Format output as JSON Lines.

Common options

Include/Exclude options

-e PATTERN, --exclude PATTERN

exclude paths matching PATTERN

--exclude-from EXCLUDEFILE

read exclude patterns from EXCLUDEFILE, one per line

--pattern PATTERN

include/exclude paths matching PATTERN

--patterns-from PATTERNFILE

read include/exclude patterns from PATTERNFILE, one per line

Description

This command finds differences (file contents, user/group/mode) between archives.

A repository location and an archive name must be specified for REPO::ARCHIVE1. ARCHIVE2 is just another archive name in same repository (no repository location allowed).

For archives created with Borg 1.1 or newer diff automatically detects whether the archives are created with the same chunker params. If so, only chunk IDs are compared, which is very fast.

For archives prior to Borg 1.1 chunk contents are compared by default. If you did not create the archives with different chunker params, pass --same-chunker-params. Note that the chunker params changed from Borg 0.xx to 1.0.

For more help on include/exclude patterns, see the borg help patterns command output.

Examples
$ borg init -e=none testrepo
$ mkdir testdir
$ cd testdir
$ echo asdf > file1
$ dd if=/dev/urandom bs=1M count=4 > file2
$ touch file3
$ borg create ../testrepo::archive1 .

$ chmod a+x file1
$ echo "something" >> file2
$ borg create ../testrepo::archive2 .

$ echo "testing 123" >> file1
$ rm file3
$ touch file4
$ borg create ../testrepo::archive3 .

$ cd ..
$ borg diff testrepo::archive1 archive2
[-rw-r--r-- -> -rwxr-xr-x] file1
   +135 B    -252 B file2

$ borg diff testrepo::archive2 archive3
    +17 B      -5 B file1
added           0 B file4
removed         0 B file3

$ borg diff testrepo::archive1 archive3
    +17 B      -5 B [-rw-r--r-- -> -rwxr-xr-x] file1
   +135 B    -252 B file2
added           0 B file4
removed         0 B file3

$ borg diff --json-lines testrepo::archive1 archive3
{"path": "file1", "changes": [{"type": "modified", "added": 17, "removed": 5}, {"type": "mode", "old_mode": "-rw-r--r--", "new_mode": "-rwxr-xr-x"}]}
{"path": "file2", "changes": [{"type": "modified", "added": 135, "removed": 252}]}
{"path": "file4", "changes": [{"type": "added", "size": 0}]}
{"path": "file3", "changes": [{"type": "removed", "size": 0}]}

borg delete

borg [common options] delete [options] [REPOSITORY_OR_ARCHIVE] [ARCHIVE...]

positional arguments

REPOSITORY_OR_ARCHIVE

repository or archive to delete

ARCHIVE

archives to delete

options

-n, --dry-run

do not change repository

--list

output verbose list of archives

-s, --stats

print statistics for the deleted archive

--cache-only

delete only the local cache for the given repository

--force

force deletion of corrupted archives, use --force --force in case --force does not work.

--keep-security-info

keep the local security info when deleting a repository

--save-space

work slower, but using less space

-c SECONDS, --checkpoint-interval SECONDS

write checkpoint every SECONDS seconds (Default: 1800)

Common options

Archive filters — Archive filters can be applied to repository targets.

-P PREFIX, --prefix PREFIX

only consider archive names starting with this prefix. (deprecated)

-a GLOB, --glob-archives GLOB

only consider archive names matching the glob. sh: rules apply (without actually using the sh: prefix), see "borg help patterns".

--sort-by KEYS

Comma-separated list of sorting keys; valid keys are: timestamp, archive, name, id; default is: timestamp

--first N

consider first N archives after other filters were applied

--last N

consider last N archives after other filters were applied

Description

This command deletes an archive from the repository or the complete repository.

Important: When deleting archives, repository disk space is not freed until you run borg compact.

When you delete a complete repository, the security info and local cache for it (if any) are also deleted. Alternatively, you can delete just the local cache with the --cache-only option, or keep the security info with the --keep-security-info option.

When in doubt, use --dry-run --list to see what would be deleted.

When using --stats, you will get some statistics about how much data was deleted - the "Deleted data" deduplicated size there is most interesting as that is how much your repository will shrink. Please note that the "All archives" stats refer to the state after deletion.

You can delete multiple archives by specifying a shell pattern to match multiple archives using the --glob-archives GLOB option (for more info on these patterns, see borg help patterns).

To avoid accidentally deleting archives, especially when using glob patterns, it might be helpful to use the --dry-run to test out the command without actually making any changes to the repository.

Examples
# delete a single backup archive:
$ borg delete /path/to/repo::Monday
# actually free disk space:
$ borg compact /path/to/repo

# delete all archives whose names begin with the machine's hostname followed by "-"
$ borg delete --glob-archives '{hostname}-*' /path/to/repo

# delete all archives whose names contain "-2012-"
$ borg delete --glob-archives '*-2012-*' /path/to/repo

# see what would be deleted if delete was run without --dry-run
$ borg delete --list --dry-run -a '*-May-*' /path/to/repo

# delete the whole repository and the related local cache:
$ borg delete /path/to/repo
You requested to completely DELETE the repository *including* all archives it contains:
repo                                 Mon, 2016-02-15 19:26:54
root-2016-02-15                      Mon, 2016-02-15 19:36:29
newname                              Mon, 2016-02-15 19:50:19
Type 'YES' if you understand this and want to continue: YES

borg prune

borg [common options] prune [options] [REPOSITORY]

positional arguments

REPOSITORY

repository to prune

options

-n, --dry-run

do not change repository

--force

force pruning of corrupted archives, use --force --force in case --force does not work.

-s, --stats

print statistics for the deleted archive

--list

output verbose list of archives it keeps/prunes

--keep-within INTERVAL

keep all archives within this time interval

--keep-last, --keep-secondly

number of secondly archives to keep

--keep-minutely

number of minutely archives to keep

-H, --keep-hourly

number of hourly archives to keep

-d, --keep-daily

number of daily archives to keep

-w, --keep-weekly

number of weekly archives to keep

-m, --keep-monthly

number of monthly archives to keep

-y, --keep-yearly

number of yearly archives to keep

--save-space

work slower, but using less space

-c SECONDS, --checkpoint-interval SECONDS

write checkpoint every SECONDS seconds (Default: 1800)

Common options

Archive filters — Archive filters can be applied to repository targets.

-P PREFIX, --prefix PREFIX

only consider archive names starting with this prefix. (deprecated)

-a GLOB, --glob-archives GLOB

only consider archive names matching the glob. sh: rules apply (without actually using the sh: prefix), see "borg help patterns".

Description

The prune command prunes a repository by deleting all archives not matching any of the specified retention options.

Important: Repository disk space is not freed until you run borg compact.

This command is normally used by automated backup scripts wanting to keep a certain number of historic backups. This retention policy is commonly referred to as GFS (Grandfather-father-son) backup rotation scheme.

Also, prune automatically removes checkpoint archives (incomplete archives left behind by interrupted backup runs) except if the checkpoint is the latest archive (and thus still needed). Checkpoint archives are not considered when comparing archive counts against the retention limits (--keep-X).

If a prefix is set with -P, then only archives that start with the prefix are considered for deletion and only those archives count towards the totals specified by the rules. Otherwise, all archives in the repository are candidates for deletion! There is no automatic distinction between archives representing different contents. These need to be distinguished by specifying matching prefixes.

If you have multiple sequences of archives with different data sets (e.g. from different machines) in one shared repository, use one prune call per data set that matches only the respective archives using the -P option.

The --keep-within option takes an argument of the form "<int><char>", where char is "H", "d", "w", "m", "y". For example, --keep-within 2d means to keep all archives that were created within the past 48 hours. "1m" is taken to mean "31d". The archives kept with this option do not count towards the totals specified by any other options.

A good procedure is to thin out more and more the older your backups get. As an example, --keep-daily 7 means to keep the latest backup on each day, up to 7 most recent days with backups (days without backups do not count). The rules are applied from secondly to yearly, and backups selected by previous rules do not count towards those of later rules. The time that each backup starts is used for pruning purposes. Dates and times are interpreted in the local timezone, and weeks go from Monday to Sunday. Specifying a negative number of archives to keep means that there is no limit. As of borg 1.2.0, borg will retain the oldest archive if any of the secondly, minutely, hourly, daily, weekly, monthly, or yearly rules was not otherwise able to meet its retention target. This enables the first chronological archive to continue aging until it is replaced by a newer archive that meets the retention criteria.

The --keep-last N option is doing the same as --keep-secondly N (and it will keep the last N archives under the assumption that you do not create more than one backup archive in the same second).

When using --stats, you will get some statistics about how much data was deleted - the "Deleted data" deduplicated size there is most interesting as that is how much your repository will shrink. Please note that the "All archives" stats refer to the state after pruning.

Examples

Be careful, prune is a potentially dangerous command, it will remove backup archives.

The default of prune is to apply to all archives in the repository unless you restrict its operation to a subset of the archives using --glob-archives. When using --glob-archives, be careful to choose a good matching pattern - e.g. do not use "foo*" if you do not also want to match "foobar".

It is strongly recommended to always run prune -v --list --dry-run ... first so you will see what it would do without it actually doing anything.

# Keep 7 end of day and 4 additional end of week archives.
# Do a dry-run without actually deleting anything.
$ borg prune -v --list --dry-run --keep-daily=7 --keep-weekly=4 /path/to/repo

# Same as above but only apply to archive names starting with the hostname
# of the machine followed by a "-" character:
$ borg prune -v --list --keep-daily=7 --keep-weekly=4 --glob-archives='{hostname}-*' /path/to/repo
# actually free disk space:
$ borg compact /path/to/repo

# Keep 7 end of day, 4 additional end of week archives,
# and an end of month archive for every month:
$ borg prune -v --list --keep-daily=7 --keep-weekly=4 --keep-monthly=-1 /path/to/repo

# Keep all backups in the last 10 days, 4 additional end of week archives,
# and an end of month archive for every month:
$ borg prune -v --list --keep-within=10d --keep-weekly=4 --keep-monthly=-1 /path/to/repo

There is also a visualized prune example in docs/misc/prune-example.txt:

borg prune visualized
=====================

Assume it is 2016-01-01, today's backup has not yet been made, you have
created at least one backup on each day in 2015 except on 2015-12-19 (no
backup made on that day), and you started backing up with borg on
2015-01-01.

This is what borg prune --keep-daily 14 --keep-monthly 6 --keep-yearly 1
would keep.

Backups kept by the --keep-daily rule are marked by a "d" to the right,
backups kept by the --keep-monthly rule are marked by a "m" to the right,
and backups kept by the --keep-yearly rule are marked by a "y" to the
right.

Calendar view
-------------

                            2015
      January               February               March
Mo Tu We Th Fr Sa Su  Mo Tu We Th Fr Sa Su  Mo Tu We Th Fr Sa Su
          1y 2  3  4                     1                     1
 5  6  7  8  9 10 11   2  3  4  5  6  7  8   2  3  4  5  6  7  8
12 13 14 15 16 17 18   9 10 11 12 13 14 15   9 10 11 12 13 14 15
19 20 21 22 23 24 25  16 17 18 19 20 21 22  16 17 18 19 20 21 22
26 27 28 29 30 31     23 24 25 26 27 28     23 24 25 26 27 28 29
                                            30 31

       April                  May                   June
Mo Tu We Th Fr Sa Su  Mo Tu We Th Fr Sa Su  Mo Tu We Th Fr Sa Su
       1  2  3  4  5               1  2  3   1  2  3  4  5  6  7
 6  7  8  9 10 11 12   4  5  6  7  8  9 10   8  9 10 11 12 13 14
13 14 15 16 17 18 19  11 12 13 14 15 16 17  15 16 17 18 19 20 21
20 21 22 23 24 25 26  18 19 20 21 22 23 24  22 23 24 25 26 27 28
27 28 29 30           25 26 27 28 29 30 31  29 30m


        July                 August              September
Mo Tu We Th Fr Sa Su  Mo Tu We Th Fr Sa Su  Mo Tu We Th Fr Sa Su
       1  2  3  4  5                  1  2      1  2  3  4  5  6
 6  7  8  9 10 11 12   3  4  5  6  7  8  9   7  8  9 10 11 12 13
13 14 15 16 17 18 19  10 11 12 13 14 15 16  14 15 16 17 18 19 20
20 21 22 23 24 25 26  17 18 19 20 21 22 23  21 22 23 24 25 26 27
27 28 29 30 31m       24 25 26 27 28 29 30  28 29 30m
                      31m

      October               November              December
Mo Tu We Th Fr Sa Su  Mo Tu We Th Fr Sa Su  Mo Tu We Th Fr Sa Su
          1  2  3  4                     1      1  2  3  4  5  6
 5  6  7  8  9 10 11   2  3  4  5  6  7  8   7  8  9 10 11 12 13
12 13 14 15 16 17 18   9 10 11 12 13 14 15  14 15 16 17d18d19 20d
19 20 21 22 23 24 25  16 17 18 19 20 21 22  21d22d23d24d25d26d27d
26 27 28 29 30 31m    23 24 25 26 27 28 29  28d29d30d31d
                      30m

List view
---------

--keep-daily 14     --keep-monthly 6     --keep-yearly 1
----------------------------------------------------------------
 1. 2015-12-31       (2015-12-31 kept     (2015-12-31 kept
 2. 2015-12-30        by daily rule)       by daily rule)
 3. 2015-12-29       1. 2015-11-30        1. 2015-01-01 (oldest)
 4. 2015-12-28       2. 2015-10-31
 5. 2015-12-27       3. 2015-09-30
 6. 2015-12-26       4. 2015-08-31
 7. 2015-12-25       5. 2015-07-31
 8. 2015-12-24       6. 2015-06-30
 9. 2015-12-23
10. 2015-12-22
11. 2015-12-21
12. 2015-12-20
    (no backup made on 2015-12-19)
13. 2015-12-18
14. 2015-12-17


Notes
-----

2015-12-31 is kept due to the --keep-daily 14 rule (because it is applied
first), not due to the --keep-monthly or --keep-yearly rule.

The --keep-yearly 1 rule does not consider the December 31st backup because it
has already been kept due to the daily rule. There are no backups available
from previous years, so the --keep-yearly target of 1 backup is not satisfied.
Because of this, the 2015-01-01 archive (the oldest archive available) is kept.

The --keep-monthly 6 rule keeps Nov, Oct, Sep, Aug, Jul and Jun. December is
not considered for this rule, because that backup was already kept because of
the daily rule.

2015-12-17 is kept to satisfy the --keep-daily 14 rule - because no backup was
made on 2015-12-19. If a backup had been made on that day, it would not keep
the one from 2015-12-17.

We did not include weekly, hourly, minutely or secondly rules to keep this
example simple. They all work in basically the same way.

The weekly rule is easy to understand roughly, but hard to understand in all
details. If interested, read "ISO 8601:2000 standard week-based year".

borg compact

borg [common options] compact [options] [REPOSITORY]

positional arguments

REPOSITORY

repository to compact

options

--cleanup-commits

cleanup commit-only 17-byte segment files

--threshold PERCENT

set minimum threshold for saved space in PERCENT (Default: 10)

Common options

Description

This command frees repository space by compacting segments.

Use this regularly to avoid running out of space - you do not need to use this after each borg command though. It is especially useful after deleting archives, because only compaction will really free repository space.

borg compact does not need a key, so it is possible to invoke it from the client or also from the server.

Depending on the amount of segments that need compaction, it may take a while, so consider using the --progress option.

A segment is compacted if the amount of saved space is above the percentage value given by the --threshold option. If omitted, a threshold of 10% is used. When using --verbose, borg will output an estimate of the freed space.

After upgrading borg (server) to 1.2+, you can use borg compact --cleanup-commits to clean up the numerous 17byte commit-only segments that borg 1.1 did not clean up due to a bug. It is enough to do that once per repository. After cleaning up the commits, borg will also do a normal compaction.

See Separate compaction in Additional Notes for more details.

Examples
# compact segments and free repo disk space
$ borg compact /path/to/repo

# same as above plus clean up 17byte commit-only segments
$ borg compact --cleanup-commits /path/to/repo

borg info

borg [common options] info [options] [REPOSITORY_OR_ARCHIVE]

positional arguments

REPOSITORY_OR_ARCHIVE

repository or archive to display information about

options

--json

format output as JSON

Common options

Archive filters — Archive filters can be applied to repository targets.

-P PREFIX, --prefix PREFIX

only consider archive names starting with this prefix. (deprecated)

-a GLOB, --glob-archives GLOB

only consider archive names matching the glob. sh: rules apply (without actually using the sh: prefix), see "borg help patterns".

--sort-by KEYS

Comma-separated list of sorting keys; valid keys are: timestamp, archive, name, id; default is: timestamp

--first N

consider first N archives after other filters were applied

--last N

consider last N archives after other filters were applied

Description

This command displays detailed information about the specified archive or repository.

Please note that the deduplicated sizes of the individual archives do not add up to the deduplicated size of the repository ("all archives"), because the two are meaning different things:

This archive / deduplicated size = amount of data stored ONLY for this archive = unique chunks of this archive. All archives / deduplicated size = amount of data stored in the repo = all chunks in the repository.

Borg archives can only contain a limited amount of file metadata. The size of an archive relative to this limit depends on a number of factors, mainly the number of files, the lengths of paths and other metadata stored for files. This is shown as utilization of maximum supported archive size.

Examples
$ borg info /path/to/repo::2017-06-29T11:00-srv
Archive name: 2017-06-29T11:00-srv
Archive fingerprint: b2f1beac2bd553b34e06358afa45a3c1689320d39163890c5bbbd49125f00fe5
Comment:
Hostname: myhostname
Username: root
Time (start): Thu, 2017-06-29 11:03:07
Time (end): Thu, 2017-06-29 11:03:13
Duration: 5.66 seconds
Number of files: 17037
Command line: /usr/sbin/borg create /path/to/repo::2017-06-29T11:00-srv /srv
Utilization of max. archive size: 0%
------------------------------------------------------------------------------
                       Original size      Compressed size    Deduplicated size
This archive:               12.53 GB             12.49 GB              1.62 kB
All archives:              121.82 TB            112.41 TB            215.42 GB

                       Unique chunks         Total chunks
Chunk index:                 1015213            626934122

$ borg info /path/to/repo --last 1
Archive name: 2017-06-29T11:00-srv
Archive fingerprint: b2f1beac2bd553b34e06358afa45a3c1689320d39163890c5bbbd49125f00fe5
Comment:
Hostname: myhostname
Username: root
Time (start): Thu, 2017-06-29 11:03:07
Time (end): Thu, 2017-06-29 11:03:13
Duration: 5.66 seconds
Number of files: 17037
Command line: /usr/sbin/borg create /path/to/repo::2017-06-29T11:00-srv /srv
Utilization of max. archive size: 0%
------------------------------------------------------------------------------
                       Original size      Compressed size    Deduplicated size
This archive:               12.53 GB             12.49 GB              1.62 kB
All archives:              121.82 TB            112.41 TB            215.42 GB

                       Unique chunks         Total chunks
Chunk index:                 1015213            626934122

$ borg info /path/to/repo
Repository ID: d857ce5788c51272c61535062e89eac4e8ef5a884ffbe976e0af9d8765dedfa5
Location: /path/to/repo
Encrypted: Yes (repokey)
Cache: /root/.cache/borg/d857ce5788c51272c61535062e89eac4e8ef5a884ffbe976e0af9d8765dedfa5
Security dir: /root/.config/borg/security/d857ce5788c51272c61535062e89eac4e8ef5a884ffbe976e0af9d8765dedfa5
------------------------------------------------------------------------------
                       Original size      Compressed size    Deduplicated size
All archives:              121.82 TB            112.41 TB            215.42 GB

                       Unique chunks         Total chunks
Chunk index:                 1015213            626934122

borg mount

borg [common options] mount [options] REPOSITORY_OR_ARCHIVE MOUNTPOINT [PATH...]

positional arguments

REPOSITORY_OR_ARCHIVE

repository or archive to mount

MOUNTPOINT

where to mount filesystem

PATH

paths to extract; patterns are supported

options

--consider-checkpoints

Show checkpoint archives in the repository contents list (default: hidden).

-f, --foreground

stay in foreground, do not daemonize

-o

Extra mount options

--numeric-owner

deprecated, use --numeric-ids instead

--numeric-ids

use numeric user and group identifiers from archive(s)

Common options

Archive filters — Archive filters can be applied to repository targets.

-P PREFIX, --prefix PREFIX

only consider archive names starting with this prefix. (deprecated)

-a GLOB, --glob-archives GLOB

only consider archive names matching the glob. sh: rules apply (without actually using the sh: prefix), see "borg help patterns".

--sort-by KEYS

Comma-separated list of sorting keys; valid keys are: timestamp, archive, name, id; default is: timestamp

--first N

consider first N archives after other filters were applied

--last N

consider last N archives after other filters were applied

Include/Exclude options

-e PATTERN, --exclude PATTERN

exclude paths matching PATTERN

--exclude-from EXCLUDEFILE

read exclude patterns from EXCLUDEFILE, one per line

--pattern PATTERN

include/exclude paths matching PATTERN

--patterns-from PATTERNFILE

read include/exclude patterns from PATTERNFILE, one per line

--strip-components NUMBER

Remove the specified number of leading path elements. Paths with fewer elements will be silently skipped.

Description

This command mounts an archive as a FUSE filesystem. This can be useful for browsing an archive or restoring individual files. Unless the --foreground option is given the command will run in the background until the filesystem is umounted.

The command borgfs provides a wrapper for borg mount. This can also be used in fstab entries: /path/to/repo /mnt/point fuse.borgfs defaults,noauto 0 0

To allow a regular user to use fstab entries, add the user option: /path/to/repo /mnt/point fuse.borgfs defaults,noauto,user 0 0

For FUSE configuration and mount options, see the mount.fuse(8) manual page.

Borg's default behavior is to use the archived user and group names of each file and map them to the system's respective user and group ids. Alternatively, using numeric-ids will instead use the archived user and group ids without any mapping.

The uid and gid mount options (implemented by Borg) can be used to override the user and group ids of all files (i.e., borg mount -o uid=1000,gid=1000).

The man page references user_id and group_id mount options (implemented by fuse) which specify the user and group id of the mount owner (aka, the user who does the mounting). It is set automatically by libfuse (or the filesystem if libfuse is not used). However, you should not specify these manually. Unlike the uid and gid mount options which affect all files, user_id and group_id affect the user and group id of the mounted (base) directory.

Additional mount options supported by borg:

  • versions: when used with a repository mount, this gives a merged, versioned view of the files in the archives. EXPERIMENTAL, layout may change in future.

  • allow_damaged_files: by default damaged files (where missing chunks were replaced with runs of zeros by borg check --repair) are not readable and return EIO (I/O error). Set this option to read such files.

  • ignore_permissions: for security reasons the default_permissions mount option is internally enforced by borg. ignore_permissions can be given to not enforce default_permissions.

The BORG_MOUNT_DATA_CACHE_ENTRIES environment variable is meant for advanced users to tweak the performance. It sets the number of cached data chunks; additional memory usage can be up to ~8 MiB times this number. The default is the number of CPU cores.

When the daemonized process receives a signal or crashes, it does not unmount. Unmounting in these cases could cause an active rsync or similar process to unintentionally delete data.

When running in the foreground ^C/SIGINT unmounts cleanly, but other signals or crashes do not.

borg umount

borg [common options] umount [options] MOUNTPOINT

positional arguments

MOUNTPOINT

mountpoint of the filesystem to umount

Common options

Description

This command un-mounts a FUSE filesystem that was mounted with borg mount.

This is a convenience wrapper that just calls the platform-specific shell command - usually this is either umount or fusermount -u.

Examples
# Mounting the repository shows all archives.
# Archives are loaded lazily, expect some delay when navigating to an archive
# for the first time.
$ borg mount /path/to/repo /tmp/mymountpoint
$ ls /tmp/mymountpoint
root-2016-02-14 root-2016-02-15
$ borg umount /tmp/mymountpoint

# Mounting a specific archive is possible as well.
$ borg mount /path/to/repo::root-2016-02-15 /tmp/mymountpoint
$ ls /tmp/mymountpoint
bin  boot  etc      home  lib  lib64  lost+found  media  mnt  opt
root  sbin  srv  tmp  usr  var
$ borg umount /tmp/mymountpoint

# The "versions view" merges all archives in the repository
# and provides a versioned view on files.
$ borg mount -o versions /path/to/repo /tmp/mymountpoint
$ ls -l /tmp/mymountpoint/home/user/doc.txt/
total 24
-rw-rw-r-- 1 user group 12357 Aug 26 21:19 doc.cda00bc9.txt
-rw-rw-r-- 1 user group 12204 Aug 26 21:04 doc.fa760f28.txt
$ borg umount /tmp/mymountpoint

# Archive filters are supported.
# These are especially handy for the "versions view",
# which does not support lazy processing of archives.
$ borg mount -o versions --glob-archives '*-my-home' --last 10 /path/to/repo /tmp/mymountpoint

# Exclusion options are supported.
# These can speed up mounting and lower memory needs significantly.
$ borg mount /path/to/repo /tmp/mymountpoint only/that/path
$ borg mount --exclude '...' /path/to/repo /tmp/mymountpoint
borgfs
$ echo '/mnt/backup /tmp/myrepo fuse.borgfs defaults,noauto 0 0' >> /etc/fstab
$ echo '/mnt/backup::root-2016-02-15 /tmp/myarchive fuse.borgfs defaults,noauto 0 0' >> /etc/fstab
$ mount /tmp/myrepo
$ mount /tmp/myarchive
$ ls /tmp/myrepo
root-2016-02-01 root-2016-02-2015
$ ls /tmp/myarchive
bin  boot  etc      home  lib  lib64  lost+found  media  mnt  opt  root  sbin  srv  tmp  usr  var

Note

borgfs will be automatically provided if you used a distribution package, pip or setup.py to install Borg. Users of the standalone binary will have to manually create a symlink (see Standalone Binary).

borg key change-passphrase

borg [common options] key change-passphrase [options] [REPOSITORY]

positional arguments

REPOSITORY

Common options

Description

The key files used for repository encryption are optionally passphrase protected. This command can be used to change this passphrase.

Please note that this command only changes the passphrase, but not any secret protected by it (like e.g. encryption/MAC keys or chunker seed). Thus, changing the passphrase after passphrase and borg key got compromised does not protect future (nor past) backups to the same repository.

Examples
# Create a key file protected repository
$ borg init --encryption=keyfile -v /path/to/repo
Initializing repository at "/path/to/repo"
Enter new passphrase:
Enter same passphrase again:
Remember your passphrase. Your data will be inaccessible without it.
Key in "/root/.config/borg/keys/mnt_backup" created.
Keep this key safe. Your data will be inaccessible without it.
Synchronizing chunks cache...
Archives: 0, w/ cached Idx: 0, w/ outdated Idx: 0, w/o cached Idx: 0.
Done.

# Change key file passphrase
$ borg key change-passphrase -v /path/to/repo
Enter passphrase for key /root/.config/borg/keys/mnt_backup:
Enter new passphrase:
Enter same passphrase again:
Remember your passphrase. Your data will be inaccessible without it.
Key updated

# Import a previously-exported key into the specified
# key file (creating or overwriting the output key)
# (keyfile repositories only)
$ BORG_KEY_FILE=/path/to/output-key borg key import /path/to/repo /path/to/exported

Fully automated using environment variables:

$ BORG_NEW_PASSPHRASE=old borg init -e=repokey repo
# now "old" is the current passphrase.
$ BORG_PASSPHRASE=old BORG_NEW_PASSPHRASE=new borg key change-passphrase repo
# now "new" is the current passphrase.

borg key export

borg [common options] key export [options] [REPOSITORY] [PATH]

positional arguments

REPOSITORY

PATH

where to store the backup

options

--paper

Create an export suitable for printing and later type-in

--qr-html

Create an html file suitable for printing and later type-in or qr scan

Common options

Description

If repository encryption is used, the repository is inaccessible without the key. This command allows one to backup this essential key. Note that the backup produced does not include the passphrase itself (i.e. the exported key stays encrypted). In order to regain access to a repository, one needs both the exported key and the original passphrase.

There are three backup formats. The normal backup format is suitable for digital storage as a file. The --paper backup format is optimized for printing and typing in while importing, with per line checks to reduce problems with manual input. The --qr-html creates a printable HTML template with a QR code and a copy of the --paper-formatted key.

For repositories using keyfile encryption the key is saved locally on the system that is capable of doing backups. To guard against loss of this key, the key needs to be backed up independently of the main data backup.

For repositories using the repokey encryption the key is saved in the repository in the config file. A backup is thus not strictly needed, but guards against the repository becoming inaccessible if the file is damaged for some reason.

Examples:

borg key export /path/to/repo > encrypted-key-backup
borg key export --paper /path/to/repo > encrypted-key-backup.txt
borg key export --qr-html /path/to/repo > encrypted-key-backup.html
# Or pass the output file as an argument instead of redirecting stdout:
borg key export /path/to/repo encrypted-key-backup
borg key export --paper /path/to/repo encrypted-key-backup.txt
borg key export --qr-html /path/to/repo encrypted-key-backup.html

borg key import

borg [common options] key import [options] [REPOSITORY] [PATH]

positional arguments

REPOSITORY

PATH

path to the backup ('-' to read from stdin)

options

--paper

interactively import from a backup done with --paper

Common options

Description

This command restores a key previously backed up with the export command.

If the --paper option is given, the import will be an interactive process in which each line is checked for plausibility before proceeding to the next line. For this format PATH must not be given.

For repositories using keyfile encryption, the key file which borg key import writes to depends on several factors. If the BORG_KEY_FILE environment variable is set and non-empty, borg key import creates or overwrites that file named by $BORG_KEY_FILE. Otherwise, borg key import searches in the $BORG_KEYS_DIR directory for a key file associated with the repository. If a key file is found in $BORG_KEYS_DIR, borg key import overwrites it; otherwise, borg key import creates a new key file in $BORG_KEYS_DIR.

borg upgrade

borg [common options] upgrade [options] [REPOSITORY]

positional arguments

REPOSITORY

path to the repository to be upgraded

options

-n, --dry-run

do not change repository

--inplace

rewrite repository in place, with no chance of going back to older versions of the repository.

--force

Force upgrade

--tam

Enable manifest authentication (in key and cache) (Borg 1.0.9 and later).

--disable-tam

Disable manifest authentication (in key and cache).

--archives-tam

add TAM authentication for all archives.

Common options

Description

Upgrade an existing, local Borg repository.

When you do not need borg upgrade

Not every change requires that you run borg upgrade.

You do not need to run it when:

  • moving your repository to a different place

  • upgrading to another point release (like 1.0.x to 1.0.y), except when noted otherwise in the changelog

  • upgrading from 1.0.x to 1.1.x, except when noted otherwise in the changelog

Borg 1.x.y upgrades

Archive TAM authentication:

Use borg upgrade --archives-tam REPO to add archive TAMs to all archives that are not TAM authenticated yet. This is a convenient method to just trust all archives present - if an archive does not have TAM authentication yet, a TAM will be added. Archives created by old borg versions < 1.0.9 do not have TAMs. Archives created by newer borg version should have TAMs already. If you have a high risk environment, you should not just run this, but first verify that the archives are authentic and not malicious (== have good content, have a good timestamp). Borg 1.2.5+ needs all archives to be TAM authenticated for safety reasons.

This upgrade needs to be done once per repository.

Manifest TAM authentication:

Use borg upgrade --tam REPO to require manifest authentication introduced with Borg 1.0.9 to address security issues. This means that modifying the repository after doing this with a version prior to 1.0.9 will raise a validation error, so only perform this upgrade after updating all clients using the repository to 1.0.9 or newer.

This upgrade should be done on each client for safety reasons.

If a repository is accidentally modified with a pre-1.0.9 client after this upgrade, use borg upgrade --tam --force REPO to remedy it.

If you routinely do this you might not want to enable this upgrade (which will leave you exposed to the security issue). You can reverse the upgrade by issuing borg upgrade --disable-tam REPO.

See https://borgbackup.readthedocs.io/en/stable/changes.html#pre-1-0-9-manifest-spoofing-vulnerability for details.

Attic and Borg 0.xx to Borg 1.x

This currently supports converting an Attic repository to Borg and also helps with converting Borg 0.xx to 1.0.

Currently, only LOCAL repositories can be upgraded (issue #465).

Please note that borg create (since 1.0.0) uses bigger chunks by default than old borg or attic did, so the new chunks won't deduplicate with the old chunks in the upgraded repository. See --chunker-params option of borg create and borg recreate.

borg upgrade will change the magic strings in the repository's segments to match the new Borg magic strings. The keyfiles found in $ATTIC_KEYS_DIR or ~/.attic/keys/ will also be converted and copied to $BORG_KEYS_DIR or ~/.config/borg/keys.

The cache files are converted, from $ATTIC_CACHE_DIR or ~/.cache/attic to $BORG_CACHE_DIR or ~/.cache/borg, but the cache layout between Borg and Attic changed, so it is possible the first backup after the conversion takes longer than expected due to the cache resync.

Upgrade should be able to resume if interrupted, although it will still iterate over all segments. If you want to start from scratch, use borg delete over the copied repository to make sure the cache files are also removed:

borg delete borg

Unless --inplace is specified, the upgrade process first creates a backup copy of the repository, in REPOSITORY.before-upgrade-DATETIME, using hardlinks. This requires that the repository and its parent directory reside on same filesystem so the hardlink copy can work. This takes longer than in place upgrades, but is much safer and gives progress information (as opposed to cp -al). Once you are satisfied with the conversion, you can safely destroy the backup copy.

WARNING: Running the upgrade in place will make the current copy unusable with older version, with no way of going back to previous versions. This can PERMANENTLY DAMAGE YOUR REPOSITORY! Attic CAN NOT READ BORG REPOSITORIES, as the magic strings have changed. You have been warned.

Examples
# Upgrade the borg repository to the most recent version.
$ borg upgrade -v /path/to/repo
making a hardlink copy in /path/to/repo.before-upgrade-2016-02-15-20:51:55
opening attic repository with borg and converting
no key file found for repository
converting repo index /path/to/repo/index.0
converting 1 segments...
converting borg 0.xx to borg current
no key file found for repository
Upgrading a passphrase encrypted attic repo

attic offered a "passphrase" encryption mode, but this was removed in borg 1.0 and replaced by the "repokey" mode (which stores the passphrase-protected encryption key into the repository config).

Thus, to upgrade a "passphrase" attic repo to a "repokey" borg repo, 2 steps are needed, in this order:

  • borg upgrade repo

  • borg key migrate-to-repokey repo

borg recreate

borg [common options] recreate [options] [REPOSITORY_OR_ARCHIVE] [PATH...]

positional arguments

REPOSITORY_OR_ARCHIVE

repository or archive to recreate

PATH

paths to recreate; patterns are supported

options

--list

output verbose list of items (files, dirs, ...)

--filter STATUSCHARS

only display items with the given status characters (listed in borg create --help)

-n, --dry-run

do not change anything

-s, --stats

print statistics at end

Common options

Include/Exclude options

-e PATTERN, --exclude PATTERN

exclude paths matching PATTERN

--exclude-from EXCLUDEFILE

read exclude patterns from EXCLUDEFILE, one per line

--pattern PATTERN

include/exclude paths matching PATTERN

--patterns-from PATTERNFILE

read include/exclude patterns from PATTERNFILE, one per line

--exclude-caches

exclude directories that contain a CACHEDIR.TAG file (http://www.bford.info/cachedir/spec.html)

--exclude-if-present NAME

exclude directories that are tagged by containing a filesystem object with the given NAME

--keep-exclude-tags

if tag objects are specified with --exclude-if-present, don't omit the tag objects themselves from the backup archive

Archive options

--target TARGET

create a new archive with the name ARCHIVE, do not replace existing archive (only applies for a single archive)

-c SECONDS, --checkpoint-interval SECONDS

write checkpoint every SECONDS seconds (Default: 1800)

--comment COMMENT

add a comment text to the archive

--timestamp TIMESTAMP

manually specify the archive creation date/time (UTC, yyyy-mm-ddThh:mm:ss format). alternatively, give a reference file/directory.

-C COMPRESSION, --compression COMPRESSION

select compression algorithm, see the output of the "borg help compression" command for details.

--recompress MODE

recompress data chunks according to MODE and --compression. Possible modes are if-different: recompress if current compression is with a different compression algorithm (the level is not considered); always: recompress even if current compression is with the same compression algorithm (use this to change the compression level); and never: do not recompress (use this option to explicitly prevent recompression). If no MODE is given, if-different will be used. Not passing --recompress is equivalent to "--recompress never".

--chunker-params PARAMS

rechunk using given chunker parameters (ALGO, CHUNK_MIN_EXP, CHUNK_MAX_EXP, HASH_MASK_BITS, HASH_WINDOW_SIZE) or default to use the chunker defaults. default: do not rechunk

Description

Recreate the contents of existing archives.

recreate is a potentially dangerous function and might lead to data loss (if used wrongly). BE VERY CAREFUL!

Important: Repository disk space is not freed until you run borg compact.

--exclude, --exclude-from, --exclude-if-present, --keep-exclude-tags and PATH have the exact same semantics as in "borg create", but they only check for files in the archives and not in the local file system. If PATHs are specified, the resulting archives will only contain files from these PATHs.

Note that all paths in an archive are relative, therefore absolute patterns/paths will not match (--exclude, --exclude-from, PATHs).

--recompress allows one to change the compression of existing data in archives. Due to how Borg stores compressed size information this might display incorrect information for archives that were not recreated at the same time. There is no risk of data loss by this.

--chunker-params will re-chunk all files in the archive, this can be used to have upgraded Borg 0.xx or Attic archives deduplicate with Borg 1.x archives.

USE WITH CAUTION. Depending on the PATHs and patterns given, recreate can be used to permanently delete files from archives. When in doubt, use --dry-run --verbose --list to see how patterns/PATHS are interpreted. See Item flags in borg create for details.

The archive being recreated is only removed after the operation completes. The archive that is built during the operation exists at the same time at "<ARCHIVE>.recreate". The new archive will have a different archive ID.

With --target the original archive is not replaced, instead a new archive is created.

When rechunking (or recompressing), space usage can be substantial - expect at least the entire deduplicated size of the archives using the previous chunker (or compression) params.

If you recently ran borg check --repair and it had to fix lost chunks with all-zero replacement chunks, please first run another backup for the same data and re-run borg check --repair afterwards to heal any archives that had lost chunks which are still generated from the input data.

Important: running borg recreate to re-chunk will remove the chunks_healthy metadata of all items with replacement chunks, so healing will not be possible any more after re-chunking (it is also unlikely it would ever work: due to the change of chunking parameters, the missing chunk likely will never be seen again even if you still have the data that produced it).

Examples
# Make old (Attic / Borg 0.xx) archives deduplicate with Borg 1.x archives.
# Archives created with Borg 1.1+ and the default chunker params are skipped
# (archive ID stays the same).
$ borg recreate /mnt/backup --chunker-params default --progress

# Create a backup with little but fast compression
$ borg create /mnt/backup::archive /some/files --compression lz4
# Then compress it - this might take longer, but the backup has already completed,
# so no inconsistencies from a long-running backup job.
$ borg recreate /mnt/backup::archive --recompress --compression zlib,9

# Remove unwanted files from all archives in a repository.
# Note the relative path for the --exclude option - archives only contain relative paths.
$ borg recreate /mnt/backup --exclude home/icke/Pictures/drunk_photos

# Change archive comment
$ borg create --comment "This is a comment" /mnt/backup::archivename ~
$ borg info /mnt/backup::archivename
Name: archivename
Fingerprint: ...
Comment: This is a comment
...
$ borg recreate --comment "This is a better comment" /mnt/backup::archivename
$ borg info /mnt/backup::archivename
Name: archivename
Fingerprint: ...
Comment: This is a better comment
...

borg import-tar

borg [common options] import-tar [options] ARCHIVE TARFILE

positional arguments

ARCHIVE

name of archive to create (must be also a valid directory name)

TARFILE

input tar file. "-" to read from stdin instead.

options

--tar-filter

filter program to pipe data through

-s, --stats

print statistics for the created archive

--list

output verbose list of items (files, dirs, ...)

--filter STATUSCHARS

only display items with the given status characters

--json

output stats as JSON (implies --stats)

--ignore-zeros

ignore zero-filled blocks in the input tarball

Common options

Archive options

--comment COMMENT

add a comment text to the archive

--timestamp TIMESTAMP

manually specify the archive creation date/time (UTC, yyyy-mm-ddThh:mm:ss format). alternatively, give a reference file/directory.

-c SECONDS, --checkpoint-interval SECONDS

write checkpoint every SECONDS seconds (Default: 1800)

--chunker-params PARAMS

specify the chunker parameters (ALGO, CHUNK_MIN_EXP, CHUNK_MAX_EXP, HASH_MASK_BITS, HASH_WINDOW_SIZE). default: buzhash,19,23,21,4095

-C COMPRESSION, --compression COMPRESSION

select compression algorithm, see the output of the "borg help compression" command for details.

Description

This command creates a backup archive from a tarball.

When giving '-' as path, Borg will read a tar stream from standard input.

By default (--tar-filter=auto) Borg will detect whether the file is compressed based on its file extension and pipe the file through an appropriate filter:

  • .tar.gz or .tgz: gzip -d

  • .tar.bz2 or .tbz: bzip2 -d

  • .tar.xz or .txz: xz -d

  • .tar.zstd or .tar.zst: zstd -d

  • .tar.lz4: lz4 -d

Alternatively, a --tar-filter program may be explicitly specified. It should read compressed data from stdin and output an uncompressed tar stream on stdout.

Most documentation of borg create applies. Note that this command does not support excluding files.

import-tar is a lossy conversion: BSD flags, ACLs, extended attributes (xattrs), atime and ctime are not exported. Timestamp resolution is limited to whole seconds, not the nanosecond resolution otherwise supported by Borg.

A --sparse option (as found in borg create) is not supported.

import-tar reads POSIX.1-1988 (ustar), POSIX.1-2001 (pax), GNU tar, UNIX V7 tar and SunOS tar with extended attributes.

To import multiple tarballs into a single archive, they can be simply concatenated (e.g. using "cat") into a single file, and imported with an --ignore-zeros option to skip through the stop markers between them.

borg export-tar

borg [common options] export-tar [options] ARCHIVE FILE [PATH...]

positional arguments

ARCHIVE

archive to export

FILE

output tar file. "-" to write to stdout instead.

PATH

paths to extract; patterns are supported

options

--tar-filter

filter program to pipe data through

--list

output verbose list of items (files, dirs, ...)

Common options

Include/Exclude options

-e PATTERN, --exclude PATTERN

exclude paths matching PATTERN

--exclude-from EXCLUDEFILE

read exclude patterns from EXCLUDEFILE, one per line

--pattern PATTERN

include/exclude paths matching PATTERN

--patterns-from PATTERNFILE

read include/exclude patterns from PATTERNFILE, one per line

--strip-components NUMBER

Remove the specified number of leading path elements. Paths with fewer elements will be silently skipped.

Description

This command creates a tarball from an archive.

When giving '-' as the output FILE, Borg will write a tar stream to standard output.

By default (--tar-filter=auto) Borg will detect whether the FILE should be compressed based on its file extension and pipe the tarball through an appropriate filter before writing it to FILE:

  • .tar.gz or .tgz: gzip

  • .tar.bz2 or .tbz: bzip2

  • .tar.xz or .txz: xz

  • .tar.zstd or .tar.zst: zstd

  • .tar.lz4: lz4

Alternatively, a --tar-filter program may be explicitly specified. It should read the uncompressed tar stream from stdin and write a compressed/filtered tar stream to stdout.

The generated tarball uses the GNU tar format.

export-tar is a lossy conversion: BSD flags, ACLs, extended attributes (xattrs), atime and ctime are not exported. Timestamp resolution is limited to whole seconds, not the nanosecond resolution otherwise supported by Borg.

A --sparse option (as found in borg extract) is not supported.

By default the entire archive is extracted but a subset of files and directories can be selected by passing a list of PATHs as arguments. The file selection can further be restricted by using the --exclude option.

For more help on include/exclude patterns, see the borg help patterns command output.

--progress can be slower than no progress display, since it makes one additional pass over the archive metadata.

Examples
# export as uncompressed tar
$ borg export-tar /path/to/repo::Monday Monday.tar

# exclude some types, compress using gzip
$ borg export-tar /path/to/repo::Monday Monday.tar.gz --exclude '*.so'

# use higher compression level with gzip
$ borg export-tar --tar-filter="gzip -9" testrepo::linux Monday.tar.gz

# export a tar, but instead of storing it on disk,
# upload it to a remote site using curl.
$ borg export-tar /path/to/repo::Monday - | curl --data-binary @- https://somewhere/to/POST

# remote extraction via "tarpipe"
$ borg export-tar /path/to/repo::Monday - | ssh somewhere "cd extracted; tar x"

borg serve

borg [common options] serve [options]

options

--restrict-to-path PATH

restrict repository access to PATH. Can be specified multiple times to allow the client access to several directories. Access to all sub-directories is granted implicitly; PATH doesn't need to directly point to a repository.

--restrict-to-repository PATH

restrict repository access. Only the repository located at PATH (no sub-directories are considered) is accessible. Can be specified multiple times to allow the client access to several repositories. Unlike --restrict-to-path sub-directories are not accessible; PATH needs to directly point at a repository location. PATH may be an empty directory or the last element of PATH may not exist, in which case the client may initialize a repository there.

--append-only

only allow appending to repository segment files. Note that this only affects the low level structure of the repository, and running delete or prune will still be allowed. See Append-only mode (forbid compaction) in Additional Notes for more details.

--storage-quota QUOTA

Override storage quota of the repository (e.g. 5G, 1.5T). When a new repository is initialized, sets the storage quota on the new repository as well. Default: no quota.

Common options

Description

This command starts a repository server process. This command is usually not used manually.

Examples

borg serve has special support for ssh forced commands (see authorized_keys example below): if the environment variable SSH_ORIGINAL_COMMAND is set it will ignore some options given on the command line and use the values from the variable instead. This only applies to a carefully controlled allowlist of safe options. This list currently contains:

  • Options that control the log level and debug topics printed such as --verbose, --info, --debug, --debug-topic, etc.

  • --lock-wait to allow the client to control how long to wait before giving up and aborting the operation when another process is holding a lock.

Environment variables (such as BORG_XXX) contained in the original command sent by the client are not interpreted, but ignored. If BORG_XXX environment variables should be set on the borg serve side, then these must be set in system-specific locations like /etc/environment or in the forced command itself (example below).

# Allow an SSH keypair to only run borg, and only have access to /path/to/repo.
# Use key options to disable unneeded and potentially dangerous SSH functionality.
# This will help to secure an automated remote backup system.
$ cat ~/.ssh/authorized_keys
command="borg serve --restrict-to-path /path/to/repo",restrict ssh-rsa AAAAB3[...]

# Set a BORG_XXX environment variable on the "borg serve" side
$ cat ~/.ssh/authorized_keys
command="export BORG_XXX=value; borg serve [...]",restrict ssh-rsa [...]

Note

The examples above use the restrict directive. This does automatically block potential dangerous ssh features, even when they are added in a future update. Thus, this option should be preferred.

If you're using openssh-server < 7.2, however, you have to explicitly specify the ssh features to restrict and cannot simply use the restrict option as it has been introduced in v7.2. We recommend to use no-port-forwarding,no-X11-forwarding,no-pty,no-agent-forwarding,no-user-rc in this case.

Details about sshd usage: sshd(8)

SSH Configuration

borg serve's pipes (stdin/stdout/stderr) are connected to the sshd process on the server side. In the event that the SSH connection between borg serve and the client is disconnected or stuck abnormally (for example, due to a network outage), it can take a long time for sshd to notice the client is disconnected. In the meantime, sshd continues running, and as a result so does the borg serve process holding the lock on the repository. This can cause subsequent borg operations on the remote repository to fail with the error: Failed to create/acquire the lock.

In order to avoid this, it is recommended to perform the following additional SSH configuration:

Either in the client side's ~/.ssh/config file, or in the client's /etc/ssh/ssh_config file:

Host backupserver
        ServerAliveInterval 10
        ServerAliveCountMax 30

Replacing backupserver with the hostname, FQDN or IP address of the borg server.

This will cause the client to send a keepalive to the server every 10 seconds. If 30 consecutive keepalives are sent without a response (a time of 300 seconds), the ssh client process will be terminated, causing the borg process to terminate gracefully.

On the server side's sshd configuration file (typically /etc/ssh/sshd_config):

ClientAliveInterval 10
ClientAliveCountMax 30

This will cause the server to send a keep alive to the client every 10 seconds. If 30 consecutive keepalives are sent without a response (a time of 300 seconds), the server's sshd process will be terminated, causing the borg serve process to terminate gracefully and release the lock on the repository.

If you then run borg commands with --lock-wait 600, this gives sufficient time for the borg serve processes to terminate after the SSH connection is torn down after the 300 second wait for the keepalives to fail.

You may, of course, modify the timeout values demonstrated above to values that suit your environment and use case.

borg config

borg [common options] config [options] [REPOSITORY] [NAME] [VALUE]

positional arguments

REPOSITORY

repository to configure

NAME

name of config key

VALUE

new value for key

options

-c, --cache

get and set values from the repo cache

-d, --delete

delete the key from the config file

-l, --list

list the configuration of the repo

Common options

Description

This command gets and sets options in a local repository or cache config file. For security reasons, this command only works on local repositories.

To delete a config value entirely, use --delete. To list the values of the configuration file or the default values, use --list. To get and existing key, pass only the key name. To set a key, pass both the key name and the new value. Keys can be specified in the format "section.name" or simply "name"; the section will default to "repository" and "cache" for the repo and cache configs, respectively.

By default, borg config manipulates the repository config file. Using --cache edits the repository cache's config file instead.

Note

The repository & cache config files are some of the only directly manipulable parts of a repository that aren't versioned or backed up, so be careful when making changes!

Examples
# find cache directory
$ cd ~/.cache/borg/$(borg config /path/to/repo id)

# reserve some space
$ borg config /path/to/repo additional_free_space 2G

# make a repo append-only
$ borg config /path/to/repo append_only 1

borg with-lock

borg [common options] with-lock [options] REPOSITORY COMMAND [ARGS...]

positional arguments

REPOSITORY

repository to lock

COMMAND

command to run

ARGS

command arguments

Common options

Description

This command runs a user-specified command while the repository lock is held.

It will first try to acquire the lock (make sure that no other operation is running in the repo), then execute the given command as a subprocess and wait for its termination, release the lock and return the user command's return code as borg's return code.

Note

If you copy a repository with the lock held, the lock will be present in the copy. Thus, before using borg on the copy from a different host, you need to use "borg break-lock" on the copied repository, because Borg is cautious and does not automatically remove stale locks made by a different host.

borg break-lock

borg [common options] break-lock [options] [REPOSITORY]

positional arguments

REPOSITORY

repository for which to break the locks

Common options

Description

This command breaks the repository and cache locks. Please use carefully and only while no borg process (on any machine) is trying to access the Cache or the Repository.

borg benchmark crud

borg [common options] benchmark crud [options] REPOSITORY PATH

positional arguments

REPOSITORY

repository to use for benchmark (must exist)

PATH

path were to create benchmark input data

Common options

Description

This command benchmarks borg CRUD (create, read, update, delete) operations.

It creates input data below the given PATH and backups this data into the given REPO. The REPO must already exist (it could be a fresh empty repo or an existing repo, the command will create / read / update / delete some archives named borg-benchmark-crud* there.

Make sure you have free space there, you'll need about 1GB each (+ overhead).

If your repository is encrypted and borg needs a passphrase to unlock the key, use:

BORG_PASSPHRASE=mysecret borg benchmark crud REPO PATH

Measurements are done with different input file sizes and counts. The file contents are very artificial (either all zero or all random), thus the measurement results do not necessarily reflect performance with real data. Also, due to the kind of content used, no compression is used in these benchmarks.

C- == borg create (1st archive creation, no compression, do not use files cache)

C-Z- == all-zero files. full dedup, this is primarily measuring reader/chunker/hasher. C-R- == random files. no dedup, measuring throughput through all processing stages.

R- == borg extract (extract archive, dry-run, do everything, but do not write files to disk)

R-Z- == all zero files. Measuring heavily duplicated files. R-R- == random files. No duplication here, measuring throughput through all processing stages, except writing to disk.

U- == borg create (2nd archive creation of unchanged input files, measure files cache speed)

The throughput value is kind of virtual here, it does not actually read the file. U-Z- == needs to check the 2 all-zero chunks' existence in the repo. U-R- == needs to check existence of a lot of different chunks in the repo.

D- == borg delete archive (delete last remaining archive, measure deletion + compaction)

D-Z- == few chunks to delete / few segments to compact/remove. D-R- == many chunks to delete / many segments to compact/remove.

Please note that there might be quite some variance in these measurements. Try multiple measurements and having a otherwise idle machine (and network, if you use it).

Miscellaneous Help

borg help patterns

The path/filenames used as input for the pattern matching start from the currently active recursion root. You usually give the recursion root(s) when invoking borg and these can be either relative or absolute paths.

Starting with Borg 1.2, paths that are matched against patterns always appear relative. If you give /absolute/ as root, the paths going into the matcher will start with absolute/. If you give ../../relative as root, the paths will be normalized as relative/.

A directory exclusion pattern can end either with or without a slash ('/'). If it ends with a slash, such as some/path/, the directory will be included but not its content. If it does not end with a slash, such as some/path, both the directory and content will be excluded.

Borg supports different pattern styles. To define a non-default style for a specific pattern, prefix it with two characters followed by a colon ':' (i.e. fm:path/*, sh:path/**).

Fnmatch, selector fm:

This is the default style for --exclude and --exclude-from. These patterns use a variant of shell pattern syntax, with '*' matching any number of characters, '?' matching any single character, '[...]' matching any single character specified, including ranges, and '[!...]' matching any character not specified. For the purpose of these patterns, the path separator (backslash for Windows and '/' on other systems) is not treated specially. Wrap meta-characters in brackets for a literal match (i.e. [?] to match the literal character ?). For a path to match a pattern, the full path must match, or it must match from the start of the full path to just before a path separator. Except for the root path, paths will never end in the path separator when matching is attempted. Thus, if a given pattern ends in a path separator, a '*' is appended before matching is attempted. A leading path separator is always removed.

Shell-style patterns, selector sh:

This is the default style for --pattern and --patterns-from. Like fnmatch patterns these are similar to shell patterns. The difference is that the pattern may include **/ for matching zero or more directory levels, * for matching zero or more arbitrary characters with the exception of any path separator. A leading path separator is always removed.

Regular expressions, selector re:

Regular expressions similar to those found in Perl are supported. Unlike shell patterns regular expressions are not required to match the full path and any substring match is sufficient. It is strongly recommended to anchor patterns to the start ('^'), to the end ('$') or both. Path separators (backslash for Windows and '/' on other systems) in paths are always normalized to a forward slash ('/') before applying a pattern. The regular expression syntax is described in the Python documentation for the re module.

Path prefix, selector pp:

This pattern style is useful to match whole sub-directories. The pattern pp:root/somedir matches root/somedir and everything therein. A leading path separator is always removed.

Path full-match, selector pf:

This pattern style is (only) useful to match full paths. This is kind of a pseudo pattern as it can not have any variable or unspecified parts - the full path must be given. pf:root/file.ext matches root/file.ext only. A leading path separator is always removed.

Implementation note: this is implemented via very time-efficient O(1) hashtable lookups (this means you can have huge amounts of such patterns without impacting performance much). Due to that, this kind of pattern does not respect any context or order. If you use such a pattern to include a file, it will always be included (if the directory recursion encounters it). Other include/exclude patterns that would normally match will be ignored. Same logic applies for exclude.

Note

re:, sh: and fm: patterns are all implemented on top of the Python SRE engine. It is very easy to formulate patterns for each of these types which requires an inordinate amount of time to match paths. If untrusted users are able to supply patterns, ensure they cannot supply re: patterns. Further, ensure that sh: and fm: patterns only contain a handful of wildcards at most.

Exclusions can be passed via the command line option --exclude. When used from within a shell, the patterns should be quoted to protect them from expansion.

The --exclude-from option permits loading exclusion patterns from a text file with one pattern per line. Lines empty or starting with the number sign ('#') after removing whitespace on both ends are ignored. The optional style selector prefix is also supported for patterns loaded from a file. Due to whitespace removal, paths with whitespace at the beginning or end can only be excluded using regular expressions.

To test your exclusion patterns without performing an actual backup you can run borg create --list --dry-run ....

Examples:

# Exclude '/home/user/file.o' but not '/home/user/file.odt':
$ borg create -e '*.o' backup /

# Exclude '/home/user/junk' and '/home/user/subdir/junk' but
# not '/home/user/importantjunk' or '/etc/junk':
$ borg create -e 'home/*/junk' backup /

# Exclude the contents of '/home/user/cache' but not the directory itself:
$ borg create -e home/user/cache/ backup /

# The file '/home/user/cache/important' is *not* backed up:
$ borg create -e home/user/cache/ backup / /home/user/cache/important

# The contents of directories in '/home' are not backed up when their name
# ends in '.tmp'
$ borg create --exclude 're:^home/[^/]+\.tmp/' backup /

# Load exclusions from file
$ cat >exclude.txt <<EOF
# Comment line
home/*/junk
*.tmp
fm:aa:something/*
re:^home/[^/]+\.tmp/
sh:home/*/.thumbnails
# Example with spaces, no need to escape as it is processed by borg
some file with spaces.txt
EOF
$ borg create --exclude-from exclude.txt backup /

A more general and easier to use way to define filename matching patterns exists with the --pattern and --patterns-from options. Using these, you may specify the backup roots, default pattern styles and patterns for inclusion and exclusion.

Root path prefix R

A recursion root path starts with the prefix R, followed by a path (a plain path, not a file pattern). Use this prefix to have the root paths in the patterns file rather than as command line arguments.

Pattern style prefix P

To change the default pattern style, use the P prefix, followed by the pattern style abbreviation (fm, pf, pp, re, sh). All patterns following this line will use this style until another style is specified.

Exclude pattern prefix -

Use the prefix -, followed by a pattern, to define an exclusion. This has the same effect as the --exclude option.

Exclude no-recurse pattern prefix !

Use the prefix !, followed by a pattern, to define an exclusion that does not recurse into subdirectories. This saves time, but prevents include patterns to match any files in subdirectories.

Include pattern prefix +

Use the prefix +, followed by a pattern, to define inclusions. This is useful to include paths that are covered in an exclude pattern and would otherwise not be backed up.

Note

Via --pattern or --patterns-from you can define BOTH inclusion and exclusion of files using pattern prefixes + and -. With --exclude and --exclude-from ONLY excludes are defined.

The first matching pattern is used, so if an include pattern matches before an exclude pattern, the file is backed up. Note that a no-recurse exclude stops examination of subdirectories so that potential includes will not match - use normal excludes for such use cases.

Example:

# Define the recursion root
R /
# Exclude all iso files in any directory
- **/*.iso
# Explicitly include all inside etc and root
+ etc/**
+ root/**
# Exclude a specific directory under each user's home directories
- home/*/.cache
# Explicitly include everything in /home
+ home/**
# Explicitly exclude some directories without recursing into them
! re:^(dev|proc|run|sys|tmp)
# Exclude all other files and directories
# that are not specifically included earlier.
- **

Note

It's possible that a sub-directory/file is matched while parent directories are not. In that case, parent directories are not backed up thus their user, group, permission, etc. can not be restored.

Note that the default pattern style for --pattern and --patterns-from is shell style (sh:), so those patterns behave similar to rsync include/exclude patterns. The pattern style can be set via the P prefix.

Patterns (--pattern) and excludes (--exclude) from the command line are considered first (in the order of appearance). Then patterns from --patterns-from are added. Exclusion patterns from --exclude-from files are appended last.

Examples:

# backup pics, but not the ones from 2018, except the good ones:
# note: using = is essential to avoid cmdline argument parsing issues.
borg create --pattern=+pics/2018/good --pattern=-pics/2018 repo::arch pics

# use a file with patterns:
borg create --patterns-from patterns.lst repo::arch

The patterns.lst file could look like that:

# "sh:" pattern style is the default, so the following line is not needed:
P sh
R /
# can be rebuild
- home/*/.cache
# they're downloads for a reason
- home/*/Downloads
# susan is a nice person
# include susans home
+ home/susan
# also back up this exact file
+ pf:home/bobby/specialfile.txt
# don't backup the other home directories
- home/*
# don't even look in /proc
! proc

You can specify recursion roots either on the command line or in a patternfile:

# these two commands do the same thing
borg create --exclude home/bobby/junk repo::arch /home/bobby /home/susan
borg create --patterns-from patternfile.lst repo::arch

The patternfile:

# note that excludes use fm: by default and patternfiles use sh: by default.
# therefore, we need to specify fm: to have the same exact behavior.
P fm
R /home/bobby
R /home/susan

- home/bobby/junk

This allows you to share the same patterns between multiple repositories without needing to specify them on the command line.

borg help placeholders

Repository (or Archive) URLs, --prefix, --glob-archives, --comment and --remote-path values support these placeholders:

{hostname}

The (short) hostname of the machine.

{fqdn}

The full name of the machine.

{reverse-fqdn}

The full name of the machine in reverse domain name notation.

{now}

The current local date and time, by default in ISO-8601 format. You can also supply your own format string, e.g. {now:%Y-%m-%d_%H:%M:%S}

{utcnow}

The current UTC date and time, by default in ISO-8601 format. You can also supply your own format string, e.g. {utcnow:%Y-%m-%d_%H:%M:%S}

{user}

The user name (or UID, if no name is available) of the user running borg.

{pid}

The current process ID.

{borgversion}

The version of borg, e.g.: 1.0.8rc1

{borgmajor}

The version of borg, only the major version, e.g.: 1

{borgminor}

The version of borg, only major and minor version, e.g.: 1.0

{borgpatch}

The version of borg, only major, minor and patch version, e.g.: 1.0.8

If literal curly braces need to be used, double them for escaping:

borg create /path/to/repo::{{literal_text}}

Examples:

borg create /path/to/repo::{hostname}-{user}-{utcnow} ...
borg create /path/to/repo::{hostname}-{now:%Y-%m-%d_%H:%M:%S} ...
borg prune --glob-archives '{hostname}-*' ...

Note

systemd uses a difficult, non-standard syntax for command lines in unit files (refer to the systemd.unit(5) manual page).

When invoking borg from unit files, pay particular attention to escaping, especially when using the now/utcnow placeholders, since systemd performs its own %-based variable replacement even in quoted text. To avoid interference from systemd, double all percent signs ({hostname}-{now:%Y-%m-%d_%H:%M:%S} becomes {hostname}-{now:%%Y-%%m-%%d_%%H:%%M:%%S}).

borg help compression

It is no problem to mix different compression methods in one repo, deduplication is done on the source data chunks (not on the compressed or encrypted data).

If some specific chunk was once compressed and stored into the repo, creating another backup that also uses this chunk will not change the stored chunk. So if you use different compression specs for the backups, whichever stores a chunk first determines its compression. See also borg recreate.

Compression is lz4 by default. If you want something else, you have to specify what you want.

Valid compression specifiers are:

none

Do not compress.

lz4

Use lz4 compression. Very high speed, very low compression. (default)

zstd[,L]

Use zstd ("zstandard") compression, a modern wide-range algorithm. If you do not explicitly give the compression level L (ranging from 1 to 22), it will use level 3. Archives compressed with zstd are not compatible with borg < 1.1.4.

zlib[,L]

Use zlib ("gz") compression. Medium speed, medium compression. If you do not explicitly give the compression level L (ranging from 0 to 9), it will use level 6. Giving level 0 (means "no compression", but still has zlib protocol overhead) is usually pointless, you better use "none" compression.

lzma[,L]

Use lzma ("xz") compression. Low speed, high compression. If you do not explicitly give the compression level L (ranging from 0 to 9), it will use level 6. Giving levels above 6 is pointless and counterproductive because it does not compress better due to the buffer size used by borg - but it wastes lots of CPU cycles and RAM.

auto,C[,L]

Use a built-in heuristic to decide per chunk whether to compress or not. The heuristic tries with lz4 whether the data is compressible. For incompressible data, it will not use compression (uses "none"). For compressible data, it uses the given C[,L] compression - with C[,L] being any valid compression specifier.

obfuscate,SPEC,C[,L]

Use compressed-size obfuscation to make fingerprinting attacks based on the observable stored chunk size more difficult. Note:

  • You must combine this with encryption, or it won't make any sense.

  • Your repo size will be bigger, of course.

  • A chunk is limited by the constant MAX_DATA_SIZE (cur. ~20MiB).

The SPEC value determines how the size obfuscation works:

Relative random reciprocal size variation (multiplicative)

Size will increase by a factor, relative to the compressed data size. Smaller factors are used often, larger factors rarely.

Available factors:

1:     0.01 ..        100
2:     0.1  ..      1,000
3:     1    ..     10,000
4:    10    ..    100,000
5:   100    ..  1,000,000
6: 1,000    .. 10,000,000

Example probabilities for SPEC 1:

90   %  0.01 ..   0.1
 9   %  0.1  ..   1
 0.9 %  1    ..  10
 0.09% 10    .. 100

Randomly sized padding up to the given size (additive)

110: 1kiB (2 ^ (SPEC - 100))
...
120: 1MiB
...
123: 8MiB (max.)

Examples:

borg create --compression lz4 REPO::ARCHIVE data
borg create --compression zstd REPO::ARCHIVE data
borg create --compression zstd,10 REPO::ARCHIVE data
borg create --compression zlib REPO::ARCHIVE data
borg create --compression zlib,1 REPO::ARCHIVE data
borg create --compression auto,lzma,6 REPO::ARCHIVE data
borg create --compression auto,lzma ...
borg create --compression obfuscate,110,none ...
borg create --compression obfuscate,3,auto,zstd,10 ...
borg create --compression obfuscate,2,zstd,6 ...

Debugging Facilities

There is a borg debug command that has some subcommands which are all not intended for normal use and potentially very dangerous if used incorrectly.

For example, borg debug put-obj and borg debug delete-obj will only do what their name suggests: put objects into repo / delete objects from repo.

Please note:

  • they will not update the chunks cache (chunks index) about the object

  • they will not update the manifest (so no automatic chunks index resync is triggered)

  • they will not check whether the object is in use (e.g. before delete-obj)

  • they will not update any metadata which may point to the object

They exist to improve debugging capabilities without direct system access, e.g. in case you ever run into some severe malfunction. Use them only if you know what you are doing or if a trusted Borg developer tells you what to do.

Borg has a --debug-topic TOPIC option to enable specific debugging messages. Topics are generally not documented.

A --debug-profile FILE option exists which writes a profile of the main program's execution to a file. The format of these files is not directly compatible with the Python profiling tools, since these use the "marshal" format, which is not intended to be secure (quoting the Python docs: "Never unmarshal data received from an untrusted or unauthenticated source.").

The borg debug profile-convert command can be used to take a Borg profile and convert it to a profile file that is compatible with the Python tools.

Additionally, if the filename specified for --debug-profile ends with ".pyprof" a Python compatible profile is generated. This is only intended for local use by developers.

Additional Notes

Here are misc. notes about topics that are maybe not covered in enough detail in the usage section.

--chunker-params

The chunker params influence how input files are cut into pieces (chunks) which are then considered for deduplication. They also have a big impact on resource usage (RAM and disk space) as the amount of resources needed is (also) determined by the total amount of chunks in the repository (see Indexes / Caches memory usage for details).

--chunker-params=buzhash,10,23,16,4095 results in a fine-grained deduplication| and creates a big amount of chunks and thus uses a lot of resources to manage them. This is good for relatively small data volumes and if the machine has a good amount of free RAM and disk space.

--chunker-params=buzhash,19,23,21,4095 (default) results in a coarse-grained deduplication and creates a much smaller amount of chunks and thus uses less resources. This is good for relatively big data volumes and if the machine has a relatively low amount of free RAM and disk space.

--chunker-params=fixed,4194304 results in fixed 4MiB sized block deduplication and is more efficient than the previous example when used for for block devices (like disks, partitions, LVM LVs) or raw disk image files.

--chunker-params=fixed,4096,512 results in fixed 4kiB sized blocks, but the first header block will only be 512B long. This might be useful to dedup files with 1 header + N fixed size data blocks. Be careful to not produce a too big amount of chunks (like using small block size for huge files).

If you already have made some archives in a repository and you then change chunker params, this of course impacts deduplication as the chunks will be cut differently.

In the worst case (all files are big and were touched in between backups), this will store all content into the repository again.

Usually, it is not that bad though:

  • usually most files are not touched, so it will just re-use the old chunks it already has in the repo

  • files smaller than the (both old and new) minimum chunksize result in only one chunk anyway, so the resulting chunks are same and deduplication will apply

If you switch chunker params to save resources for an existing repo that already has some backup archives, you will see an increasing effect over time, when more and more files have been touched and stored again using the bigger chunksize and all references to the smaller older chunks have been removed (by deleting / pruning archives).

If you want to see an immediate big effect on resource usage, you better start a new repository when changing chunker params.

For more details, see Chunks.

--noatime / --noctime

You can use these borg create options to not store the respective timestamp into the archive, in case you do not really need it.

Besides saving a little space for the not archived timestamp, it might also affect metadata stream deduplication: if only this timestamp changes between backups and is stored into the metadata stream, the metadata stream chunks won't deduplicate just because of that.

--nobsdflags / --noflags

You can use this to not query and store (or not extract and set) flags - in case you don't need them or if they are broken somehow for your fs.

On Linux, dealing with the flags needs some additional syscalls. Especially when dealing with lots of small files, this causes a noticeable overhead, so you can use this option also for speeding up operations.

--umask

borg uses a safe default umask of 077 (that means the files borg creates have only permissions for owner, but no permissions for group and others) - so there should rarely be a need to change the default behaviour.

This option only affects the process to which it is given. Thus, when you run borg in client/server mode and you want to change the behaviour on the server side, you need to use borg serve --umask=XXX ... as a ssh forced command in authorized_keys. The --umask value given on the client side is not transferred to the server side.

Also, if you choose to use the --umask option, always be consistent and use the same umask value so you do not create a mixup of permissions in a borg repository or with other files borg creates.

--read-special

The --read-special option is special - you do not want to use it for normal full-filesystem backups, but rather after carefully picking some targets for it.

The option --read-special triggers special treatment for block and char device files as well as FIFOs. Instead of storing them as such a device (or FIFO), they will get opened, their content will be read and in the backup archive they will show up like a regular file.

Symlinks will also get special treatment if (and only if) they point to such a special file: instead of storing them as a symlink, the target special file will get processed as described above.

One intended use case of this is backing up the contents of one or multiple block devices, like e.g. LVM snapshots or inactive LVs or disk partitions.

You need to be careful about what you include when using --read-special, e.g. if you include /dev/zero, your backup will never terminate.

Restoring such files' content is currently only supported one at a time via --stdout option (and you have to redirect stdout to where ever it shall go, maybe directly into an existing device file of your choice or indirectly via dd).

To some extent, mounting a backup archive with the backups of special files via borg mount and then loop-mounting the image files from inside the mount point will work. If you plan to access a lot of data in there, it likely will scale and perform better if you do not work via the FUSE mount.

Example

Imagine you have made some snapshots of logical volumes (LVs) you want to backup.

Note

For some scenarios, this is a good method to get "crash-like" consistency (I call it crash-like because it is the same as you would get if you just hit the reset button or your machine would abruptly and completely crash). This is better than no consistency at all and a good method for some use cases, but likely not good enough if you have databases running.

Then you create a backup archive of all these snapshots. The backup process will see a "frozen" state of the logical volumes, while the processes working in the original volumes continue changing the data stored there.

You also add the output of lvdisplay to your backup, so you can see the LV sizes in case you ever need to recreate and restore them.

After the backup has completed, you remove the snapshots again.

$ # create snapshots here
$ lvdisplay > lvdisplay.txt
$ borg create --read-special /path/to/repo::arch lvdisplay.txt /dev/vg0/*-snapshot
$ # remove snapshots here

Now, let's see how to restore some LVs from such a backup.

$ borg extract /path/to/repo::arch lvdisplay.txt
$ # create empty LVs with correct sizes here (look into lvdisplay.txt).
$ # we assume that you created an empty root and home LV and overwrite it now:
$ borg extract --stdout /path/to/repo::arch dev/vg0/root-snapshot > /dev/vg0/root
$ borg extract --stdout /path/to/repo::arch dev/vg0/home-snapshot > /dev/vg0/home
Separate compaction

Borg does not auto-compact the segment files in the repository at commit time (at the end of each repository-writing command) any more.

This is new since borg 1.2.0 and requires borg >= 1.2.0 on client and server.

This causes a similar behaviour of the repository as if it was in append-only mode (see below) most of the time (until borg compact is invoked or an old client triggers auto-compaction).

This has some notable consequences:

  • repository space is not freed immediately when deleting / pruning archives

  • commands finish quicker

  • repository is more robust and might be easier to recover after damages (as it contains data in a more sequential manner, historic manifests, multiple commits - until you run borg compact)

  • user can choose when to run compaction (it should be done regularly, but not necessarily after each single borg command)

  • user can choose from where to invoke borg compact to do the compaction (from client or from server, it does not need a key)

  • less repo sync data traffic in case you create a copy of your repository by using a sync tool (like rsync, rclone, ...)

You can manually run compaction by invoking the borg compact command.

Append-only mode (forbid compaction)

A repository can be made "append-only", which means that Borg will never overwrite or delete committed data (append-only refers to the segment files, but borg will also reject to delete the repository completely).

If borg compact command is used on a repo in append-only mode, there will be no warning or error, but no compaction will happen.

append-only is useful for scenarios where a backup client machine backups remotely to a backup server using borg serve, since a hacked client machine cannot delete backups on the server permanently.

To activate append-only mode, set append_only to 1 in the repository config:

borg config /path/to/repo append_only 1

Note that you can go back-and-forth between normal and append-only operation with borg config; it's not a "one way trip."

In append-only mode Borg will create a transaction log in the transactions file, where each line is a transaction and a UTC timestamp.

In addition, borg serve can act as if a repository is in append-only mode with its option --append-only. This can be very useful for fine-tuning access control in .ssh/authorized_keys:

command="borg serve --append-only ..." ssh-rsa <key used for not-always-trustable backup clients>
command="borg serve ..." ssh-rsa <key used for backup management>

Running borg init via a borg serve --append-only server will not create an append-only repository. Running borg init --append-only creates an append-only repository regardless of server settings.

Example

Suppose an attacker remotely deleted all backups, but your repository was in append-only mode. A transaction log in this situation might look like this:

transaction 1, UTC time 2016-03-31T15:53:27.383532
transaction 5, UTC time 2016-03-31T15:53:52.588922
transaction 11, UTC time 2016-03-31T15:54:23.887256
transaction 12, UTC time 2016-03-31T15:55:54.022540
transaction 13, UTC time 2016-03-31T15:55:55.472564

From your security logs you conclude the attacker gained access at 15:54:00 and all the backups where deleted or replaced by compromised backups. From the log you know that transactions 11 and later are compromised. Note that the transaction ID is the name of the last file in the transaction. For example, transaction 11 spans files 6 to 11.

In a real attack you'll likely want to keep the compromised repository intact to analyze what the attacker tried to achieve. It's also a good idea to make this copy just in case something goes wrong during the recovery. Since recovery is done by deleting some files, a hard link copy (cp -al) is sufficient.

The first step to reset the repository to transaction 5, the last uncompromised transaction, is to remove the hints.N, index.N and integrity.N files in the repository (these files are always expendable). In this example N is 13.

Then remove or move all segment files from the segment directories in data/ starting with file 6:

rm data/**/{6..13}

That's all to do in the repository.

If you want to access this rolled back repository from a client that already has a cache for this repository, the cache will reflect a newer repository state than what you actually have in the repository now, after the rollback.

Thus, you need to clear the cache:

borg delete --cache-only repo

The cache will get rebuilt automatically. Depending on repo size and archive count, it may take a while.

You also will need to remove ~/.config/borg/security/REPOID/manifest-timestamp.

Drawbacks

As data is only appended, and nothing removed, commands like prune or delete won't free disk space, they merely tag data as deleted in a new transaction.

Be aware that as soon as you write to the repo in non-append-only mode (e.g. prune, delete or create archives from an admin machine), it will remove the deleted objects permanently (including the ones that were already marked as deleted, but not removed, in append-only mode). Automated edits to the repository (such as a cron job running borg prune) will render append-only mode moot if data is deleted.

Even if an archive appears to be available, it is possible an attacker could delete just a few chunks from an archive and silently corrupt its data. While in append-only mode, this is reversible, but borg check should be run before a writing/pruning operation on an append-only repository to catch accidental or malicious corruption:

# run without append-only mode
borg check --verify-data repo && borg compact repo

Aside from checking repository & archive integrity you may want to also manually check backups to ensure their content seems correct.

Further considerations

Append-only mode is not respected by tools other than Borg. rm still works on the repository. Make sure that backup client machines only get to access the repository via borg serve.

Ensure that no remote access is possible if the repository is temporarily set to normal mode for e.g. regular pruning.

Further protections can be implemented, but are outside of Borg's scope. For example, file system snapshots or wrapping borg serve to set special permissions or ACLs on new data files.

SSH batch mode

When running Borg using an automated script, ssh might still ask for a password, even if there is an SSH key for the target server. Use this to make scripts more robust:

export BORG_RSH='ssh -oBatchMode=yes'

Deployment

This chapter details deployment strategies for the following scenarios.

Central repository server with Ansible or Salt

This section will give an example how to setup a borg repository server for multiple clients.

Machines

There are multiple machines used in this section and will further be named by their respective fully qualified domain name (fqdn).

  • The backup server: backup01.srv.local

  • The clients:

    • John Doe's desktop: johndoe.clnt.local

    • Webserver 01: web01.srv.local

    • Application server 01: app01.srv.local

User and group

The repository server needs to have only one UNIX user for all the clients. Recommended user and group with additional settings:

  • User: backup

  • Group: backup

  • Shell: /bin/bash (or other capable to run the borg serve command)

  • Home: /home/backup

Most clients shall initiate a backup from the root user to catch all users, groups and permissions (e.g. when backing up /home).

Folders

The following folder tree layout is suggested on the repository server:

  • User home directory, /home/backup

  • Repositories path (storage pool): /home/backup/repos

  • Clients restricted paths (/home/backup/repos/<client fqdn>):

    • johndoe.clnt.local: /home/backup/repos/johndoe.clnt.local

    • web01.srv.local: /home/backup/repos/web01.srv.local

    • app01.srv.local: /home/backup/repos/app01.srv.local

Restrictions

Borg is instructed to restrict clients into their own paths: borg serve --restrict-to-path /home/backup/repos/<client fqdn>

The client will be able to access any file or subdirectory inside of /home/backup/repos/<client fqdn> but no other directories. You can allow a client to access several separate directories by passing multiple --restrict-to-path flags, for instance: borg serve --restrict-to-path /home/backup/repos/<client fqdn> --restrict-to-path /home/backup/repos/<other client fqdn>, which could make sense if multiple machines belong to one person which should then have access to all the backups of their machines.

There is only one ssh key per client allowed. Keys are added for johndoe.clnt.local, web01.srv.local and app01.srv.local. But they will access the backup under only one UNIX user account as: backup@backup01.srv.local. Every key in $HOME/.ssh/authorized_keys has a forced command and restrictions applied as shown below:

command="cd /home/backup/repos/<client fqdn>;
         borg serve --restrict-to-path /home/backup/repos/<client fqdn>",
         restrict <keytype> <key> <host>

Note

The text shown above needs to be written on a single line!

The options which are added to the key will perform the following:

  1. Change working directory

  2. Run borg serve restricted to the client base path

  3. Restrict ssh and do not allow stuff which imposes a security risk

Due to the cd command we use, the server automatically changes the current working directory. Then client doesn't need to have knowledge of the absolute or relative remote repository path and can directly access the repositories at <user>@<host>:<repo>.

Note

The setup above ignores all client given commandline parameters which are normally appended to the borg serve command.

Client

The client needs to initialize the pictures repository like this:

borg init backup@backup01.srv.local:pictures

Or with the full path (should actually never be used, as only for demonstrational purposes). The server should automatically change the current working directory to the <client fqdn> folder.

borg init backup@backup01.srv.local:/home/backup/repos/johndoe.clnt.local/pictures

When johndoe.clnt.local tries to access a not restricted path the following error is raised. John Doe tries to backup into the Web 01 path:

borg init backup@backup01.srv.local:/home/backup/repos/web01.srv.local/pictures
~~~ SNIP ~~~
Remote: borg.remote.PathNotAllowed: /home/backup/repos/web01.srv.local/pictures
~~~ SNIP ~~~
Repository path not allowed
Ansible

Ansible takes care of all the system-specific commands to add the user, create the folder, install and configure software.

- hosts: backup01.srv.local
  vars:
    user: backup
    group: backup
    home: /home/backup
    pool: "{{ home }}/repos"
    auth_users:
      - host: johndoe.clnt.local
        key: "{{ lookup('file', '/path/to/keys/johndoe.clnt.local.pub') }}"
      - host: web01.clnt.local
        key: "{{ lookup('file', '/path/to/keys/web01.clnt.local.pub') }}"
      - host: app01.clnt.local
        key: "{{ lookup('file', '/path/to/keys/app01.clnt.local.pub') }}"
  tasks:
  - package: name=borg state=present
  - group: name="{{ group }}" state=present
  - user: name="{{ user }}" shell=/bin/bash home="{{ home }}" createhome=yes group="{{ group }}" groups= state=present
  - file: path="{{ home }}" owner="{{ user }}" group="{{ group }}" mode=0700 state=directory
  - file: path="{{ home }}/.ssh" owner="{{ user }}" group="{{ group }}" mode=0700 state=directory
  - file: path="{{ pool }}" owner="{{ user }}" group="{{ group }}" mode=0700 state=directory
  - authorized_key: user="{{ user }}"
                    key="{{ item.key }}"
                    key_options='command="cd {{ pool }}/{{ item.host }};borg serve --restrict-to-path {{ pool }}/{{ item.host }}",restrict'
    with_items: "{{ auth_users }}"
  - file: path="{{ home }}/.ssh/authorized_keys" owner="{{ user }}" group="{{ group }}" mode=0600 state=file
  - file: path="{{ pool }}/{{ item.host }}" owner="{{ user }}" group="{{ group }}" mode=0700 state=directory
    with_items: "{{ auth_users }}"
Salt

This is a configuration similar to the one above, configured to be deployed with Salt running on a Debian system.

Install borg backup from pip:
  pkg.installed:
    - pkgs:
      - python3
      - python3-dev
      - python3-pip
      - python-virtualenv
      - libssl-dev
      - openssl
      - libacl1-dev
      - libacl1
      - build-essential
      - libfuse-dev
      - fuse
      - pkg-config
  pip.installed:
    - pkgs: ["borgbackup"]
    - bin_env: /usr/bin/pip3

Setup backup user:
  user.present:
    - name: backup
    - fullname: Backup User
    - home: /home/backup
    - shell: /bin/bash
# CAUTION!
# If you change the ssh command= option below, it won't necessarily get pushed to the backup
# server correctly unless you delete the ~/.ssh/authorized_keys file and re-create it!
{% for host in backupclients %}
Give backup access to {{host}}:
  ssh_auth.present:
    - user: backup
    - source: salt://conf/ssh-pubkeys/{{host}}-backup.id_ecdsa.pub
    - options:
      - command="cd /home/backup/repos/{{host}}; borg serve --restrict-to-path /home/backup/repos/{{host}}"
      - restrict
{% endfor %}
Enhancements

As this section only describes a simple and effective setup it could be further enhanced when supporting (a limited set) of client supplied commands. A wrapper for starting borg serve could be written. Or borg itself could be enhanced to autodetect it runs under SSH by checking the SSH_ORIGINAL_COMMAND environment variable. This is left open for future improvements.

When extending ssh autodetection in borg no external wrapper script is necessary and no other interpreter or application has to be deployed.

See also

Hosting repositories

This sections shows how to securely provide repository storage for users. Optionally, each user can have a storage quota.

Repositories are accessed through SSH. Each user of the service should have her own login which is only able to access the user's files. Technically it would be possible to have multiple users share one login, however, separating them is better. Separate logins increase isolation and are thus an additional layer of security and safety for both the provider and the users.

For example, if a user manages to breach borg serve then she can only damage her own data (assuming that the system does not have further vulnerabilities).

Use the standard directory structure of the operating system. Each user is assigned a home directory and repositories of the user reside in her home directory.

The following ~user/.ssh/authorized_keys file is the most important piece for a correct deployment. It allows the user to login via their public key (which must be provided by the user), and restricts SSH access to safe operations only.

command="borg serve --restrict-to-repository /home/<user>/repository",restrict
<key type> <key> <key host>

Note

The text shown above needs to be written on a single line!

Warning

If this file should be automatically updated (e.g. by a web console), pay utmost attention to sanitizing user input. Strip all whitespace around the user-supplied key, ensure that it only contains ASCII with no control characters and that it consists of three parts separated by a single space. Ensure that no newlines are contained within the key.

The restrict keyword enables all restrictions, i.e. disables port, agent and X11 forwarding, as well as disabling PTY allocation and execution of ~/.ssh/rc. If any future restriction capabilities are added to authorized_keys files they will be included in this set.

The command keyword forces execution of the specified command line upon login. This must be borg serve. The --restrict-to-repository option permits access to exactly one repository. It can be given multiple times to permit access to more than one repository.

The repository may not exist yet; it can be initialized by the user, which allows for encryption.

Storage quotas can be enabled by adding the --storage-quota option to the borg serve command line:

restrict,command="borg serve --storage-quota 20G ..." ...

The storage quotas of repositories are completely independent. If a client is able to access multiple repositories, each repository can be filled to the specified quota.

If storage quotas are used, ensure that all deployed Borg releases support storage quotas.

Refer to Storage quotas for more details on storage quotas.

Specificities: Append-only repositories

Running borg init via a borg serve --append-only server will not create a repository that is configured to be append-only by its repository config.

But, --append-only arguments in authorized_keys will override the repository config, therefore append-only mode can be enabled on a key by key basis.

Refer to the sshd(8) man page for more details on SSH options. See also borg serve

Automated backups to a local hard drive

This guide shows how to automate backups to a hard drive directly connected to your computer. If a backup hard drive is connected, backups are automatically started, and the drive shut-down and disconnected when they are done.

This guide is written for a Linux-based operating system and makes use of systemd and udev.

Overview

An udev rule is created to trigger on the addition of block devices. The rule contains a tag that triggers systemd to start a oneshot service. The oneshot service executes a script in the standard systemd service environment, which automatically captures stdout/stderr and logs it to the journal.

The script mounts the added block device, if it is a registered backup drive, and creates backups on it. When done, it optionally unmounts the file system and spins the drive down, so that it may be physically disconnected.

Configuring the system

First, create the /etc/backups directory (as root). All configuration goes into this directory.

Find out the ID of the partition table of your backup disk (here assumed to be /dev/sdz):

lsblk --fs -o +PTUUID /dev/sdz

Then, create /etc/backups/40-backup.rules with the following content (all on one line):

ACTION=="add", SUBSYSTEM=="block", ENV{ID_PART_TABLE_UUID}=="<the PTUUID you just noted>", TAG+="systemd", ENV{SYSTEMD_WANTS}="automatic-backup.service"

The "systemd" tag in conjunction with the SYSTEMD_WANTS environment variable has systemd launch the "automatic-backup" service, which we will create next, as the /etc/backups/automatic-backup.service file:

[Service]
Type=oneshot
ExecStart=/etc/backups/run.sh

Now, create the main backup script, /etc/backups/run.sh. Below is a template, modify it to suit your needs (e.g. more backup sets, dumping databases etc.).

#!/bin/bash -ue

# The udev rule is not terribly accurate and may trigger our service before
# the kernel has finished probing partitions. Sleep for a bit to ensure
# the kernel is done.
#
# This can be avoided by using a more precise udev rule, e.g. matching
# a specific hardware path and partition.
sleep 5

#
# Script configuration
#

# The backup partition is mounted there
MOUNTPOINT=/mnt/backup

# This is the location of the Borg repository
TARGET=$MOUNTPOINT/borg-backups/backup.borg

# Archive name schema
DATE=$(date --iso-8601)-$(hostname)

# This is the file that will later contain UUIDs of registered backup drives
DISKS=/etc/backups/backup.disks

# Find whether the connected block device is a backup drive
for uuid in $(lsblk --noheadings --list --output uuid)
do
        if grep --quiet --fixed-strings $uuid $DISKS; then
                break
        fi
        uuid=
done

if [ ! $uuid ]; then
        echo "No backup disk found, exiting"
        exit 0
fi

echo "Disk $uuid is a backup disk"
partition_path=/dev/disk/by-uuid/$uuid
# Mount file system if not already done. This assumes that if something is already
# mounted at $MOUNTPOINT, it is the backup drive. It won't find the drive if
# it was mounted somewhere else.
findmnt $MOUNTPOINT >/dev/null || mount $partition_path $MOUNTPOINT
drive=$(lsblk --inverse --noheadings --list --paths --output name $partition_path | head --lines 1)
echo "Drive path: $drive"

#
# Create backups
#

# Options for borg create
BORG_OPTS="--stats --one-file-system --compression lz4 --checkpoint-interval 86400"

# Set BORG_PASSPHRASE or BORG_PASSCOMMAND somewhere around here, using export,
# if encryption is used.

# No one can answer if Borg asks these questions, it is better to just fail quickly
# instead of hanging.
export BORG_RELOCATED_REPO_ACCESS_IS_OK=no
export BORG_UNKNOWN_UNENCRYPTED_REPO_ACCESS_IS_OK=no

# Log Borg version
borg --version

echo "Starting backup for $DATE"

# This is just an example, change it however you see fit
borg create $BORG_OPTS \
  --exclude root/.cache \
  --exclude var/lib/docker/devicemapper \
  $TARGET::$DATE-$$-system \
  / /boot

# /home is often a separate partition / file system.
# Even if it isn't (add --exclude /home above), it probably makes sense
# to have /home in a separate archive.
borg create $BORG_OPTS \
  --exclude 'sh:home/*/.cache' \
  $TARGET::$DATE-$$-home \
  /home/

echo "Completed backup for $DATE"

# Just to be completely paranoid
sync

if [ -f /etc/backups/autoeject ]; then
        umount $MOUNTPOINT
        hdparm -Y $drive
fi

if [ -f /etc/backups/backup-suspend ]; then
        systemctl suspend
fi

Create the /etc/backups/autoeject file to have the script automatically eject the drive after creating the backup. Rename the file to something else (e.g. /etc/backup/autoeject-no) when you want to do something with the drive after creating backups (e.g running check).

Create the /etc/backups/backup-suspend file if the machine should suspend after completing the backup. Don't forget to physically disconnect the device before resuming, otherwise you'll enter a cycle. You can also add an option to power down instead.

Create an empty /etc/backups/backup.disks file, you'll register your backup drives there.

The last part is to actually enable the udev rules and services:

ln -s /etc/backups/40-backup.rules /etc/udev/rules.d/40-backup.rules
ln -s /etc/backups/automatic-backup.service /etc/systemd/system/automatic-backup.service
systemctl daemon-reload
udevadm control --reload
Adding backup hard drives

Connect your backup hard drive. Format it, if not done already. Find the UUID of the file system that backups should be stored on:

lsblk -o+uuid,label

Note the UUID into the /etc/backup/backup.disks file.

Mount the drive to /mnt/backup.

Initialize a Borg repository at the location indicated by TARGET:

borg init --encryption ... /mnt/backup/borg-backups/backup.borg

Unmount and reconnect the drive, or manually start the automatic-backup service to start the first backup:

systemctl start --no-block automatic-backup

See backup logs using journalctl:

journalctl -fu automatic-backup [-n number-of-lines]
Security considerations

The script as shown above will mount any file system with an UUID listed in /etc/backup/backup.disks. The UUID check is a safety / annoyance-reduction mechanism to keep the script from blowing up whenever a random USB thumb drive is connected. It is not meant as a security mechanism. Mounting file systems and reading repository data exposes additional attack surfaces (kernel file system drivers, possibly user space services and Borg itself). On the other hand, someone standing right next to your computer can attempt a lot of attacks, most of which are easier to do than e.g. exploiting file systems (installing a physical key logger, DMA attacks, stealing the machine, ...).

Borg ensures that backups are not created on random drives that "just happen" to contain a Borg repository. If an unknown unencrypted repository is encountered, then the script aborts (BORG_UNKNOWN_UNENCRYPTED_REPO_ACCESS_IS_OK=no).

Backups are only created on hard drives that contain a Borg repository that is either known (by ID) to your machine or you are using encryption and the passphrase of the repository has to match the passphrase supplied to Borg.

Backing up entire disk images

Backing up disk images can still be efficient with Borg because its deduplication technique makes sure only the modified parts of the file are stored. Borg also has optional simple sparse file support for extract.

It is of utmost importancy to pin down the disk you want to backup. You need to use the SERIAL for that. Use:

# You can find the short disk serial by:
# udevadm info --query=property --name=nvme1n1 | grep ID_SERIAL_SHORT | cut -d '=' -f 2

DISK_SERIAL="7VS0224F"
DISK_ID=$(readlink -f /dev/disk/by-id/*"${DISK_SERIAL}") # Returns /dev/nvme1n1

mapfile -t PARTITIONS < <(lsblk -o NAME,TYPE -p -n -l "$DISK_ID" | awk '$2 == "part" {print $1}')
echo "Partitions of $DISK_ID:"
echo "${PARTITIONS[@]}"
echo "Disk Identifier: $DISK_ID"

# Use the following line to perform a borg backup for the full disk:
# borg create --read-special /path/to/repo::{now} "$DISK_ID"

# Use the following to perform a borg backup for all partitions of the disk
# borg create --read-special /path/to/repo::{now} "${PARTITIONS[@]}"

# Example output:
# Partitions of /dev/nvme1n1:
# /dev/nvme1n1p1
# /dev/nvme1n1p2
# /dev/nvme1n1p3
# Disk Identifier: /dev/nvme1n1
# borg create --read-special /path/to/repo::{now} /dev/nvme1n1
# borg create --read-special /path/to/repo::{now} /dev/nvme1n1p1 /dev/nvme1n1p2 /dev/nvme1n1p3
Decreasing the size of image backups

Disk images are as large as the full disk when uncompressed and might not get much smaller post-deduplication after heavy use because virtually all file systems don't actually delete file data on disk but instead delete the filesystem entries referencing the data. Therefore, if a disk nears capacity and files are deleted again, the change will barely decrease the space it takes up when compressed and deduplicated. Depending on the filesystem, there are several ways to decrease the size of a disk image:

Using ntfsclone (NTFS, i.e. Windows VMs)

ntfsclone can only operate on filesystems with the journal cleared (i.e. turned-off machines), which somewhat limits its utility in the case of VM snapshots. However, when it can be used, its special image format is even more efficient than just zeroing and deduplicating. For backup, save the disk header and the contents of each partition:

HEADER_SIZE=$(sfdisk -lo Start $DISK | grep -A1 -P 'Start$' | tail -n1 | xargs echo)
PARTITIONS=$(sfdisk -lo Device,Type $DISK | sed -e '1,/Device\s*Type/d')
dd if=$DISK count=$HEADER_SIZE | borg create repo::hostname-partinfo -
echo "$PARTITIONS" | grep NTFS | cut -d' ' -f1 | while read x; do
    PARTNUM=$(echo $x | grep -Eo "[0-9]+$")
    ntfsclone -so - $x | borg create repo::hostname-part$PARTNUM -
done
# to backup non-NTFS partitions as well:
echo "$PARTITIONS" | grep -v NTFS | cut -d' ' -f1 | while read x; do
    PARTNUM=$(echo $x | grep -Eo "[0-9]+$")
    borg create --read-special repo::hostname-part$PARTNUM $x
done

Restoration is a similar process:

borg extract --stdout repo::hostname-partinfo | dd of=$DISK && partprobe
PARTITIONS=$(sfdisk -lo Device,Type $DISK | sed -e '1,/Device\s*Type/d')
borg list --format {archive}{NL} repo | grep 'part[0-9]*$' | while read x; do
    PARTNUM=$(echo $x | grep -Eo "[0-9]+$")
    PARTITION=$(echo "$PARTITIONS" | grep -E "$DISKp?$PARTNUM" | head -n1)
    if echo "$PARTITION" | cut -d' ' -f2- | grep -q NTFS; then
        borg extract --stdout repo::$x | ntfsclone -rO $(echo "$PARTITION" | cut -d' ' -f1) -
    else
        borg extract --stdout repo::$x | dd of=$(echo "$PARTITION" | cut -d' ' -f1)
    fi
done

Note

When backing up a disk image (as opposed to a real block device), mount it as a loopback image to use the above snippets:

DISK=$(losetup -Pf --show /path/to/disk/image)
# do backup as shown above
losetup -d $DISK
Using zerofree (ext2, ext3, ext4)

zerofree works similarly to ntfsclone in that it zeros out unused chunks of the FS, except it works in place, zeroing the original partition. This makes the backup process a bit simpler:

sfdisk -lo Device,Type $DISK | sed -e '1,/Device\s*Type/d' | grep Linux | cut -d' ' -f1 | xargs -n1 zerofree
borg create --read-special repo::hostname-disk $DISK

Because the partitions were zeroed in place, restoration is only one command:

borg extract --stdout repo::hostname-disk | dd of=$DISK

Note

The "traditional" way to zero out space on a partition, especially one already mounted, is to simply dd from /dev/zero to a temporary file and delete it. This is ill-advised for the reasons mentioned in the zerofree man page:

  • it is slow

  • it makes the disk image (temporarily) grow to its maximal extent

  • it (temporarily) uses all free space on the disk, so other concurrent write actions may fail.

Virtual machines

If you use non-snapshotting backup tools like Borg to back up virtual machines, then the VMs should be turned off for the duration of the backup. Backing up live VMs can (and will) result in corrupted or inconsistent backup contents: a VM image is just a regular file to Borg with the same issues as regular files when it comes to concurrent reading and writing from the same file.

For backing up live VMs use filesystem snapshots on the VM host, which establishes crash-consistency for the VM images. This means that with most file systems (that are journaling) the FS will always be fine in the backup (but may need a journal replay to become accessible).

Usually this does not mean that file contents on the VM are consistent, since file contents are normally not journaled. Notable exceptions are ext4 in data=journal mode, ZFS and btrfs (unless nodatacow is used).

Applications designed with crash-consistency in mind (most relational databases like PostgreSQL, SQLite etc. but also for example Borg repositories) should always be able to recover to a consistent state from a backup created with crash-consistent snapshots (even on ext4 with data=writeback or XFS). Other applications may require a lot of work to reach application-consistency; it's a broad and complex issue that cannot be explained in entirety here.

Hypervisor snapshots capturing most of the VM's state can also be used for backups and can be a better alternative to pure file system based snapshots of the VM's disk, since no state is lost. Depending on the application this can be the easiest and most reliable way to create application-consistent backups.

Borg doesn't intend to address these issues due to their huge complexity and platform/software dependency. Combining Borg with the mechanisms provided by the platform (snapshots, hypervisor features) will be the best approach to start tackling them.

Backing up in pull mode

Typically the borg client connects to a backup server using SSH as a transport when initiating a backup. This is referred to as push mode.

If you however require the backup server to initiate the connection or prefer it to initiate the backup run, one of the following workarounds is required to allow such a pull mode setup.

A common use case for pull mode is to backup a remote server to a local personal computer.

SSHFS

Assuming you have a pull backup system set up with borg, where a backup server pulls the data from the target via SSHFS. In this mode, the backup client's file system is mounted remotely on the backup server. Pull mode is even possible if the SSH connection must be established by the client via a remote tunnel. Other network file systems like NFS or SMB could be used as well, but SSHFS is very simple to set up and probably the most secure one.

There are some restrictions caused by SSHFS. For example, unless you define UID and GID mappings when mounting via sshfs, owners and groups of the mounted file system will probably change, and you may not have access to those files if BorgBackup is not run with root privileges.

SSHFS is a FUSE file system and uses the SFTP protocol, so there may be also other unsupported features that the actual implementations of sshfs, libfuse and sftp on the backup server do not support, like file name encodings, ACLs, xattrs or flags. So there is no guarantee that you are able to restore a system completely in every aspect from such a backup.

Warning

To mount the client's root file system you will need root access to the client. This contradicts to the usual threat model of BorgBackup, where clients don't need to trust the backup server (data is encrypted). In pull mode the server (when logged in as root) could cause unlimited damage to the client. Therefore, pull mode should be used only from servers you do fully trust!

Warning

Additionally, while being chrooted into the client's root file system, code from the client will be executed. Thus, you should only do that when fully trusting the client.

Warning

The chroot method was chosen to get the right user and group name-id mappings, assuming they only come from files (/etc/passwd and group). This assumption might be wrong, e.g. if users/groups also come from ldap or other providers. Thus, it might be better to use --numeric-ids and not archive any user or group names (but just the numeric IDs) and not use chroot.

Creating a backup

Generally, in a pull backup situation there is no direct way for borg to know the client's original UID:GID name mapping of files, because Borg would use /etc/passwd and /etc/group of the backup server to map the names. To derive the right names, Borg needs to have access to the client's passwd and group files and use them in the backup process.

The solution to this problem is chrooting into an sshfs mounted directory. In this example the whole client root file system is mounted. We use the stand-alone BorgBackup executable and copy it into the mounted file system to make Borg available after entering chroot; this can be skipped if Borg is already installed on the client.

# Mount client root file system.
mkdir /tmp/sshfs
sshfs root@host:/ /tmp/sshfs
# Mount BorgBackup repository inside it.
mkdir /tmp/sshfs/borgrepo
mount --bind /path/to/repo /tmp/sshfs/borgrepo
# Make borg executable available.
cp /usr/local/bin/borg /tmp/sshfs/usr/local/bin/borg
# Mount important system directories and enter chroot.
cd /tmp/sshfs
for i in dev proc sys; do mount --bind /$i $i; done
chroot /tmp/sshfs

Now we are on the backup system but inside a chroot with the client's root file system. We have a copy of Borg binary in /usr/local/bin and the repository in /borgrepo. Borg will back up the client's user/group names, and we can create the backup, retaining the original paths, excluding the repository:

borg create --exclude borgrepo --files-cache ctime,size /borgrepo::archive /

For the sake of simplicity only borgrepo is excluded here. You may want to set up an exclude file with additional files and folders to be excluded. Also note that we have to modify Borg's file change detection behaviour – SSHFS cannot guarantee stable inode numbers, so we have to supply the --files-cache option.

Finally, we need to exit chroot, unmount all the stuff and clean up:

exit # exit chroot
rm /tmp/sshfs/usr/local/bin/borg
cd /tmp/sshfs
for i in dev proc sys borgrepo; do umount ./$i; done
rmdir borgrepo
cd ~
umount /tmp/sshfs
rmdir /tmp/sshfs

Thanks to secuser on IRC for this how-to!

Restore methods

The counterpart of a pull backup is a push restore. Depending on the type of restore – full restore or partial restore – there are different methods to make sure the correct IDs are restored.

Partial restore

In case of a partial restore, using the archived UIDs/GIDs might lead to wrong results if the name-to-ID mapping on the target system has changed compared to backup time (might be the case e.g. for a fresh OS install).

The workaround again is chrooting into an sshfs mounted directory, so Borg is able to map the user/group names of the backup files to the actual IDs on the client. This example is similar to the backup above – only the Borg command is different:

# Mount client root file system.
mkdir /tmp/sshfs
sshfs root@host:/ /tmp/sshfs
# Mount BorgBackup repository inside it.
mkdir /tmp/sshfs/borgrepo
mount --bind /path/to/repo /tmp/sshfs/borgrepo
# Make borg executable available.
cp /usr/local/bin/borg /tmp/sshfs/usr/local/bin/borg
# Mount important system directories and enter chroot.
cd /tmp/sshfs
for i in dev proc sys; do mount --bind /$i $i; done
chroot /tmp/sshfs

Now we can run

borg extract /borgrepo::archive PATH

to partially restore whatever we like. Finally, do the clean-up:

exit # exit chroot
rm /tmp/sshfs/usr/local/bin/borg
cd /tmp/sshfs
for i in dev proc sys borgrepo; do umount ./$i; done
rmdir borgrepo
cd ~
umount /tmp/sshfs
rmdir /tmp/sshfs
Full restore

When doing a full restore, we restore all files (including the ones containing the ID-to-name mapping, /etc/passwd and /etc/group). Everything will be consistent automatically if we restore the numeric IDs stored in the archive. So there is no need for a chroot environment; we just mount the client file system and extract a backup, utilizing the --numeric-ids option:

sshfs root@host:/ /mnt/sshfs
cd /mnt/sshfs
borg extract --numeric-ids /path/to/repo::archive
cd ~
umount /mnt/sshfs
Simple (lossy) full restore

Using borg export-tar it is possible to stream a backup to the client and directly extract it without the need of mounting with SSHFS:

borg export-tar /path/to/repo::archive - | ssh root@host 'tar -C / -x'

Note that in this scenario the tar format is the limiting factor – it cannot restore all the advanced features that BorgBackup supports. See borg export-tar for limitations.

socat

In this setup a SSH connection from the backup server to the client is established that uses SSH reverse port forwarding to transparently tunnel data between UNIX domain sockets on the client and server and the socat tool to connect these with the borg client and server processes, respectively.

The program socat has to be available on the backup server and on the client to be backed up.

When pushing a backup the borg client (holding the data to be backed up) connects to the backup server via ssh, starts borg serve on the backup server and communicates via standard input and output (transported via SSH) with the process on the backup server.

With the help of socat this process can be reversed. The backup server will create a connection to the client (holding the data to be backed up) and will pull the data.

In the following example borg-server connects to borg-client to pull a backup.

To provide a secure setup sockets should be stored in /run/borg, only accessible to the users that run the backup process. So on both systems, borg-server and borg-client the folder /run/borg has to be created:

sudo mkdir -m 0700 /run/borg

On borg-server the socket file is opened by the user running the borg serve process writing to the repository so the user has to have read and write permissions on /run/borg:

borg-server:~$ sudo chown borgs /run/borg

On borg-client the socket file is created by ssh, so the user used to connect to borg-client has to have read and write permissions on /run/borg:

borg-client:~$ sudo chown borgc /run/borg

On borg-server, we have to start the command borg serve and make its standard input and output available to a unix socket:

borg-server:~$ socat UNIX-LISTEN:/run/borg/reponame.sock,fork EXEC:"borg serve --append-only --restrict-to-path /path/to/repo"

Socat will wait until a connection is opened. Then socat will execute the command given, redirecting Standard Input and Output to the unix socket. The optional arguments for borg serve are not necessary but a sane default.

Note

When used in production you may also use systemd socket-based activation instead of socat on the server side. You would wrap the borg serve command in a service unit and configure a matching socket unit to start the service whenever a client connects to the socket.

Now we need a way to access the unix socket on borg-client (holding the data to be backed up), as we created the unix socket on borg-server Opening a SSH connection from the borg-server to the borg-client with reverse port forwarding can do this for us:

borg-server:~$ ssh -R /run/borg/reponame.sock:/run/borg/reponame.sock borgc@borg-client

Note

As the default value of OpenSSH for StreamLocalBindUnlink is no, the socket file created by sshd is not removed. Trying to connect a second time, will print a short warning, and the forwarding does not take place:

Warning: remote port forwarding failed for listen path /run/borg/reponame.sock

When you are done, you have to manually remove the socket file, otherwise you may see an error like this when trying to execute borg commands:

Remote: YYYY/MM/DD HH:MM:SS socat[XXX] E connect(5, AF=1 "/run/borg/reponame.sock", 13): Connection refused
Connection closed by remote host. Is borg working on the server?

When a process opens the socket on borg-client, SSH will forward all data to the socket on borg-server.

The next step is to tell borg on borg-client to use the unix socket to communicate with the borg serve command on borg-server via the socat socket instead of SSH:

borg-client:~$ export BORG_RSH="sh -c 'exec socat STDIO UNIX-CONNECT:/run/borg/reponame.sock'"

The default value for BORG_RSH is ssh. By default Borg uses SSH to create the connection to the backup server. Therefore Borg parses the repo URL and adds the server name (and other arguments) to the SSH command. Those arguments can not be handled by socat. We wrap the command with sh to ignore all arguments intended for the SSH command.

All Borg commands can now be executed on borg-client. For example to create a backup execute the borg create command:

borg-client:~$ borg create ssh://borg-server/path/to/repo::archive /path_to_backup

When automating backup creation, the interactive ssh session may seem inappropriate. An alternative way of creating a backup may be the following command:

borg-server:~$ ssh \
   -R /run/borg/reponame.sock:/run/borg/reponame.sock \
   borgc@borg-client \
   borg create \
   --rsh "sh -c 'exec socat STDIO UNIX-CONNECT:/run/borg/reponame.sock'" \
   ssh://borg-server/path/to/repo::archive /path_to_backup \
   ';' rm /run/borg/reponame.sock

This command also automatically removes the socket file after the borg create command is done.

ssh-agent

In this scenario borg-server initiates an SSH connection to borg-client and forwards the authentication agent connection.

After that, it works similar to the push mode: borg-client initiates another SSH connection back to borg-server using the forwarded authentication agent connection to authenticate itself, starts borg serve and communicates with it.

Using this method requires ssh access of user borgs to borgc@borg-client, where:

  • borgs is the user on the server side with read/write access to local borg repository.

  • borgc is the user on the client side with read access to files meant to be backed up.

Applying this method for automated backup operations

Assume that the borg-client host is untrusted. Therefore we do some effort to prevent a hostile user on the borg-client side to do something harmful. In case of a fully trusted borg-client the method could be simplified.

Preparing the server side

Do this once for each client on borg-server to allow borgs to connect itself on borg-server using a dedicated ssh key:

borgs@borg-server$ install -m 700 -d ~/.ssh/
borgs@borg-server$ ssh-keygen -N '' -t rsa  -f ~/.ssh/borg-client_key
borgs@borg-server$ { echo -n 'command="borg serve --append-only --restrict-to-repo ~/repo",restrict '; cat ~/.ssh/borg-client_key.pub; } >> ~/.ssh/authorized_keys
borgs@borg-server$ chmod 600 ~/.ssh/authorized_keys

install -m 700 -d ~/.ssh/

Create directory ~/.ssh with correct permissions if it does not exist yet.

ssh-keygen -N '' -t rsa  -f ~/.ssh/borg-client_key

Create an ssh key dedicated to communication with borg-client.

Note

Another more complex approach is using a unique ssh key for each pull operation. This is more secure as it guarantees that the key will not be used for other purposes.

{ echo -n 'command="borg serve --append-only --restrict-to-repo ~/repo",restrict '; cat ~/.ssh/borg-client_key.pub; } >> ~/.ssh/authorized_keys

Add borg-client's ssh public key to ~/.ssh/authorized_keys with forced command and restricted mode. The borg client is restricted to use one repo at the specified path and to append-only operation. Commands like delete, prune and compact have to be executed another way, for example directly on borg-server side or from a privileged, less restricted client (using another authorized_keys entry).

chmod 600 ~/.ssh/authorized_keys

Fix permissions of ~/.ssh/authorized_keys.

Pull operation

Initiating borg command execution from borg-server (e.g. init):

borgs@borg-server$ (
  eval $(ssh-agent) > /dev/null
  ssh-add -q ~/.ssh/borg-client_key
  echo 'your secure borg key passphrase' | \
    ssh -A -o StrictHostKeyChecking=no borgc@borg-client "BORG_PASSPHRASE=\$(cat) borg --rsh 'ssh -o StrictHostKeyChecking=no' init --encryption repokey ssh://borgs@borg-server/~/repo"
  kill "${SSH_AGENT_PID}"
)

Parentheses around commands are needed to avoid interference with a possibly already running ssh-agent. Parentheses are not needed when using a dedicated bash process.

eval $(ssh-agent) > /dev/null

Run the SSH agent in the background and export related environment variables to the current bash session.

ssh-add -q ~/.ssh/borg-client_key

Load the SSH private key dedicated to communication with the borg-client into the SSH agent. Look at man 1 ssh-add for a more detailed explanation.

Note

Care needs to be taken when loading keys into the SSH agent. Users on the borg-client having read/write permissions to the agent's UNIX-domain socket (at least borgc and root in our case) can access the agent on borg-server through the forwarded connection and can authenticate using any of the identities loaded into the agent (look at man 1 ssh for more detailed explanation). Therefore there are some security considerations:

  • Private keys loaded into the agent must not be used to enable access anywhere else.

  • The keys meant to be loaded into the agent must be specified explicitly, not from default locations.

  • The borg-client's entry in borgs@borg-server:~/.ssh/authorized_keys must be as restrictive as possible.

echo 'your secure borg key passphrase' | ssh -A -o StrictHostKeyChecking=no borgc@borg-client "BORG_PASSPHRASE=\$(cat) borg --rsh 'ssh -o StrictHostKeyChecking=no' init --encryption repokey ssh://borgs@borg-server/~/repo"

Run the borg init command on borg-client.

ssh://borgs@borg-server/~/repo refers to the repository repo within borgs's home directory on borg-server.

StrictHostKeyChecking=no is used to automatically add host keys to ~/.ssh/known_hosts without user intervention.

kill "${SSH_AGENT_PID}"

Kill ssh-agent with loaded keys when it is not needed anymore.

Frequently asked questions

Usage & Limitations

What is the difference between a repo on an external hard drive vs. repo on a server?

If Borg is running in client/server mode, the client uses SSH as a transport to talk to the remote agent, which is another Borg process (Borg is installed on the server, too) started automatically by the client. The Borg server is doing storage-related low-level repo operations (get, put, commit, check, compact), while the Borg client does the high-level stuff: deduplication, encryption, compression, dealing with archives, backups, restores, etc., which reduces the amount of data that goes over the network.

When Borg is writing to a repo on a locally mounted remote file system, e.g. SSHFS, the Borg client only can do file system operations and has no agent running on the remote side, so every operation needs to go over the network, which is slower.

Can I backup from multiple servers into a single repository?

Yes, this is possible from the technical standpoint, but it is not recommended from the security perspective. BorgBackup is built upon a defined Attack model that cannot provide its guarantees for multiple clients using the same repository. See Isn't BorgBackup's AES-CTR crypto broken? for a detailed explanation.

Also, in order for the deduplication used by Borg to work, it needs to keep a local cache containing checksums of all file chunks already stored in the repository. This cache is stored in ~/.cache/borg/. If Borg detects that a repository has been modified since the local cache was updated it will need to rebuild the cache. This rebuild can be quite time consuming.

So, yes it's possible. But it will be most efficient if a single repository is only modified from one place. Also keep in mind that Borg will keep an exclusive lock on the repository while creating or deleting archives, which may make simultaneous backups fail.

Can I back up to multiple, swapped backup targets?

It is possible to swap your backup disks if each backup medium is assigned its own repository by creating a new one with borg init.

Can I copy or synchronize my repo to another location?

If you want to have redundant backup repositories (preferably at separate locations), the recommended way to do that is like this:

  • borg init repo1

  • borg init repo2

  • client machine ---borg create---> repo1

  • client machine ---borg create---> repo2

This will create distinct repositories (separate repo ID, separate keys) and nothing bad happening in repo1 will influence repo2.

Some people decide against above recommendation and create identical copies of a repo (using some copy / sync / clone tool).

While this might be better than having no redundancy at all, you have to be very careful about how you do that and what you may / must not do with the result (if you decide against our recommendation).

What you would get with this is:

  • client machine ---borg create---> repo

  • repo ---copy/sync---> copy-of-repo

There is no special borg command to do the copying, you could just use any reliable tool that creates an identical copy (cp, rsync, rclone might be options).

But think about whether that is really what you want. If something goes wrong in repo, you will have the same issue in copy-of-repo.

Make sure you do the copy/sync while no backup is running, see borg with-lock about how to do that.

Also, you must not run borg against multiple instances of the same repo (like repo and copy-of-repo) as that would create severe issues:

  • Data loss: they have the same repository ID, so the borg client will think they are identical and e.g. use the same local cache for them (which is an issue if they happen to be not the same). See #4272 for an example.

  • Encryption security issues if you would update repo and copy-of-repo independently, due to AES counter reuse.

See also: My repository is corrupt, how can I restore from an older copy of it?

"this is either an attack or unsafe" warning

About the warning:

Cache, or information obtained from the security directory is newer than repository - this is either an attack or unsafe (multiple repos with same ID)

"unsafe": If not following the advice from the previous section, you can easily run into this by yourself by restoring an older copy of your repository.

"attack": maybe an attacker has replaced your repo by an older copy, trying to trick you into AES counter reuse, trying to break your repo encryption.

Borg users have also reported that fs issues (like hw issues / I/O errors causing the fs to become read-only) can cause this warning, see #7853.

If you'ld decide to ignore this and accept unsafe operation for this repository, you could delete the manifest-timestamp and the local cache:

borg config repo id   # shows the REPO_ID
rm ~/.config/borg/security/REPO_ID/manifest-timestamp
borg delete --cache-only REPO

This is an unsafe and unsupported way to use borg, you have been warned.

Which file types, attributes, etc. are not preserved?
  • UNIX domain sockets (because it does not make sense - they are meaningless without the running process that created them and the process needs to recreate them in any case). So, don't panic if your backup misses a UDS!

  • The precise on-disk (or rather: not-on-disk) representation of the holes in a sparse file. Archive creation has no special support for sparse files, holes are backed up as (deduplicated and compressed) runs of zero bytes. Archive extraction has optional support to extract all-zero chunks as holes in a sparse file.

  • Some filesystem specific attributes, like btrfs NOCOW, see Support for file metadata.

  • For hardlinked symlinks, the hardlinking can not be archived (and thus, the hardlinking will not be done at extraction time). The symlinks will be archived and extracted as non-hardlinked symlinks, see #2379.

Are there other known limitations?
  • A single archive can only reference a limited volume of file/dir metadata, usually corresponding to tens or hundreds of millions of files/dirs. When trying to go beyond that limit, you will get a fatal IntegrityError exception telling that the (archive) object is too big. An easy workaround is to create multiple archives with fewer items each. See also the Note about archive limitations and #1452.

    borg info shows how large (relative to the maximum size) existing archives are.

  • borg extract only supports restoring into an empty destination. After that, the destination will exactly have the contents of the extracted archive. If you extract into a non-empty destination, borg will (for example) not remove files which are in the destination, but not in the archive. See #4598 for a workaround and more details.

If a backup stops mid-way, does the already-backed-up data stay there?

Yes, Borg supports resuming backups.

During a backup a special checkpoint archive named <archive-name>.checkpoint is saved every checkpoint interval (the default value for this is 30 minutes) containing all the data backed-up until that point.

This checkpoint archive is a valid archive, but it is only a partial backup (not all files that you wanted to backup are contained in it). Having it in the repo until a successful, full backup is completed is useful because it references all the transmitted chunks up to the checkpoint. This means that in case of an interruption, you only need to retransfer the data since the last checkpoint.

If a backup was interrupted, you normally do not need to do anything special, just invoke borg create as you always do. If the repository is still locked, you may need to run borg break-lock before the next backup. You may use the same archive name as in previous attempt or a different one (e.g. if you always include the current datetime), it does not matter.

Borg always does full single-pass backups, so it will start again from the beginning - but it will be much faster, because some of the data was already stored into the repo (and is still referenced by the checkpoint archive), so it does not need to get transmitted and stored again.

Once your backup has finished successfully, you can delete all <archive-name>.checkpoint archives. If you run borg prune, it will also care for deleting unneeded checkpoints.

Note: the checkpointing mechanism creates hidden, partial files in an archive, so that checkpoints even work while a big file is being processed. They are named <filename>.borg_part_<N> and all operations usually ignore these files, but you can make them considered by giving the option --consider-part-files. You usually only need that option if you are really desperate (e.g. if you have no completed backup of that file and you'ld rather get a partial file extracted than nothing). You do not want to give that option under any normal circumstances.

Note that checkpoints inside files are created only since version 1.1, make sure you have an up-to-date version of borgbackup if you want to continue instead of retransferring a huge file. In some cases, there is only an outdated version shipped with your distribution (e.g. Debian). See Installation.

How can I backup huge file(s) over a unstable connection?

This is not a problem anymore.

For more details, see If a backup stops mid-way, does the already-backed-up data stay there?.

How can I switch append-only mode on and off?

You could do that (via borg config REPO append_only 0/1), but using different ssh keys and different entries in authorized_keys is much easier and also maybe has less potential of things going wrong somehow.

My machine goes to sleep causing Broken pipe

When backing up your data over the network, your machine should not go to sleep. On macOS you can use caffeinate to avoid that.

How can I restore huge file(s) over an unstable connection?

If you cannot manage to extract the whole big file in one go, you can extract all the part files and manually concatenate them together.

For more details, see If a backup stops mid-way, does the already-backed-up data stay there?.

How can I compare contents of an archive to my local filesystem?

You can instruct export-tar to send a tar stream to the stdout, and then use tar to perform the comparison:

borg export-tar /path/to/repo::archive-name - | tar --compare -f - -C /path/to/compare/to
My repository is corrupt, how can I restore from an older copy of it?

If your repositories are encrypted and have the same ID, the recommended method is to delete the corrupted repository, but keep its security info, and then copy the working repository to the same location:

borg delete --keep-security-info /path/to/repo
rsync -aH /path/to/repo-working/ /path/to/repo  # Note the trailing slash.

A plain delete command would remove the security info in ~/.config/borg/security, including the nonce value. In BorgBackup Encryption is AES-CTR, where the nonce is a counter. When the working repo was used later for creating new archives, Borg would re-use nonce values due to starting from a lower counter value given by the older copy of the repository. To prevent this, the keep-security-info option is applied so that the client-side nonce counter is kept.

Can Borg add redundancy to the backup data to deal with hardware malfunction?

No, it can't. While that at first sounds like a good idea to defend against some defect HDD sectors or SSD flash blocks, dealing with this in a reliable way needs a lot of low-level storage layout information and control which we do not have (and also can't get, even if we wanted).

So, if you need that, consider RAID or a filesystem that offers redundant storage or just make backups to different locations / different hardware.

See also #225.

Can Borg verify data integrity of a backup archive?

Yes, if you want to detect accidental data damage (like bit rot), use the check operation. It will notice corruption using CRCs and hashes. If you want to be able to detect malicious tampering also, use an encrypted repo. It will then be able to check using CRCs and HMACs.

Can I use Borg on SMR hard drives?

SMR (shingled magnetic recording) hard drives are very different from regular hard drives. Applications have to behave in certain ways or performance will be heavily degraded.

Borg 1.1 ships with default settings suitable for SMR drives, and has been successfully tested on Seagate Archive v2 drives using the ext4 file system.

Some Linux kernel versions between 3.19 and 4.5 had various bugs handling device-managed SMR drives, leading to IO errors, unresponsive drives and unreliable operation in general.

For more details, refer to #2252.

I get an IntegrityError or similar - what now?

A single error does not necessarily indicate bad hardware or a Borg bug. All hardware exhibits a bit error rate (BER). Hard drives are typically specified as exhibiting fewer than one error every 12 to 120 TB (one bit error in 10e14 to 10e15 bits). The specification is often called unrecoverable read error rate (URE rate).

Apart from these very rare errors there are two main causes of errors:

  1. Defective hardware: described below.

  2. Bugs in software (Borg, operating system, libraries): Ensure software is up to date. Check whether the issue is caused by any fixed bugs described in Important notes.

Finding defective hardware

Note

Hardware diagnostics are operating system dependent and do not apply universally. The commands shown apply for popular Unix-like systems. Refer to your operating system's manual.

Checking hard drives

Find the drive containing the repository and use findmnt, mount or lsblk to learn the device path (typically /dev/...) of the drive. Then, smartmontools can retrieve self-diagnostics of the drive in question:

# smartctl -a /dev/sdSomething

The Offline_Uncorrectable, Current_Pending_Sector and Reported_Uncorrect attributes indicate data corruption. A high UDMA_CRC_Error_Count usually indicates a bad cable.

I/O errors logged by the system (refer to the system journal or dmesg) can point to issues as well. I/O errors only affecting the file system easily go unnoticed, since they are not reported to applications (e.g. Borg), while these errors can still corrupt data.

Drives can corrupt some sectors in one event, while remaining reliable otherwise. Conversely, drives can fail completely with no advance warning. If in doubt, copy all data from the drive in question to another drive -- just in case it fails completely.

If any of these are suspicious, a self-test is recommended:

# smartctl -t long /dev/sdSomething

Running fsck if not done already might yield further insights.

Checking memory

Intermittent issues, such as borg check finding errors inconsistently between runs, are frequently caused by bad memory.

Run memtest86+ (or an equivalent memory tester) to verify that the memory subsystem is operating correctly.

Checking processors

Processors rarely cause errors. If they do, they are usually overclocked or otherwise operated outside their specifications. We do not recommend to operate hardware outside its specifications for productive use.

Tools to verify correct processor operation include Prime95 (mprime), linpack, and the Intel Processor Diagnostic Tool (applies only to Intel processors).

Repairing a damaged repository

With any defective hardware found and replaced, the damage done to the repository needs to be ascertained and fixed.

borg check provides diagnostics and --repair options for repositories with issues. We recommend to first run without --repair to assess the situation. If the found issues and proposed repairs seem right, re-run "check" with --repair enabled.

How probable is it to get a hash collision problem?

If you noticed, there are some issues (#170 (warning: hell) and #4884) about the probability of a chunk having the same hash as another chunk, making the file corrupted because it grabbed the wrong chunk. This is called the Birthday Problem.

There is a lot of probability in here so, I can give you my interpretation of such math but it's honestly better that you read it yourself and grab your own resolution from that.

Assuming that all your chunks have a size of \(2^{21}\) bytes (approximately 2.1 MB) and we have a "perfect" hash algorithm, we can think that the probability of collision would be of \(p^2/2^{n+1}\) then, using SHA-256 (\(n=256\)) and for example we have 1000 million chunks (\(p=10^9\)) (1000 million chunks would be about 2100TB). The probability would be around to 0.0000000000000000000000000000000000000000000000000000000000043.

A mass-murderer space rock happens about once every 30 million years on average. This leads to a probability of such an event occurring in the next second to about \(10^{-15}\). That's 45 orders of magnitude more probable than the SHA-256 collision. Briefly stated, if you find SHA-256 collisions scary then your priorities are wrong. This example was grabbed from this SO answer, it's great honestly.

Still, the real question is if Borg tries to not make this happen?

Well... it used to not check anything but there was a feature added which saves the size of the chunks too, so the size of the chunks is compared to the size that you got with the hash and if the check says there is a mismatch it will raise an exception instead of corrupting the file. This doesn't save us from everything but reduces the chances of corruption. There are other ways of trying to escape this but it would affect performance so much that it wouldn't be worth it and it would contradict Borg's design, so if you don't want this to happen, simply don't use Borg.

Why is the time elapsed in the archive stats different from wall clock time?

Borg needs to write the time elapsed into the archive metadata before finalizing the archive, compacting the segments, and committing the repo & cache. This means when Borg is run with e.g. the time command, the duration shown in the archive stats may be shorter than the full time the command runs for.

How do I configure different prune policies for different directories?

Say you want to prune /var/log faster than the rest of /. How do we implement that? The answer is to backup to different archive names and then implement different prune policies for different --glob-archives matching patterns.

For example, you could have a script that does:

borg create --exclude var/log $REPOSITORY:main-$(date +%Y-%m-%d) /
borg create $REPOSITORY:logs-$(date +%Y-%m-%d) /var/log

Then you would have two different prune calls with different policies:

borg prune --verbose --list -d 30 --glob-archives 'main-*' "$REPOSITORY"
borg prune --verbose --list -d 7  --glob-archives 'logs-*' "$REPOSITORY"

This will keep 7 days of logs and 30 days of everything else.

How do I remove files from an existing backup?

A file is only removed from a BorgBackup repository if all archives that contain the file are deleted and the corresponding data chunks are removed from the repository There are two ways how to remove files from a repository.

1. Use borg delete to remove all archives that contain the files. This will of course delete everything in the archive, not only some files.

2. If you really want to remove only some specific files, you can run the borg recreate command to rewrite all archives with a different --exclude pattern. See the examples in the manpage for more information.

Finally, run borg compact with the --threshold 0 option to delete the data chunks from the repository.

Can I safely change the compression level or algorithm?

The compression level and algorithm don't affect deduplication. Chunk ID hashes are calculated before compression. New compression settings will only be applied to new chunks, not existing chunks. So it's safe to change them.

Security

Isn't BorgBackup's AES-CTR crypto broken?

If a nonce (counter) value is reused, AES-CTR mode crypto is broken.

To exploit the AES counter management issue, an attacker would need to have access to the borg repository.

By tampering with the repo, the attacker could bring the repo into a state so that it reports a lower "highest used counter value" than the one that actually was used. The client would usually notice that, because it rather trusts the clientside stored "highest used counter value" than trusting the server.

But there are situations, where this is simply not possible:

  • If clients A and B used the repo, the client A can only know its own highest CTR value, but not the one produced by B. That is only known to (B and) the server (the repo) and thus the client A needs to trust the server about the value produced by B in that situation. You can't do much about this except not having multiple clients per repo.

  • Even if there is only one client, if client-side information is completely lost (e.g. due to disk defect), the client also needs to trust the value from server side. You can avoid this by not continuing to write to the repository after you have lost clientside borg information.

How important is the $HOME/.config/borg directory?

The Borg config directory has content that you should take care of:

security subdirectory

Each directory here represents one Borg repository by its ID and contains the last known status. If a repository's status is different from this information at the beginning of BorgBackup operation, Borg outputs warning messages and asks for confirmation, so make sure you do not lose or manipulate these files. However, apart from those warnings, a loss of these files can be recovered.

keys subdirectory

All your borg keyfile keys are stored in this directory. Please note that borg repokey keys are stored inside the repository. You MUST make sure to have an independent backup of these keyfiles, otherwise you cannot access your backups anymore if you lose them. You also MUST keep these files secret; everyone who gains access to your repository and has the corresponding keyfile (and the key passphrase) can extract it.

Make sure that only you have access to the Borg config directory.

Do I need to take security precautions regarding the cache?

The cache contains a lot of metadata information about the files in your repositories and it is not encrypted.

However, the assumption is that the cache is being stored on the very same system which also contains the original files which are being backed up. So someone with access to the cache files would also have access the original files anyway.

The Internals section contains more details about The cache. If you ever need to move the cache to a different location, this can be achieved by using the appropriate Environment Variables.

How can I specify the encryption passphrase programmatically?

There are several ways to specify a passphrase without human intervention:

Setting BORG_PASSPHRASE

The passphrase can be specified using the BORG_PASSPHRASE environment variable. This is often the simplest option, but can be insecure if the script that sets it is world-readable.

Note

Be careful how you set the environment; using the env command, a system() call or using inline shell scripts (e.g. BORG_PASSPHRASE=hunter2 borg ...) might expose the credentials in the process list directly and they will be readable to all users on a system. Using export in a shell script file should be safe, however, as the environment of a process is accessible only to that user.

Using BORG_PASSCOMMAND with a properly permissioned file

Another option is to create a file with a password in it in your home directory and use permissions to keep anyone else from reading it. For example, first create a key:

(umask 0077; head -c 32 /dev/urandom | base64 -w 0 > ~/.borg-passphrase)

Then in an automated script one can put:

export BORG_PASSCOMMAND="cat $HOME/.borg-passphrase"

and Borg will automatically use that passphrase.

Using keyfile-based encryption with a blank passphrase

It is possible to encrypt your repository in keyfile mode instead of the default repokey mode and use a blank passphrase for the key file (simply press Enter twice when borg init asks for the password). See Repository encryption for more details.

Using BORG_PASSCOMMAND with macOS Keychain

macOS has a native manager for secrets (such as passphrases) which is safer than just using a file as it is encrypted at rest and unlocked manually (fortunately, the login keyring automatically unlocks when you login). With the built-in security command, you can access it from the command line, making it useful for BORG_PASSCOMMAND.

First generate a passphrase and use security to save it to your login (default) keychain:

security add-generic-password -D secret -U -a $USER -s borg-passphrase -w $(head -c 32 /dev/urandom | base64 -w 0)

In your backup script retrieve it in the BORG_PASSCOMMAND:

export BORG_PASSCOMMAND="security find-generic-password -a $USER -s borg-passphrase -w"
Using BORG_PASSCOMMAND with GNOME Keyring

GNOME also has a keyring daemon that can be used to store a Borg passphrase. First ensure libsecret-tools, gnome-keyring and libpam-gnome-keyring are installed. If libpam-gnome-keyring wasn't already installed, ensure it runs on login:

sudo sh -c "echo session optional pam_gnome_keyring.so auto_start >> /etc/pam.d/login"
sudo sh -c "echo password optional pam_gnome_keyring.so >> /etc/pam.d/passwd"
# you may need to relogin afterwards to activate the login keyring

Then add a secret to the login keyring:

head -c 32 /dev/urandom | base64 -w 0 | secret-tool store borg-repository repo-name --label="Borg Passphrase"

If a dialog box pops up prompting you to pick a password for a new keychain, use your login password. If there is a checkbox for automatically unlocking on login, check it to allow backups without any user intervention whatsoever.

Once the secret is saved, retrieve it in a backup script using BORG_PASSCOMMAND:

export BORG_PASSCOMMAND="secret-tool lookup borg-repository repo-name"

Note

For this to automatically unlock the keychain it must be run in the dbus session of an unlocked terminal; for example, running a backup script as a cron job might not work unless you also export DISPLAY=:0 so secret-tool can pick up your open session. It gets even more complicated when you are running the tool as a different user (e.g. running a backup as root with the password stored in the user keyring).

Using BORG_PASSCOMMAND with KWallet

KDE also has a keychain feature in the form of KWallet. The command-line tool kwalletcli can be used to store and retrieve secrets. Ensure kwalletcli is installed, generate a passphrase, and store it in your "wallet":

head -c 32 /dev/urandom | base64 -w 0 | kwalletcli -Pe borg-passphrase -f Passwords

Once the secret is saved, retrieve it in a backup script using BORG_PASSCOMMAND:

export BORG_PASSCOMMAND="kwalletcli -e borg-passphrase -f Passwords"
When backing up to remote encrypted repos, is encryption done locally?

Yes, file and directory metadata and data is locally encrypted, before leaving the local machine. We do not mean the transport layer encryption by that, but the data/metadata itself. Transport layer encryption (e.g. when ssh is used as a transport) applies additionally.

When backing up to remote servers, do I have to trust the remote server?

Yes and No.

No, as far as data confidentiality is concerned - if you use encryption, all your files/dirs data and metadata are stored in their encrypted form into the repository.

Yes, as an attacker with access to the remote server could delete (or otherwise make unavailable) all your backups.

How can I protect against a hacked backup client?

Assume you backup your backup client machine C to the backup server S and C gets hacked. In a simple push setup, the attacker could then use borg on C to delete all backups residing on S.

These are your options to protect against that:

  • Do not allow to permanently delete data from the repo, see Append-only mode (forbid compaction).

  • Use a pull-mode setup using ssh -R, see Backing up in pull mode for more information.

  • Mount C's filesystem on another machine and then create a backup of it.

  • Do not give C filesystem-level access to S.

See Hosting repositories for a detailed protection guide.

How can I protect against a hacked backup server?

Just in case you got the impression that pull-mode backups are way more safe than push-mode, you also need to consider the case that your backup server S gets hacked. In case S has access to a lot of clients C, that might bring you into even bigger trouble than a hacked backup client in the previous FAQ entry.

These are your options to protect against that:

  • Use the standard push-mode setup (see also previous FAQ entry).

  • Mount (the repo part of) S's filesystem on C.

  • Do not give S file-system level access to C.

  • Have your backup server at a well protected place (maybe not reachable from the internet), configure it safely, apply security updates, monitor it, ...

How can I protect against theft, sabotage, lightning, fire, ...?

In general: if your only backup medium is nearby the backupped machine and always connected, you can easily get into trouble: they likely share the same fate if something goes really wrong.

Thus:

  • have multiple backup media

  • have media disconnected from network, power, computer

  • have media at another place

  • have a relatively recent backup on your media

How do I report a security issue with Borg?

Send a private email to the security contact if you think you have discovered a security issue. Please disclose security issues responsibly.

How important are the nonce files?

Borg uses AES-CTR encryption. An essential part of AES-CTR is a sequential counter that must never repeat. If the same value of the counter is used twice in the same repository, an attacker can decrypt the data. The counter is stored in the home directory of each user ($HOME/.config/borg/security/$REPO_ID/nonce) as well as in the repository (/path/to/repo/nonce). When creating a new archive borg uses the highest of the two values. The value of the counter in the repository may be higher than your local value if another user has created an archive more recently than you did.

Since the nonce is not necessary to read the data that is already encrypted, borg info, borg list, borg extract and borg mount should work just fine without it.

If the nonce file stored in the repo is lost, but you still have your local copy, borg will recreate the repository nonce file the next time you run borg create. This should be safe for repositories that are only used from one user account on one machine.

For repositories that are used by multiple users and/or from multiple machines it is safest to avoid running any commands that modify the repository after the nonce is deleted or if you suspect it may have been tampered with. See Attack model.

Common issues

/path/to/repo is not a valid repository. Check repo config.

There can be many causes of this error. E.g. you have incorrectly specified the repository path.

You will also get this error if you try to access a repository that uses the argon2 key algorithm using an old version of borg. We recommend upgrading to the latest stable version and trying again. We are sorry. We should have thought abount forward compatibility and implemented a more helpful error message.

Why does Borg extract hang after some time?

When I do a borg extract, after a while all activity stops, no cpu usage, no downloads.

This may happen when the SSH connection is stuck on server side. You can configure SSH on client side to prevent this by sending keep-alive requests, for example in ~/.ssh/config:

Host borg.example.com
    # Client kills connection after 3*30 seconds without server response:
    ServerAliveInterval 30
    ServerAliveCountMax 3

You can also do the opposite and configure SSH on server side in /etc/ssh/sshd_config, to make the server send keep-alive requests to the client:

# Server kills connection after 3*30 seconds without client response:
ClientAliveInterval 30
ClientAliveCountMax 3
How can I deal with my very unstable SSH connection?

If you have issues with lost connections during long-running borg commands, you could try to work around:

  • Make partial extracts like borg extract REPO PATTERN to do multiple smaller extraction runs that complete before your connection has issues.

  • Try using borg mount REPO MOUNTPOINT and rsync -avH from MOUNTPOINT to your desired extraction directory. If the connection breaks down, just repeat that over and over again until rsync does not find anything to do any more. Due to the way borg mount works, this might be less efficient than borg extract for bigger volumes of data.

Why do I get "connection closed by remote" after a while?

When doing a backup to a remote server (using a ssh: repo URL), it sometimes stops after a while (some minutes, hours, ... - not immediately) with "connection closed by remote" error message. Why?

That's a good question and we are trying to find a good answer in #636.

Why am I seeing idle borg serve processes on the repo server?

Maybe the ssh connection between client and server broke down and that was not yet noticed on the server. Try these settings:

# /etc/ssh/sshd_config on borg repo server - kill connection to client
# after ClientAliveCountMax * ClientAliveInterval seconds with no response
ClientAliveInterval 20
ClientAliveCountMax 3

If you have multiple borg create ... ; borg create ... commands in a already serialized way in a single script, you need to give them --lock-wait N (with N being a bit more than the time the server needs to terminate broken down connections and release the lock).

The borg cache eats way too much disk space, what can I do?

This may especially happen if borg needs to rebuild the local "chunks" index - either because it was removed, or because it was not coherent with the repository state any more (e.g. because another borg instance changed the repository).

To optimize this rebuild process, borg caches per-archive information in the chunks.archive.d/ directory. It won't help the first time it happens, but it will make the subsequent rebuilds faster (because it needs to transfer less data from the repository). While being faster, the cache needs quite some disk space, which might be unwanted.

There is a temporary (but maybe long lived) hack to avoid using lots of disk space for chunks.archive.d (see #235 for details):

# this assumes you are working with the same user as the backup.
cd ~/.cache/borg/$(borg config /path/to/repo id)
rm -rf chunks.archive.d ; touch chunks.archive.d

This deletes all the cached archive chunk indexes and replaces the directory that kept them with a file, so borg won't be able to store anything "in" there in future.

This has some pros and cons, though:

  • much less disk space needs for ~/.cache/borg.

  • chunk cache resyncs will be slower as it will have to transfer chunk usage metadata for all archives from the repository (which might be slow if your repo connection is slow) and it will also have to build the hashtables from that data. chunk cache resyncs happen e.g. if your repo was written to by another machine (if you share same backup repo between multiple machines) or if your local chunks cache was lost somehow.

The long term plan to improve this is called "borgception", see #474.

Can I backup my root partition (/) with Borg?

Backing up your entire root partition works just fine, but remember to exclude directories that make no sense to backup, such as /dev, /proc, /sys, /tmp and /run, and to use --one-file-system if you only want to backup the root partition (and not any mounted devices e.g.).

If it crashes with a UnicodeError, what can I do?

Check if your encoding is set correctly. For most POSIX-like systems, try:

export LANG=en_US.UTF-8  # or similar, important is correct charset

If that does not help:

  • check for typos, check if you really used export.

  • check if you have set LC_ALL - if so, try not setting it.

  • check if you generated the respective locale via locale-gen.

I can't extract non-ascii filenames by giving them on the commandline!?

This might be due to different ways to represent some characters in unicode or due to other non-ascii encoding issues.

If you run into that, try this:

  • avoid the non-ascii characters on the commandline by e.g. extracting the parent directory (or even everything)

  • mount the repo using FUSE and use some file manager

What's the expected backup performance?

A first backup will usually be somehow "slow" because there is a lot of data to process. Performance here depends on a lot of factors, so it is hard to give specific numbers.

Subsequent backups are usually very fast if most files are unchanged and only a few are new or modified. The high performance on unchanged files primarily depends only on a few factors (like fs recursion + metadata reading performance and the files cache working as expected) and much less on other factors.

E.g., for this setup:

  • server grade machine (4C/8T 2013 Xeon, 64GB RAM, 2x good 7200RPM disks)

  • local zfs filesystem (mirrored) containing the backup source data

  • repository is remote (does not matter much for unchanged files)

  • backup job runs while machine is otherwise idle

The observed performance is that Borg can process about 1 million unchanged files (and a few small changed ones) in 4 minutes!

If you are seeing much less than that in similar circumstances, read the next few FAQ entries below.

Why is backup slow for me?

So, if you feel your Borg backup is too slow somehow, you should find out why.

The usual way to approach this is to add --list --filter=AME --stats to your borg create call to produce more log output, including a file list (with file status characters) and also some statistics at the end of the backup.

Then you do the backup and look at the log output:

  • stats: Do you really have little changes or are there more changes than you thought? In the stats you can see the overall volume of changed data, which needed to be added to the repo. If that is a lot, that can be the reason why it is slow.

  • A status ("added") in the file list: If you see that often, you have a lot of new files (files that Borg did not find in the files cache). If you think there is something wrong with that (the file was there already in the previous backup), please read the FAQ entries below.

  • M status ("modified") in the file list: If you see that often, Borg thinks that a lot of your files might be modified (Borg found them in the files cache, but the metadata read from the filesystem did not match the metadata stored in the files cache). In such a case, Borg will need to process the files' contents completely, which is much slower than processing unmodified files (Borg does not read their contents!). The metadata values used in this comparison are determined by the --files-cache option and could be e.g. size, ctime and inode number (see the borg create docs for more details and potential issues). You can use the stat command on files to manually look at fs metadata to debug if there is any unexpected change triggering the M status. Also, the --debug-topic=files_cache option of borg create provides a lot of debug output helping to analyse why the files cache does not give its expected high performance.

When borg runs inside a virtual machine, there are some more things to look at:

Some hypervisors (e.g. kvm on proxmox) give some broadly compatible CPU type to the VM (usually to ease migration between VM hosts of potentially different hardware CPUs).

It is broadly compatible because they leave away modern CPU features that could be not present in older or other CPUs, e.g. hardware acceleration for AES crypto, for sha2 hashes, for (P)CLMUL(QDQ) computations useful for crc32.

So, basically you pay for compatibility with bad performance. If you prefer better performance, you should try to expose the host CPU's misc. hw acceleration features to the VM which runs borg.

On Linux, check /proc/cpuinfo for the CPU flags inside the VM. For kvm check the docs about "Host model" and "Host passthrough".

See also the next few FAQ entries for more details.

I am seeing 'A' (added) status for an unchanged file!?

The files cache is used to determine whether Borg already "knows" / has backed up a file and if so, to skip the file from chunking. It intentionally excludes files that have a timestamp which is the same as the newest timestamp in the created archive.

So, if you see an 'A' status for unchanged file(s), they are likely the files with the most recent timestamp in that archive.

This is expected: it is to avoid data loss with files that are backed up from a snapshot and that are immediately changed after the snapshot (but within timestamp granularity time, so the timestamp would not change). Without the code that removes these files from the files cache, the change that happened right after the snapshot would not be contained in the next backup as Borg would think the file is unchanged.

This does not affect deduplication, the file will be chunked, but as the chunks will often be the same and already stored in the repo (except in the above mentioned rare condition), it will just re-use them as usual and not store new data chunks.

If you want to avoid unnecessary chunking, just create or touch a small or empty file in your backup source file set (so that one has the latest timestamp, not your 50GB VM disk image) and, if you do snapshots, do the snapshot after that.

Since only the files cache is used in the display of files status, those files are reported as being added when, really, chunks are already used.

By default, ctime (change time) is used for the timestamps to have a rather safe change detection (see also the --files-cache option).

Furthermore, pathnames recorded in files cache are always absolute, even if you specify source directories with relative pathname. If relative pathnames are stable, but absolute are not (for example if you mount a filesystem without stable mount points for each backup or if you are running the backup from a filesystem snapshot whose name is not stable), borg will assume that files are different and will report them as 'added', even though no new chunks will be actually recorded for them. To avoid this, you could bind mount your source directory in a directory with the stable path.

It always chunks all my files, even unchanged ones!

Borg maintains a files cache where it remembers the timestamp, size and inode of files. When Borg does a new backup and starts processing a file, it first looks whether the file has changed (compared to the values stored in the files cache). If the values are the same, the file is assumed unchanged and thus its contents won't get chunked (again).

Borg can't keep an infinite history of files of course, thus entries in the files cache have a "maximum time to live" which is set via the environment variable BORG_FILES_CACHE_TTL (and defaults to 20). Every time you do a backup (on the same machine, using the same user), the cache entries' ttl values of files that were not "seen" are incremented by 1 and if they reach BORG_FILES_CACHE_TTL, the entry is removed from the cache.

So, for example, if you do daily backups of 26 different data sets A, B, C, ..., Z on one machine (using the default TTL), the files from A will be already forgotten when you repeat the same backups on the next day and it will be slow because it would chunk all the files each time. If you set BORG_FILES_CACHE_TTL to at least 26 (or maybe even a small multiple of that), it would be much faster.

Besides using a higher BORG_FILES_CACHE_TTL (which also increases memory usage), there is also BORG_FILES_CACHE_SUFFIX which can be used to have separate (smaller) files caches for each backup set instead of the default one (big) unified files cache.

Another possible reason is that files don't always have the same path, for example if you mount a filesystem without stable mount points for each backup or if you are running the backup from a filesystem snapshot whose name is not stable. If the directory where you mount a filesystem is different every time, Borg assumes they are different files. This is true even if you backup these files with relative pathnames - borg uses full pathnames in files cache regardless.

It is possible for some filesystems, such as mergerfs or network filesystems, to return inconsistent inode numbers across runs, causing borg to consider them changed. A workaround is to set the option --files-cache=ctime,size to exclude the inode number comparison from the files cache check so that files with different inode numbers won't be treated as modified.

Is there a way to limit bandwidth with Borg?

To limit upload (i.e. borg create) bandwidth, use the --remote-ratelimit option.

There is no built-in way to limit download (i.e. borg extract) bandwidth, but limiting download bandwidth can be accomplished with pipeviewer:

Create a wrapper script: /usr/local/bin/pv-wrapper

#!/bin/sh
    ## -q, --quiet              do not output any transfer information at all
    ## -L, --rate-limit RATE    limit transfer to RATE bytes per second
RATE=307200
pv -q -L $RATE  | "$@"

Add BORG_RSH environment variable to use pipeviewer wrapper script with ssh.

export BORG_RSH='/usr/local/bin/pv-wrapper ssh'

Now Borg will be bandwidth limited. The nice thing about pv is that you can change rate-limit on the fly:

pv -R $(pidof pv) -L 102400
How can I avoid unwanted base directories getting stored into archives?

Possible use cases:

  • Another file system is mounted and you want to backup it with original paths.

  • You have created a BTRFS snapshot in a /.snapshots directory for backup.

To achieve this, run borg create within the mountpoint/snapshot directory:

# Example: Some file system mounted in /mnt/rootfs.
cd /mnt/rootfs
borg create /path/to/repo::rootfs_backup .
I am having troubles with some network/FUSE/special filesystem, why?

Borg is doing nothing special in the filesystem, it only uses very common and compatible operations (even the locking is just "rename").

So, if you are encountering issues like slowness, corruption or malfunction when using a specific filesystem, please try if you can reproduce the issues with a local (non-network) and proven filesystem (like ext4 on Linux).

If you can't reproduce the issue then, you maybe have found an issue within the filesystem code you used (not with Borg). For this case, it is recommended that you talk to the developers / support of the network fs and maybe open an issue in their issue tracker. Do not file an issue in the Borg issue tracker.

If you can reproduce the issue with the proven filesystem, please file an issue in the Borg issue tracker about that.

Why does running 'borg check --repair' warn about data loss?

Repair usually works for recovering data in a corrupted archive. However, it's impossible to predict all modes of corruption. In some very rare instances, such as malfunctioning storage hardware, additional repo corruption may occur. If you can't afford to lose the repo, it's strongly recommended that you perform repair on a copy of the repo.

In other words, the warning is there to emphasize that Borg:
  • Will perform automated routines that modify your backup repository

  • Might not actually fix the problem you are experiencing

  • Might, in very rare cases, further corrupt your repository

In the case of malfunctioning hardware, such as a drive or USB hub corrupting data when read or written, it's best to diagnose and fix the cause of the initial corruption before attempting to repair the repo. If the corruption is caused by a one time event such as a power outage, running borg check --repair will fix most problems.

Why isn't there more progress / ETA information displayed?

Some borg runs take quite a bit, so it would be nice to see a progress display, maybe even including a ETA (expected time of "arrival" [here rather "completion"]).

For some functionality, this can be done: if the total amount of work is more or less known, we can display progress. So check if there is a --progress option.

But sometimes, the total amount is unknown (e.g. for borg create we just do a single pass over the filesystem, so we do not know the total file count or data volume before reaching the end). Adding another pass just to determine that would take additional time and could be incorrect, if the filesystem is changing.

Even if the fs does not change and we knew count and size of all files, we still could not compute the borg create ETA as we do not know the amount of changed chunks, how the bandwidth of source and destination or system performance might fluctuate.

You see, trying to display ETA would be futile. The borg developers prefer to rather not implement progress / ETA display than doing futile attempts.

See also: https://xkcd.com/612/

Why am I getting 'Operation not permitted' errors when backing up on sshfs?

By default, sshfs is not entirely POSIX-compliant when renaming files due to a technicality in the SFTP protocol. Fortunately, it also provides a workaround to make it behave correctly:

sshfs -o workaround=rename user@host:dir /mnt/dir
Can I disable checking for free disk space?

In some cases, the free disk space of the target volume is reported incorrectly. This can happen for CIFS- or FUSE shares. If you are sure that your target volume will always have enough disk space, you can use the following workaround to disable checking for free disk space:

borg config -- $REPO_LOCATION additional_free_space -2T
How do I rename a repository?

There is nothing special that needs to be done, you can simply rename the directory that corresponds to the repository. However, the next time borg interacts with the repository (i.e, via borg list), depending on the value of BORG_RELOCATED_REPO_ACCESS_IS_OK, borg may warn you that the repository has been moved. You will be given a prompt to confirm you are OK with this.

If BORG_RELOCATED_REPO_ACCESS_IS_OK is unset, borg will interactively ask for each repository whether it's OK.

It may be useful to set BORG_RELOCATED_REPO_ACCESS_IS_OK=yes to avoid the prompts when renaming multiple repositories or in a non-interactive context such as a script. See Deployment for an example.

The repository quota size is reached, what can I do?

The simplest solution is to increase or disable the quota and resume the backup:

borg config /path/to/repo storage_quota 0

If you are bound to the quota, you have to free repository space. The first to try is running borg compact to free unused backup space (see also Separate compaction):

borg compact /path/to/repo

If your repository is already compacted, run borg prune or borg delete to delete archives that you do not need anymore, and then run borg compact again.

My backup disk is full, what can I do?

Borg cannot work if you really have zero free space on the backup disk, so the first thing you must do is deleting some files to regain free disk space. See Important note about free space for further details.

Some Borg commands that do not change the repository might work under disk-full conditions, but generally this should be avoided. If your backup disk is already full when Borg starts a write command like borg create, it will abort immediately and the repository will stay as-is.

If you run a backup that stops due to a disk running full, Borg will roll back, delete the new new segment file and thus freeing disk space automatically. There may be a checkpoint archive left that has been saved before the disk got full. You can keep it to speed up the next backup or delete it to get back more disk space.

Miscellaneous

macOS: borg mounts not shown in Finder's side bar

https://github.com/osxfuse/osxfuse/wiki/Mount-options#local

Read the above first and use this on your own risk:

borg mount -olocal REPO MOUNTPOINT
Requirements for the borg single-file binary, esp. (g)libc?

We try to build the binary on old, but still supported systems - to keep the minimum requirement for the (g)libc low. The (g)libc can't be bundled into the binary as it needs to fit your kernel and OS, but Python and all other required libraries will be bundled into the binary.

If your system fulfills the minimum (g)libc requirement (see the README that is released with the binary), there should be no problem. If you are slightly below the required version, maybe just try. Due to the dynamic loading (or not loading) of some shared libraries, it might still work depending on what libraries are actually loaded and used.

In the borg git repository, there is scripts/glibc_check.py that can determine (based on the symbols' versions they want to link to) whether a set of given (Linux) binaries works with a given glibc version.

Why was Borg forked from Attic?

Borg was created in May 2015 in response to the difficulty of getting new code or larger changes incorporated into Attic and establishing a bigger developer community / more open development.

More details can be found in ticket 217 that led to the fork.

Borg intends to be:

  • simple:

    • as simple as possible, but no simpler

    • do the right thing by default, but offer options

  • open:

    • welcome feature requests

    • accept pull requests of good quality and coding style

    • give feedback on PRs that can't be accepted "as is"

    • discuss openly, don't work in the dark

  • changing:

    • Borg is not compatible with Attic

    • do not break compatibility accidentally, without a good reason or without warning. allow compatibility breaking for other cases.

    • if major version number changes, it may have incompatible changes

Migrating from Attic

What are the differences between Attic and Borg?

Borg is a fork of Attic and maintained by "The Borg collective".

Here's a (incomplete) list of some major changes:

  • lots of attic issues fixed (see issue #5), including critical data corruption bugs and security issues.

  • more open, faster paced development (see issue #1)

  • less chunk management overhead (less memory and disk usage for chunks index)

  • faster remote cache resync (useful when backing up multiple machines into same repo)

  • compression: no, lz4, zstd, zlib or lzma compression, adjustable compression levels

  • repokey replaces problematic passphrase mode (you can't change the passphrase nor the pbkdf2 iteration count in "passphrase" mode)

  • simple sparse file support, great for virtual machine disk files

  • can read special files (e.g. block devices) or from stdin, write to stdout

  • rename-based locking is more compatible than attic's posix locking

  • uses fadvise to not spoil / blow up the fs cache

  • better error messages / exception handling

  • better logging, screen output, progress indication

  • tested on misc. Linux systems, 32 and 64bit, FreeBSD, OpenBSD, NetBSD, macOS

Please read the Change Log (or docs/changes.rst in the source distribution) for more information.

Borg is not compatible with original Attic (but there is a one-way conversion).

How do I migrate from Attic to Borg?

Use borg upgrade. This is a one-way process that cannot be reversed.

There are some caveats:

  • The upgrade can only be performed on local repositories. It cannot be performed on remote repositories.

  • If the repository is in "keyfile" encryption mode, the keyfile must exist locally or it must be manually moved after performing the upgrade:

    1. Get the repository ID with borg config /path/to/repo id.

    2. Locate the attic key file at ~/.attic/keys/. The correct key for the repository starts with the line ATTIC_KEY <repository id>.

    3. Copy the attic key file to ~/.config/borg/keys/

    4. Change the first line from ATTIC_KEY ... to BORG_KEY ....

    5. Verify that the repository is now accessible (e.g. borg list <repository>).

  • Attic and Borg use different "chunker params". This means that data added by Borg won't deduplicate with the existing data stored by Attic. The effect is lessened if the files cache is used with Borg.

  • Repositories in "passphrase" mode must be migrated to "repokey" mode using Upgrading a passphrase encrypted attic repo. Borg does not support the "passphrase" mode any other way.

Why is my backup bigger than with attic?

Attic was rather unflexible when it comes to compression, it always compressed using zlib level 6 (no way to switch compression off or adjust the level or algorithm).

The default in Borg is lz4, which is fast enough to not use significant CPU time in most cases, but can only achieve modest compression. It still compresses easily compressed data fairly well.

Borg also offers zstd, zlib and lzma compression, choose wisely.

Which choice is the best option depends on a number of factors, like bandwidth to the repository, how well the data compresses, available CPU power and so on.

Support

Support and Services

Please see https://www.borgbackup.org/ for free and paid support and service options.

Security

In case you discover a security issue, please use this contact for reporting it privately and please, if possible, use encrypted E-Mail:

Thomas Waldmann <tw@waldmann-edv.de>

GPG Key Fingerprint: 6D5B EF9A DD20 7580 5747 B70F 9F88 FB52 FAF7 B393

The public key can be fetched from any GPG keyserver, but be careful: you must use the full fingerprint to check that you got the correct key.

Verifying signed releases

Releases are signed with the same GPG key and a .asc file is provided for each binary.

To verify a signature, the public key needs to be known to GPG. It can be imported into the local keystore from a keyserver with the fingerprint:

gpg --recv-keys "6D5B EF9A DD20 7580 5747 B70F 9F88 FB52 FAF7 B393"

If GPG successfully imported the key, the output should include (among other things):

...
gpg: Total number processed: 1
...

To verify for example the signature of the borg-linux64 binary:

gpg --verify borg-linux64.asc

GPG outputs if it finds a good signature. The output should look similar to this:

gpg: Signature made Sat 30 Dec 2017 01:07:36 PM CET using RSA key ID 51F78E01
gpg: Good signature from "Thomas Waldmann <email>"
gpg: aka "Thomas Waldmann <email>"
gpg: aka "Thomas Waldmann <email>"
gpg: aka "Thomas Waldmann <email>"
gpg: WARNING: This key is not certified with a trusted signature!
gpg: There is no indication that the signature belongs to the owner.
Primary key fingerprint: 6D5B EF9A DD20 7580 5747 B70F 9F88 FB52 FAF7 B393
Subkey fingerprint: 2F81 AFFB AB04 E11F E8EE 65D4 243A CFA9 51F7 8E01

If you want to make absolutely sure that you have the right key, you need to verify it via another channel and assign a trust-level to it.

Important notes

This section provides information about security and corruption issues.

Pre-1.2.5 archives spoofing vulnerability (CVE-2023-36811)

A flaw in the cryptographic authentication scheme in Borg allowed an attacker to fake archives and potentially indirectly cause backup data loss in the repository.

The attack requires an attacker to be able to

  1. insert files (with no additional headers) into backups

  2. gain write access to the repository

This vulnerability does not disclose plaintext to the attacker, nor does it affect the authenticity of existing archives.

Creating plausible fake archives may be feasible for empty or small archives, but is unlikely for large archives.

The fix enforces checking the TAM authentication tag of archives at critical places. Borg now considers archives without TAM as garbage or an attack.

We are not aware of others having discovered, disclosed or exploited this vulnerability.

Below, if we speak of borg 1.2.6, we mean a borg version >= 1.2.6 or a borg version that has the relevant security patches for this vulnerability applied (could be also an older version in that case).

Steps you must take to upgrade a repository (this applies to all kinds of repos no matter what encryption mode they use, including "none"):

  1. Upgrade all clients using this repository to borg 1.2.6. Note: it is not required to upgrade a server, except if the server-side borg is also used as a client (and not just for "borg serve").

    Do not run borg check with borg 1.2.6 before completing the upgrade steps:

    • borg check would complain about archives without a valid archive TAM.

    • borg check --repair would remove such archives!

  2. Run: BORG_WORKAROUNDS=ignore_invalid_archive_tam borg info --debug <repo> 2>&1 | grep TAM | grep -i manifest

    1. If you get "TAM-verified manifest", continue with 3.

    2. If you get "Manifest TAM not found and not required", run borg upgrade --tam --force <repository> on every client.

  3. Run: BORG_WORKAROUNDS=ignore_invalid_archive_tam borg list --consider-checkpoints --format='{name} {time} tam:{tam}{NL}' <repo>

    "tam:verified" means that the archive has a valid TAM authentication. "tam:none" is expected as output for archives created by borg <1.0.9. "tam:none" is also expected for archives resulting from a borg rename or borg recreate operation (see #7791). "tam:none" could also come from archives created by an attacker. You should verify that "tam:none" archives are authentic and not malicious (== have good content, have correct timestamp, can be extracted successfully). In case you find crappy/malicious archives, you must delete them before proceeding. In low-risk, trusted environments, you may decide on your own risk to skip step 3 and just trust in everything being OK.

  4. If there are no tam:none archives left at this point, you can skip this step. Run BORG_WORKAROUNDS=ignore_invalid_archive_tam borg upgrade --archives-tam <repo>. This will unconditionally add a correct archive TAM to all archives not having one. borg check would consider TAM-less or invalid-TAM archives as garbage or a potential attack. To see that all archives now are "tam:verified" run: borg list --consider-checkpoints --format='{name} {time} tam:{tam}{NL}' <repo>

  5. Please note that you should never use BORG_WORKAROUNDS=ignore_invalid_archive_tam for normal production operations - it is only needed once to get the archives in a repository into a good state. All archives have a valid TAM now.

Vulnerability time line:

  • 2023-06-13: Vulnerability discovered during code review by Thomas Waldmann

  • 2023-06-13...: Work on fixing the issue, upgrade procedure, docs.

  • 2023-06-30: CVE was assigned via Github CNA

  • 2023-06-30 .. 2023-08-29: Fixed issue, code review, docs, testing.

  • 2023-08-30: Released fixed version 1.2.5 (broken upgrade procedure for some repos)

  • 2023-08-31: Released fixed version 1.2.6 (fixes upgrade procedure)

Pre-1.1.11 potential index corruption / data loss issue

A bug was discovered in our hashtable code, see issue #4829. The code is used for the client-side chunks cache and the server-side repo index.

Although borg uses the hashtables very heavily, the index corruption did not happen too frequently, because it needed specific conditions to happen.

Data loss required even more specific conditions, so it should be rare (and also detectable via borg check).

You might be affected if borg crashed with / complained about:

  • AssertionError: Corrupted segment reference count - corrupted index or hints

  • ObjectNotFound: Object with key ... not found in repository ...

  • Index mismatch for key b'...'. (..., ...) != (-1, -1)

  • ValueError: stats_against: key contained in self but not in master_index.

Advised procedure to fix any related issue in your indexes/caches:

  • install fixed borg code (on client AND server)

  • for all of your clients and repos remove the cache by:

    borg delete --cache-only YOURREPO

    (later, the cache will be re-built automatically)

  • for all your repos, rebuild the repo index by:

    borg check --repair YOURREPO

    This will also check all archives and detect if there is any data-loss issue.

Affected branches / releases:

  • fd06497 introduced the bug into 1.1-maint branch - it affects all borg 1.1.x since 1.1.0b4.

  • fd06497 introduced the bug into master branch - it affects all borg 1.2.0 alpha releases.

  • c5cd882 introduced the bug into 1.0-maint branch - it affects all borg 1.0.x since 1.0.11rc1.

The bug was fixed by:

  • 701159a fixes the bug in 1.1-maint branch - will be released with borg 1.1.11.

  • fa63150 fixes the bug in master branch - will be released with borg 1.2.0a8.

  • 7bb90b6 fixes the bug in 1.0-maint branch. Branch is EOL, no new release is planned as of now.

Pre-1.1.4 potential data corruption issue

A data corruption bug was discovered in borg check --repair, see issue #3444.

This is a 1.1.x regression, releases < 1.1 (e.g. 1.0.x) are not affected.

To avoid data loss, you must not run borg check --repair using an unfixed version of borg 1.1.x. The first official release that has the fix is 1.1.4.

Package maintainers may have applied the fix to updated packages of 1.1.x (x<4) though, see the package maintainer's package changelog to make sure.

If you never had missing item metadata chunks, the bug has not affected you even if you did run borg check --repair with an unfixed version.

When borg check --repair tried to repair corrupt archives that miss item metadata chunks, the resync to valid metadata in still present item metadata chunks malfunctioned. This was due to a broken validator that considered all (even valid) item metadata as invalid. As they were considered invalid, borg discarded them. Practically, that means the affected files, directories or other fs objects were discarded from the archive.

Due to the malfunction, the process was extremely slow, but if you let it complete, borg would have created a "repaired" archive that has lost a lot of items. If you interrupted borg check --repair because it was so strangely slow (killing borg somehow, e.g. Ctrl-C) the transaction was rolled back and no corruption occurred.

The log message indicating the precondition for the bug triggering looks like:

item metadata chunk missing [chunk: 001056_bdee87d...a3e50d]

If you never had that in your borg check --repair runs, you're not affected.

But if you're unsure or you actually have seen that, better check your archives. By just using "borg list repo::archive" you can see if all expected filesystem items are listed.

Pre-1.0.9 manifest spoofing vulnerability (CVE-2016-10099)

A flaw in the cryptographic authentication scheme in Borg allowed an attacker to spoof the manifest. The attack requires an attacker to be able to

  1. insert files (with no additional headers) into backups

  2. gain write access to the repository

This vulnerability does not disclose plaintext to the attacker, nor does it affect the authenticity of existing archives.

The vulnerability allows an attacker to create a spoofed manifest (the list of archives). Creating plausible fake archives may be feasible for small archives, but is unlikely for large archives.

The fix adds a separate authentication tag to the manifest. For compatibility with prior versions this authentication tag is not required by default for existing repositories. Repositories created with 1.0.9 and later require it.

Steps you should take:

  1. Upgrade all clients to 1.0.9 or later.

  2. Run borg upgrade --tam <repository> on every client for each repository.

  3. This will list all archives, including archive IDs, for easy comparison with your logs.

  4. Done.

Prior versions can access and modify repositories with this measure enabled, however, to 1.0.9 or later their modifications are indiscernible from an attack and will raise an error until the below procedure is followed. We are aware that this can be annoying in some circumstances, but don't see a way to fix the vulnerability otherwise.

In case a version prior to 1.0.9 is used to modify a repository where above procedure was completed, and now you get an error message from other clients:

  1. borg upgrade --tam --force <repository> once with any client suffices.

This attack is mitigated by:

  • Noting/logging borg list, borg info, or borg create --stats, which contain the archive IDs.

We are not aware of others having discovered, disclosed or exploited this vulnerability.

Vulnerability time line:

  • 2016-11-14: Vulnerability and fix discovered during review of cryptography by Marian Beermann (@enkore)

  • 2016-11-20: First patch

  • 2016-12-20: Released fixed version 1.0.9

  • 2017-01-02: CVE was assigned

  • 2017-01-15: Released fixed version 1.1.0b3 (fix was previously only available from source)

Pre-1.0.9 potential data loss

If you have archives in your repository that were made with attic <= 0.13 (and later migrated to borg), running borg check would report errors in these archives. See issue #1837.

The reason for this is a invalid (and useless) metadata key that was always added due to a bug in these old attic versions.

If you run borg check --repair, things escalate quickly: all archive items with invalid metadata will be killed. Due to that attic bug, that means all items in all archives made with these old attic versions.

Pre-1.0.4 potential repo corruption

Some external errors (like network or disk I/O errors) could lead to corruption of the backup repository due to issue #1138.

A sign that this happened is if "E" status was reported for a file that can not be explained by problems with the source file. If you still have logs from "borg create -v --list", you can check for "E" status.

Here is what could cause corruption and what you can do now:

  1. I/O errors (e.g. repo disk errors) while writing data to repo.

This could lead to corrupted segment files.

Fix:

# check for corrupt chunks / segments:
borg check -v --repository-only REPO

# repair the repo:
borg check -v --repository-only --repair REPO

# make sure everything is fixed:
borg check -v --repository-only REPO
  1. Unreliable network / unreliable connection to the repo.

This could lead to archive metadata corruption.

Fix:

# check for corrupt archives:
borg check -v --archives-only REPO

# delete the corrupt archives:
borg delete --force REPO::CORRUPT_ARCHIVE

# make sure everything is fixed:
borg check -v --archives-only REPO
  1. In case you want to do more intensive checking.

The best check that everything is ok is to run a dry-run extraction:

borg extract -v --dry-run REPO::ARCHIVE

Upgrade Notes

borg 1.1.x to 1.2.x

Some things can be recommended for the upgrade process from borg 1.1.x (please also read the important compatibility notes below):

  • first upgrade to a recent 1.1.x release - especially if you run some older 1.1.* or even 1.0.* borg release.

  • using that, run at least one borg create (your normal backup), prune and especially a check to see everything is in a good state.

  • check the output of borg check - if there is anything special, consider a borg check --repair followed by another borg check.

  • if everything is fine so far (borg check reports no issues), you can consider upgrading to 1.2.x. if not, please first fix any already existing issue.

  • if you want to play safer, first create a backup of your borg repository.

  • upgrade to latest borg 1.2.x release (you could use the fat binary from github releases page)

  • borg 1.2.6 has a security fix for the pre-1.2.5 archives spoofing vulnerability (CVE-2023-36811), see details and necessary upgrade procedure described above.

  • run borg compact --cleanup-commits to clean up a ton of 17 bytes long files in your repo caused by a borg 1.1 bug

  • run borg check again (now with borg 1.2.x) and check if there is anything special.

  • run borg info (with borg 1.2.x) to build the local pre12-meta cache (can take significant time, but after that it will be fast) - for more details see below.

  • check the compatibility notes (see below) and adapt your scripts, if needed.

  • if you run into any issues, please check the github issue tracker before posting new issues there or elsewhere.

If you follow this procedure, you can help avoiding that we get a lot of "borg 1.2" issue reports that are not really 1.2 issues, but existed before and maybe just were not noticed.

Compatibility notes:

  • matching of path patterns has been aligned with borg storing relative paths. Borg archives file paths without leading slashes. Previously, include/exclude patterns could contain leading slashes. You should check your patterns and remove leading slashes.

  • dropped support / testing for older Pythons, minimum requirement is 3.8. In case your OS does not provide Python >= 3.8, consider using our binary, which does not need an external Python interpreter. Or continue using borg 1.1.x, which is still supported.

  • freeing repository space only happens when "borg compact" is invoked.

  • mount: the default for --numeric-ids is False now (same as borg extract)

  • borg create --noatime is deprecated. Not storing atime is the default behaviour now (use --atime if you want to store the atime).

  • --prefix is deprecated, use -a / --glob-archives, see #6806

  • list: corrected mix-up of "isomtime" and "mtime" formats. Previously, "isomtime" was the default but produced a verbose human format, while "mtime" produced a ISO-8601-like format. The behaviours have been swapped (so "mtime" is human, "isomtime" is ISO-like), and the default is now "mtime". "isomtime" is now a real ISO-8601 format ("T" between date and time, not a space).

  • create/recreate --list: file status for all files used to get announced AFTER the file (with borg < 1.2). Now, file status is announced BEFORE the file contents are processed. If the file status changes later (e.g. due to an error or a content change), the updated/final file status will be printed again.

  • removed deprecated-since-long stuff (deprecated since):

    • command "borg change-passphrase" (2017-02), use "borg key ..."

    • option "--keep-tag-files" (2017-01), use "--keep-exclude-tags"

    • option "--list-format" (2017-10), use "--format"

    • option "--ignore-inode" (2017-09), use "--files-cache" w/o "inode"

    • option "--no-files-cache" (2017-09), use "--files-cache=disabled"

  • removed BORG_HOSTNAME_IS_UNIQUE env var. to use borg you must implement one of these 2 scenarios:

      1. the combination of FQDN and result of uuid.getnode() must be unique and stable (this should be the case for almost everybody, except when having duplicate FQDN and MAC address or all-zero MAC address)

      1. if you are aware that 1) is not the case for you, you must set BORG_HOST_ID env var to something unique.

  • exit with 128 + signal number, #5161. if you have scripts expecting rc == 2 for a signal exit, you need to update them to check for >= 128.

Change Log

Version 1.2.7 (2023-12-02)

For upgrade and compatibility hints, please also read the section "Upgrade Notes" above.

Fixes:

  • docs: CVE-2023-36811 upgrade steps: consider checkpoint archives, #7802

  • check/compact: fix spurious reappearance of orphan chunks since borg 1.2, #6687 - this consists of 2 fixes:

    • for existing chunks: check --repair: recreate shadow index, #6687

    • for newly created chunks: update shadow index when doing a double-put, #5661

  • LockRoster.modify: no KeyError if element was already gone, #7937

  • create --X-from-command: run subcommands with a clean environment, #7916

  • list --sort-by: support "archive" as alias of "name", #7873

  • fix rc and msg if arg parsing throws an exception, #7885

Other changes:

  • support and test on Python 3.12

  • include unistd.h in _chunker.c (fix for Python 3.13)

  • allow msgpack 1.0.6 and 1.0.7

  • TAM issues: show tracebacks, improve borg check logging, #7797

  • replace "datetime.utcfromtimestamp" with custom helper to avoid deprecation warnings when using Python 3.12

  • vagrant:

    • use generic/debian9 box, fixes #7579

    • add VM with debian bookworm / test on OpenSSL 3.0.x.

  • docs:

    • not only attack/unsafe, can also be a fs issue, #7853

    • point to CVE-2023-36811 upgrade steps from borg 1.1 to 1.2 upgrade steps, #7899

    • upgrade steps needed for all kinds of repos (including "none" encryption mode), #7813

    • upgrade steps: talk about consequences of borg check, #7816

    • upgrade steps: remove period that could be interpreted as part of the command

    • automated-local.rst: use GPT UUID for consistent udev rule

    • create disk/partition sector backup by disk serial number, #7934

    • update macOS hint about full disk access

    • clarify borg prune -a option description, #7871

    • readthedocs: also build offline docs (HTMLzip), #7835

    • frontends: add "check.rebuild_refcounts" message

Version 1.2.6 (2023-08-31)

Fixes:

  • The upgrade procedure docs as published with borg 1.2.5 did not work, if the repository had archives resulting from a borg rename or borg recreate operation.

    The updated docs now use BORG_WORKAROUNDS=ignore_invalid_archive_tam at some places to avoid that issue, #7791.

    See: fix pre-1.2.5 archives spoofing vulnerability (CVE-2023-36811), details and necessary upgrade procedure described above.

Other changes:

  • updated 1.2.5 changelog entry: 1.2.5 already has the fix for rename/recreate.

  • remove cython restrictions. recommended is to build with cython 0.29.latest, because borg 1.2.x uses this since years and it is very stable. you can also try to build with cython 3.0.x, there is a good chance that it works. as a 3rd option, we also bundle the *.c files cython outputs in the release pypi package, so you can also just use these and not need cython at all.

Version 1.2.5 (2023-08-30)

Fixes:

  • Security: fix pre-1.2.5 archives spoofing vulnerability (CVE-2023-36811), see details and necessary upgrade procedure described above.

  • rename/recreate: correctly update resulting archive's TAM, see #7791

  • create: do not try to read parent dir of recursion root, #7746

  • extract: fix false warning about pattern never matching, #4110

  • diff: remove surrogates before output, #7535

  • compact: clear empty directories at end of compact process, #6823

  • create --files-cache=size: fix crash, #7658

  • keyfiles: improve key sanity check, #7561

  • only warn about "invalid" chunker params, #7590

  • ProgressIndicatorPercent: fix space computation for wide chars, #3027

  • improve argparse validator error messages

New features:

  • mount: make up volname if not given (macOS), #7690. macFUSE supports a volname mount option to give what finder displays on the desktop / in the directory view. if the user did not specify it, we make something up, because otherwise it would be "macFUSE Volume 0 (Python)" and hide the mountpoint directory name.

  • BORG_WORKAROUNDS=authenticated_no_key to extract from authenticated repos without key, #7700

Other changes:

  • add utcnow() helper function to avoid deprecated datetime.utcnow()

  • stay on latest Cython 0.29 (0.29.36) for borg 1.2.x (do not use Cython 3.0 yet)

  • docs:

    • move upgrade notes to own section, see #7546

    • mount -olocal: how to show mount in finder's sidebar, #5321

    • list: fix --pattern examples, #7611

    • improve patterns help

    • incl./excl. options, path-from-stdin exclusiveness

    • obfuscation docs: markup fix, note about MAX_DATA_SIZE

    • --one-file-system: add macOS apfs notes, #4876

    • improve --one-file-system help string, #5618

    • rewrite borg check docs

    • improve the docs for --keep-within, #7687

    • fix borg init command in environment.rst.inc

    • 1.1.x upgrade notes: more precise borg upgrade instructions, #3396

  • tests:

    • fix repo reopen

    • avoid long ids in pytest output

    • check buzhash chunksize distribution, see #7586

Version 1.2.4 (2023-03-24)

New features:

  • import-tar: add --ignore-zeros to process concatenated tars, #7432.

  • debug id-hash: computes file/chunk content id-hash, #7406

  • diff: --content-only does not show mode/ctime/mtime changes, #7248

  • diff: JSON strings in diff output are now sorted alphabetically

Bug fixes:

  • xattrs: fix namespace processing on FreeBSD, #6997

  • diff: fix path related bug seen when addressing deferred items.

  • debug get-obj/put-obj: always give chunkid as cli param, see #7290 (this is an incompatible change, see also borg debug id-hash)

  • extract: fix mtime when ResourceFork xattr is set (macOS specific), #7234

  • recreate: without --chunker-params, do not re-chunk, #7337

  • recreate: when --target is given, do not detect "nothing to do". use case: borg recreate -a src --target dst can be used to make a copy of an archive inside the same repository, #7254.

  • set .hardlink_master for ALL hardlinkable items, #7175

  • locking: fix host, pid, tid order. tid (thread id) must be parsed as hex from lock file name.

  • update development.lock.txt, including a setuptools security fix, #7227

Other changes:

  • requirements: allow msgpack 1.0.5 also

  • upgrade Cython to 0.29.33

  • hashindex minor fixes, refactor, tweaks, tests

  • use os.replace not os.rename

  • remove BORG_LIBB2_PREFIX (not used any more)

  • docs:

    • BORG_KEY_FILE: clarify docs, #7444

    • update FAQ about locale/unicode issues, #6999

    • improve mount options rendering, #7359

    • make timestamps in manual pages reproducible

    • installation: update Fedora in distribution list, #7357

  • tests:

    • fix test_size_on_disk_accurate for large st_blksize, #7250

    • add same_ts_ns function and use it for relaxed timestamp comparisons

    • "auto" compressor tests: don't assume a specific size, do not assume zlib is better than lz4, #7363

    • add test for extracted directory mtime

  • vagrant:

    • upgrade local freebsd 12.1 box -> generic/freebsd13 box (13.1)

    • use pythons > 3.8 which work on freebsd 13.1

    • pyenv: also install python 3.11.1 for testing

    • pyenv: use python 3.10.1, 3.10.0 build is broken on freebsd

Version 1.2.3 (2022-12-24)

Fixes:

  • create: fix --list --dry-run output for directories, #7209

  • diff/recreate: normalize chunker params before comparing them, #7079

  • check: fix uninitialised variable if repo is completely empty, #7034

  • xattrs: improve error handling, #6988

  • fix args.paths related argparsing, #6994

  • archive.save(): always use metadata from stats (e.g. nfiles, size, ...), #7072

  • tar_filter: recognize .tar.zst as zstd, #7093

  • get_chunker: fix missing sparse=False argument, #7056

  • file_integrity.py: make sure file_fd is always closed on exit

  • repository: cleanup(): close segment before unlinking

  • repository: use os.replace instead of os.rename

Other changes:

  • remove python < 3.7 compatibility code

  • do not use version_tuple placeholder in setuptools_scm template

  • CI: fix tox4 passenv issue, #7199

  • vagrant: update to python 3.9.16, use the openbsd 7.1 box

  • misc. test suite and docs fixes / improvements

  • remove deprecated --prefix from docs, #7109

  • Windows: use MSYS2 for Github CI, remove Appveyor CI

Version 1.2.2 (2022-08-20)

New features:

  • prune/delete --checkpoint-interval=1800 and ctrl-c/SIGINT support, #6284

Fixes:

  • SaveFile: use a custom mkstemp with mode support, #6933, #6400, #6786. This fixes umask/mode/ACL issues (and also "chmod not supported" exceptions seen in 1.2.1) of files updated using SaveFile, e.g. the repo config.

  • hashindex_compact: fix eval order (check idx before use), #5899

  • create --paths-from-(stdin|command): normalize paths, #6778

  • secure_erase: avoid collateral damage, #6768. If a hardlink copy of a repo was made and a new repo config shall be saved, do NOT fill in random garbage before deleting the previous repo config, because that would damage the hardlink copy.

  • list: fix {flags:<WIDTH>} formatting, #6081

  • check: try harder to create the key, #5719

  • misc commands: ctrl-c must not kill other subprocesses, #6912

    • borg create with a remote repo via ssh

    • borg create --content-from-command

    • borg create --paths-from-command

    • (de)compression filter process of import-tar / export-tar

Other changes:

  • deprecate --prefix, use -a / --glob-archives, see #6806

  • make setuptools happy ("package would be ignored"), #6874

  • fix pyproject.toml to create a fixed _version.py file, compatible with both old and new setuptools_scm version, #6875

  • automate asciinema screencasts

  • CI: test on macOS 12 without fuse / fuse tests (too troublesome on github CI due to kernel extensions needed by macFUSE)

  • tests: fix test_obfuscate byte accounting

  • repository: add debug logging for issue #6687

  • _chunker.c: fix warnings on macOS

  • requirements.lock.txt: use the latest cython 0.29.32

  • docs:

    • add info on man page installation, #6894

    • update archive_progress json description about "finished", #6570

    • json progress_percent: some values are optional, #4074

    • FAQ: full quota / full disk, #5960

    • correct shell syntax for installation using git

Version 1.2.1 (2022-06-06)

Fixes:

  • create: skip with warning if opening the parent dir of recursion root fails, #6374

  • create: fix crash. metadata stream can produce all-zero chunks, #6587

  • fix crash when computing stats, escape % chars in archive name, #6500

  • fix transaction rollback: use files cache filename as found in txn.active/, #6353

  • import-tar: kill filter process in case of borg exceptions, #6401 #6681

  • import-tar: fix mtime type bug

  • ensure_dir: respect umask for created directory modes, #6400

  • SaveFile: respect umask for final file mode, #6400

  • check archive: improve error handling for corrupt archive metadata block, make robust_iterator more robust, #4777

  • pre12-meta cache: do not use the cache if want_unique is True, #6612

  • fix scp-style repo url parsing for ip v6 address, #6526

  • mount -o versions: give clear error msg instead of crashing. it does not make sense to request versions view if you only look at 1 archive, but the code shall not crash in that case as it did, but give a clear error msg.

  • show_progress: add finished=true/false to archive_progress json, #6570

  • delete/prune: fix --iec mode output (decimal vs. binary units), #6606

  • info: fix authenticated mode repo to show "Encrypted: No", #6462

  • diff: support presence change for blkdev, chrdev and fifo items, #6615

New features:

  • delete: add repository id and location to prompt, #6453

  • borg debug dump-repo-objs --ghost: new --segment=S --offset=O options

Other changes:

  • support python 3.11

  • allow msgpack 1.0.4, #6716

  • load_key: no key is same as empty key, #6441

  • give a more helpful error msg for unsupported key formats, #6561

  • better error msg for defect or unsupported repo configs, #6566

  • docs:

    • document borg 1.2 pattern matching behavior change, #6407 Make clear that absolute paths always go into the matcher as if they are relative (without leading slash). Adapt all examples accordingly.

    • authentication primitives: improved security and performance infos

    • mention BORG_FILES_CACHE_SUFFIX as alternative to BORG_FILES_CACHE_TTL, #5602

    • FAQ: add a hint about --debug-topic=files_cache

    • improve borg check --max-duration description

    • fix values of TAG bytes, #6515

    • borg compact --cleanup-commits also runs a normal compaction, #6324

    • virtualization speed tips

    • recommend umask for passphrase file perms

    • borg 1.2 is security supported

    • update link to ubuntu packages, #6485

    • use --numeric-ids in pull mode docs

    • remove blake2 docs, blake2 code not bundled any more, #6371

    • clarify on-disk order and size of segment file log entry fields, #6357

    • docs building: do not transform --/--- to unicode dashes

  • tests:

    • check that borg does not require pytest for normal usage, fixes #6563

    • fix OpenBSD symlink mode test failure, #2055

  • vagrant:

    • darwin64: remove fakeroot, #6314

    • update development.lock.txt

    • use pyinstaller 4.10 and python 3.9.13 for binary build

    • upgrade VMCPUS and xdistn from 4 to 16, maybe this speeds up the tests

  • crypto:

    • use hmac.compare_digest instead of ==, #6470

    • hmac_sha256: replace own cython wrapper code by hmac.digest python stdlib (since py38)

    • hmac and blake2b minor optimizations and cleanups

    • removed some unused crypto related code, #6472

    • avoid losing the key (potential use-after-free). this never could happen in 1.2 due to the way we use the code. The issue was discovered in master after other changes, so we also "fixed" it here before it bites us.

  • setup / build:

    • add pyproject.toml, fix sys.path, #6466

    • setuptools_scm: also require it via pyproject.toml

    • allow extra compiler flags for every extension build

    • fix misc. C / Cython compiler warnings, deprecation warnings

    • fix zstd.h include for bundled zstd, #6369

  • source using python 3.8 features: pyupgrade --py38-plus ./**/*.py

Version 1.2.0 (2022-02-22 22:02:22 :-)

Fixes:

  • diff: reduce memory consumption, fix is_hardlink_master, #6295

  • compact: fix / improve freeable / freed space log output

    • derive really freed space from quota use before/after, #5679

    • do not say "freeable", but "maybe freeable" (based on hint, unsure)

  • fix race conditions in internal SaveFile function, #6306 #6028

  • implement internal safe_unlink (was: truncate_and_unlink) function more safely: usually it does not truncate any more, only under "disk full" circumstances and only if there is only one hardlink. see: https://github.com/borgbackup/borg/discussions/6286

Other changes:

  • info: use a pre12-meta cache to accelerate stats for borg < 1.2 archives. the first time borg info is invoked on a borg 1.1 repo, it can take a rather long time computing and caching some stats values for 1.1 archives, which borg 1.2 archives have in their archive metadata structure. be patient, esp. if you have lots of old archives. following invocations are much faster due to the cache. related change: add archive name to calc_stats progress display.

  • docs:

    • add borg 1.2 upgrade notes, #6217

    • link to borg placeholders and borg patterns help

    • init: explain the encryption modes better

    • clarify usage of patternfile roots

    • put import-tar docs into same file as export-tar docs

    • explain the difference between a path that ends with or without a slash, #6297

Version 1.2.0rc1 (2022-02-05)

Fixes:

  • repo::archive location placeholder expansion fixes, #5826, #5998

  • repository: fix intermediate commits, shall be at end of current segment

  • delete: don't commit if nothing was deleted, avoid cache sync, #6060

  • argument parsing: accept some options only once, #6026

  • disallow overwriting of existing keyfiles on init, #6036

  • if ensure_dir() fails, give more informative error message, #5952

New features:

  • delete --force: do not ask when deleting a repo, #5941

Other changes:

  • requirements: exclude broken or incompatible-with-pyinstaller setuptools

  • add a requirements.d/development.lock.txt and use it for vagrant

  • tests:

    • added nonce-related tests

    • refactor: remove assert_true

    • vagrant: macos box tuning, netbsd box fixes, #5370, #5922

  • docs:

    • update install docs / requirements docs, #6180

    • borg mount / FUSE "versions" view is not experimental any more

    • --pattern* is not experimental any more, #6134

    • impact of deleting path/to/repo/nonce, #5858

    • key export: add examples, #6204

    • ~/.config/borg/keys is not used for repokey keys, #6107

    • excluded parent dir's metadata can't restore

Version 1.2.0b4 (2022-01-23)

Fixes:

  • create: fix passing device nodes and symlinks to --paths-from-stdin, #6009

  • create --dry-run: fix display of kept tagfile, #5834

  • check --repair: fix missing parameter in "did not consistently fail" msg, #5822

  • fix hardlinkable file type check, #6037

  • list: remove placeholders for shake_* hashes, #6082

  • prune: handle case of calling prune_split when there are no archives, #6015

  • benchmark crud: make sure cleanup of borg-test-data files/dir happens, #5630

  • do not show archive name in repository-related error msgs, #6014

  • prettier error msg (no stacktrace) if exclude file is missing, #5734

  • do not require BORG_CONFIG_DIR if BORG_{SECURITY,KEYS}_DIR are set, #5979

  • fix pyinstaller detection for dir-mode, #5897

  • atomically create the CACHE_TAG file, #6028

  • deal with the SaveFile/SyncFile race, docs, see #6056 708a5853

  • avoid expanding path into LHS of formatting operation + tests, #6064 #6063

  • repository: quota / compactable computation fixes

  • info: emit repo info even if repo has 0 archives + test, #6120

New features:

  • check --repair: significantly speed up search for next valid object in segment, #6022

  • check: add progress indicator for archive check, #5809

  • create: add retry_erofs workaround for O_NOATIME issue on volume shadow copies in WSL1, #6024

  • create: allow --files-cache=size (this is potentially dangerous, use on your own risk), #5686

  • import-tar: implement import-tar to complement export-tar, #2233

  • implement BORG_SELFTEST env variable (can be carefully used to speedup borg hosting), #5871

  • key export: print key if path is '-' or not given, #6092

  • list --format: Add command_line to format keys

Other changes:

  • pypi metadata: alpha -> beta

  • require python 3.8+, #5975

  • use pyinstaller 4.7

  • allow msgpack 1.0.3

  • upgrade to bundled xxhash to 0.8.1

  • import-tar / export-tar: tar file related changes:

    • check for short tarfile extensions

    • add .lz4 and .zstd

    • fix docs about extensions and decompression commands

  • add github codeql analysis, #6148

  • vagrant:

    • box updates / add new boxes / remove outdated and broken boxes

    • use Python 3.9.10 (incl. binary builds) and 3.10.0

    • fix pyenv initialisation, #5798

    • fix vagrant scp on macOS, #5921

    • use macfuse instead of osxfuse

  • shell completions:

    • update shell completions to 1.1.17, #5923

    • remove BORG_LIBC completion, since 9914968 borg no longer uses find_library().

  • docs:

    • fixed readme.rst irc webchat link (we use libera chat now, not freenode)

    • fix exceptions thrown by setup.py build_man

    • check --repair: recommend checking hw before check --repair, #5855

    • check --verify-data: clarify and document conflict with --repository-only, #5808

    • serve: improve ssh forced commands docs, #6083

    • list: improve docs for borg list --format, #6061

    • list: remove --list-format from borg list

    • FAQ: fix manifest-timestamp path (inside security dir)

    • fix the broken link to .nix file

    • document behavior for filesystems with inconsistent inodes, #5770

    • clarify user_id vs uid for fuse, #5723

    • clarify pattern usage with commands, #5176

    • clarify pp vs. pf pattern type, #5300

    • update referenced freebsd/macOS versions used for binary build, #5942

    • pull mode: add some warnings, #5827

    • clarify "you will need key and passphrase" borg init warning, #4622

    • add missing leading slashes in help patterns, #5857

    • add info on renaming repositories, #5240

    • check: add notice about defective hardware, #5753

    • mention tar --compare (compare archive to fs files), #5880

    • add note about grandfather-father-son backup retention policy / rotation scheme, #6006

    • permissions note rewritten to make it less confusing

    • create github security policy

    • remove leftovers of BORG_HOSTNAME_IS_UNIQUE

    • excluded parent dir's metadata can't restore. (#6062)

    • if parent dir is not extracted, we do not have its metadata

    • clarify who starts the remote agent

Version 1.2.0b3 (2021-05-12)

Fixes:

  • create: fix --progress --log-json, #4360#issuecomment-774580052

  • do not load files cache for commands not using it, #5673

  • fix repeated cache tag file writing bug

New features:

  • create/recreate: print preliminary file status early, #5417

  • create/extract: add --noxattrs and --noacls options, #3955

  • create: verbose files cache logging via --debug-topic=files_cache, #5659

  • mount: implement --numeric-ids (default: False!), #2377

  • diff: add --json-lines option

  • info / create --stats: add --iec option to print sizes in powers of 1024.

Other changes:

  • create: add --upload-(ratelimit|buffer), deprecate --remote-* options, #5611

  • create/extract/mount: add --numeric-ids, deprecate --numeric-owner option, #5724

  • config: accept non-int value for max_segment_size / storage_quota

  • use PyInstaller v4.3, #5671

  • vagrant: use Python 3.9.5 to build binaries

  • tox.ini: modernize and enable execution without preinstalling deps

  • cleanup code style checks

  • get rid of distutils, use setuptools+packaging

  • github CI: test on Python 3.10-dev

  • check: missing / healed chunks: always tell chunk ID, #5704

  • docs:

    • remove bad /var/cache exclusion in example commands, #5625

    • misc. fixes and improvements, esp. for macOS

    • add unsafe workaround to use an old repo copy, #5722

Version 1.2.0b2 (2021-02-06)

Fixes:

  • create: do not recurse into duplicate roots, #5603

  • create: only print stats if not ctrl-c'ed, fixes traceback, #5668

  • extract: improve exception handling when setting xattrs, #5092. emit a warning message giving the path, xattr key and error message. continue trying to restore other xattrs and bsdflags of the same file after an exception with xattr-setting happened.

  • export-tar: fix memory leak with ssh: remote repository, #5568. fix potential memory leak with ssh: remote repository with partial extraction.

  • remove empty shadowed_segments lists, #5275

  • fix bad default: manifest.archives.list(consider_checkpoints=False), fixes tracebacks / KeyErros for missing objects in ChunkIndex, #5668

New features:

  • create: improve sparse file support

    • create --sparse (detect sparse file holes) and file map support, only for the "fixed" chunker, #14

    • detect all-zero chunks in read data in "buzhash" and "fixed" chunkers

    • cached_hash: use a small LRU cache to accelerate all-zero chunks hashing

    • use cached_hash also to generate all-zero replacement chunks

  • create --remote-buffer, add a upload buffer for remote repos, #5574

  • prune: keep oldest archive when retention target not met

Other changes:

  • use blake2 from python 3.6+ hashlib (this removes the requirement for libb2 and the bundled blake2 code)

  • also accept msgpack up to 1.0.2. exclude 1.0.1 though, which had some issues (not sure they affect borg).

  • create: add repository location to --stats output, #5491

  • check: debug log the segment filename

  • delete: add a --list switch to borg delete, #5116

  • borg debug dump-hints - implemented to e.g. to look at shadow_index

  • Tab completion support for additional archives for 'borg delete'

  • refactor: have one borg.constants.zero all-zero bytes object

  • refactor shadow_index updating repo.put/delete, #5661, #5636.

  • docs:

    • add another case of attempted hardlink usage

    • fix description of borg upgrade hardlink usage, #5518

    • use HTTPS everywhere

    • add examples for --paths-from-stdin, --paths-from-command, --paths-separator, #5644

    • fix typos/grammar

    • update docs for dev environment installation instructions

    • recommend running tests only on installed versions for setup

    • add badge with current status of package

  • vagrant:

    • use brew install --cask ..., #5557

    • use Python 3.9.1 and PyInstaller 4.1 to build the borg binary

Version 1.2.0b1 (2020-12-06)

Fixes:

  • BORG_CACHE_DIR crashing borg if empty, atomic handling of recursive directory creation, #5216

  • fix --dry-run and --stats coexistence, #5415

  • allow EIO with warning when trying to hardlink, #4336

  • export-tar: set tar format to GNU_FORMAT explicitly, #5274

  • use --timestamp for {utcnow} and {now} if given, #5189

  • make timestamp helper timezone-aware

New features:

  • create: implement --paths-from-stdin and --paths-from-command, see #5492. These switches read paths to archive from stdin. Delimiter can specified by --paths-delimiter=DELIM. Paths read will be added honoring every option but exclusion options and --one-file-system. borg won't recurse into directories.

  • 'obfuscate' pseudo compressor obfuscates compressed chunk size in repo

  • add pyfuse3 (successor of llfuse) as an alternative lowlevel fuse implementation to llfuse (deprecated), #5407. FUSE implementation can be switched via env var BORG_FUSE_IMPL.

  • allow appending to the files cache filename with BORG_FILES_CACHE_SUFFIX

  • create: implement --stdin-mode, --stdin-user and --stdin-group, #5333

Other changes:

  • split recursive directory walking/processing into directory walking and item processing.

  • fix warning by importing setuptools before distutils.

  • debug info: include infos about FUSE implementation, #5546

  • testing:

    • add a test for the hashindex corruption bug, #5531 #4829

    • move away from travis-ci, use github actions, #5528 #5467

    • test both on fuse2 and fuse3

    • upload coverage reports to codecov

    • fix spurious failure in test_cache_files, #5438

    • add tests for Location.with_timestamp

    • tox: add a non-fuse env to the envlist

  • vagrant:

    • use python 3.7.latest and pyinstaller 4.0 for binary creation

    • pyinstaller: compute basepath from spec file location

    • vagrant: updates/fixes for archlinux box, #5543

  • docs:

    • "filename with spaces" example added to exclude file, #5236

    • add a hint about sleeping computer, #5301

    • how to adjust macOS >= Catalina security settings, #5303

    • process/policy for adding new compression algorithms

    • updated docs about hacked backup client, #5480

    • improve ansible deployment docs, make it more generic

    • how to approach borg speed issues, give speed example, #5371

    • fix mathematical inaccuracy about chunk size, #5336

    • add example for excluding content using --pattern cli option

    • clarify borg create's '--one-file-system' option, #4009

    • improve docs/FAQ about append-only remote repos, #5497

    • fix reST markup issues, labels

    • add infos about contributor retirement status

Version 1.2.0a9 (2020-10-05)

Fixes:

  • fix memory leak related to preloading, #5202

  • check --repair: fix potential data loss, #5325

  • persist shadow_index in between borg runs, #4830

  • fix hardlinked CACHEDIR.TAG processing, #4911

  • --read-special: .part files also should be regular files, #5217

  • allow server side enforcing of umask, --umask is for the local borg process only (see docs), #4947

  • exit with 128 + signal number, #5161

  • borg config --list does not show last_segment_checked, #5159

  • locking:

    • fix ExclusiveLock race condition bug, #4923

    • fix race condition in lock migration, #4953

    • fix locking on openindiana, #5271

New features:

  • --content-from-command: create archive using stdout of given command, #5174

  • allow key-import + BORG_KEY_FILE to create key files

  • build directory-based binary for macOS to avoid Gatekeeper delays

Other changes:

  • upgrade bundled zstd to 1.4.5

  • upgrade bundled xxhash to 0.8.0, #5362

  • if self test fails, also point to OS and hardware, #5334

  • misc. shell completions fixes/updates, rewrite zsh completion

  • prettier error message when archive gets too big, #5307

  • stop relying on false exiting with status code 1

  • rephrase some warnings, #5164

  • parseformat: unnecessary calls removed, #5169

  • testing:

    • enable Python3.9 env for test suite and VMs, #5373

    • drop python 3.5, #5344

    • misc. vagrant fixes/updates

    • misc. testing fixes, #5196

  • docs:

    • add ssh-agent pull backup method to doc, #5288

    • mention double --force in prune docs

    • update Homebrew install instructions, #5185

    • better description of how cache and rebuilds of it work and how the workaround applies to that

    • point to borg create --list item flags in recreate usage, #5165

    • add a note to create from stdin regarding files cache, #5180

    • add security faq explaining AES-CTR crypto issues, #5254

    • clarify --exclude-if-present in recreate, #5193

    • add socat pull mode, #5150, #900

    • move content of resources doc page to community project, #2088

    • explain hash collision, #4884

    • clarify --recompress option, #5154

Version 1.2.0a8 (2020-04-22)

Fixes:

  • fixed potential index corruption / data loss issue due to bug in hashindex_set, #4829. Please read and follow the more detailed notes close to the top of this document.

  • fix crash when upgrading erroneous hints file, #4922

  • commit-time free space calc: ignore bad compact map entries, #4796

  • info: if the archive doesn't exist, print a pretty message, #4793

  • --prefix / -P: fix processing, avoid argparse issue, #4769

  • ignore EACCES (errno 13) when hardlinking, #4730

  • add a try catch when formatting the info string, #4818

  • check: do not stumble over invalid item key, #4845

  • update prevalence of env vars to set config and cache paths

  • mount: fix FUSE low linear read speed on large files, #5032

  • extract: fix confusing output of borg extract --list --strip-components, #4934

  • recreate: support --timestamp option, #4745

  • fix ProgressIndicator msgids (JSON output), #4935

  • fuse: set f_namemax in statfs result, #2684

  • accept absolute paths on windows

  • pyinstaller: work around issue with setuptools > 44

New features:

  • chunker speedup (plus regression test)

  • added --consider-checkpoints and related test, #4788

  • added --noflags option, deprecate --nobsdflags option, #4489

  • compact: add --threshold option, #4674

  • mount: add birthtime to FUSE entries

  • support platforms with no os.link, #4901 - if we don't have os.link, we just extract another copy instead of making a hardlink.

  • move sync_file_range to its own extension for better platform compatibility.

  • new --bypass-lock option to bypass locking, e.g. for read-only repos

  • accept absolute paths by removing leading slashes in patterns of all sorts but re: style, #4029

  • delete: new --keep-security-info option

Other changes:

  • support msgpack 0.6.2 and 1.0.0, #5065

  • upgrade bundled zstd to 1.4.4

  • upgrade bundled lz4 to 1.9.2

  • upgrade xxhash to 0.7.3

  • require recent enough llfuse for birthtime support, #5064

  • only store compressed data if the result actually is smaller, #4516

  • check: improve error output for matching index size, see #4829

  • ignore --stats when given with --dry-run, but continue, #4373

  • replaced usage of os.statvfs with shutil.disk_usage (better cross-platform support).

  • fuse: remove unneeded version check and compat code, micro opts

  • docs:

    • improve description of path variables

    • document how to completely delete data, #2929

    • add FAQ about Borg config dir, #4941

    • add docs about errors not printed as JSON, #4073

    • update usage_general.rst.inc

    • added "Will move with BORG_CONFIG_DIR variable unless specified." to BORG_SECURITY_DIR info.

    • put BORG_SECURITY_DIR immediately below BORG_CONFIG_DIR (and moved BORG_CACHE_DIR up before them).

    • add paragraph regarding cache security assumptions, #4900

    • tell about borg cache security precautions

    • add FAQ describing difference between a local repo vs. repo on a server.

    • document how to test exclusion patterns without performing an actual backup

    • create: tell that "Calculating size" time and space needs are caused by --progress

    • fix/improve documentation for @api decorator, #4674

    • add a pull backup / push restore how-to, #1552

    • fix man pages creation, #4752

    • more general FAQ for backup and retain original paths, #4532

    • explain difference between --exclude and --pattern, #4118

    • add FAQ for preventing SSH timeout in extract, #3866

    • improve password FAQ (decrease pw length, add -w 0 option to base64 to prevent line wrap), #4591

    • add note about patterns and stored paths, #4160

    • add upgrade of tools to pip installation how-to, #5090

    • document one cause of orphaned chunks in check command, #2295

    • clean up the whole check usage paragraph

    • FAQ: linked recommended restrictions to ssh public keys on borg servers, #4946

    • fixed "doc downplays severity of Nonce reuse issue", #4883

    • borg repo restore instructions needed, #3428

    • new FAQ: A repo is corrupt and must be replaced with an older repo.

    • clarify borg init's encryption modes

  • native windows port:

    • update README_WINDOWS.rst

    • updated pyinstaller spec file to support windows builds

  • testing / CI:

    • improved travis config / install script, improved macOS builds

    • allow osx builds to fail, #4955

    • Windows 10 build on Appveyor CI

  • vagrant:

    • upgrade pyinstaller to v3.5 + patch

    • use py369 for binary build, add py380 for tests

    • fix issue in stretch VM hanging at grub installation

    • add a debian buster and a ubuntu focal VM

    • update darwin box to 10.12

    • upgrade FreeBSD box to 12.1

    • fix debianoid virtualenv packages

    • use pyenv in freebsd64 VM

    • remove the flake8 test

    • darwin: avoid error if pkg is already installed

    • debianoid: don't interactively ask questions

Version 1.2.0a7 (2019-09-07)

Fixes:

  • slave hardlinks extraction issue, see #4350

  • extract: fix KeyError for "partial" extraction, #4607

  • preload chunks for hardlink slaves w/o preloaded master, #4350

  • fix preloading for old remote servers, #4652

  • fix partial extract for hardlinked contentless file types, #4725

  • Repository.open: use stat() to check for repo dir, #4695

  • Repository.check_can_create_repository: use stat() to check, ~ #4695.

  • SecurityManager.known(): check all files, #4614

  • after double-force delete, warn about necessary repair, #4704

  • cope with ANY error when importing pytest into borg.testsuite, #4652

  • fix invalid archive error message

  • setup.py: fix detection of missing Cython

  • filter out selinux xattrs, #4574

  • location arg - should it be optional? #4541

  • enable placeholder usage in --comment, #4559

  • use whitelist approach for borg serve, #4097

New features:

  • minimal native Windows support, see windows readme (work in progress)

  • create: first ctrl-c (SIGINT) triggers checkpoint and abort, #4606

  • new BORG_WORKAROUNDS mechanism, basesyncfile, #4710

  • remove WSL autodetection. if WSL still has this problem, you need to set BORG_WORKAROUNDS=basesyncfile in the borg process environment to work around it.

  • support xxh64 checksum in addition to the hashlib hashes in borg list

  • enable placeholder usage in all extra archive arguments

  • enable placeholder usage in --comment, #4559

  • enable placeholder usage in --glob-archives, #4495

  • ability to use a system-provided version of "xxhash"

  • create:

    • changed the default behaviour to not store the atime of fs items. atime is often rather not interesting and fragile - it easily changes even if nothing else has changed and, if stored into the archive, spoils deduplication of the archive metadata stream.

    • if you give the --noatime option, borg will output a deprecation warning because it is currently ignored / does nothing. Please remove the --noatime option when using borg 1.2.

    • added a --atime option for storing files' atime into an archive

Other changes:

  • argparser: always use REPOSITORY in metavar

  • do not check python/libc for borg serve, #4483

  • small borg compact improvements, #4522

  • compact: log freed space at INFO level

  • tests:

    • tox / travis: add testing on py38-dev

    • fix broken test that relied on improper zlib assumptions

    • pure-py msgpack warning shall not make a lot of tests fail, #4558

    • rename test_mount_hardlinks to test_fuse_mount_hardlinks (master)

    • vagrant: add up-to-date openindiana box (py35, openssl10)

    • get rid of confusing coverage warning, #2069

  • docs:

    • reiterate that 'file cache names are absolute' in FAQ, mention bind mount solution, #4738

    • add restore docs, #4670

    • updated docs to cover use of temp directory on remote, #4545

    • add a push-style example to borg-create(1), #4613

    • timestamps in the files cache are now usually ctime, #4583

    • benchmark crud: clarify that space is used until compact

    • update documentation of borg create, corrects a mention of borg 1.1 as a future version.

    • fix osxfuse github link in installation docs

    • how to supply a passphrase, use crypto devices, #4549

    • extract: document limitation "needs empty destination", #4598

    • update macOS Brew link

    • add note about software for automating backup

    • compact: improve docs,

    • README: new URL for funding options

Version 1.2.0a6 (2019-04-22)

Fixes:

  • delete / prune: consider part files correctly for stats, #4507

  • fix "all archives" stats considering part files, #4329

  • create: only run stat_simple_attrs() once

  • create: --stats does not work with --dry-run, exit with error msg, #4373

  • give "invalid repo" error msg if repo config not found, #4411

New features:

  • display msgpack version as part of sysinfo (e.g. in tracebacks)

Other changes:

  • docs:

    • sdd "SSH Configuration" section, #4493, #3988, #636, #4485

    • better document borg check --max-duration, #4473

    • sorted commands help in multiple steps, #4471

  • testing:

    • travis: use py 3.5.3 and 3.6.7 on macOS to get a pyenv-based python build with openssl 1.1

    • vagrant: use py 3.5.3 and 3.6.8 on darwin64 VM to build python and borg with openssl 1.1

    • pytest: -v and default XDISTN to 1, #4481

Version 1.2.0a5 (2019-03-21)

Fixes:

  • warn if a file has changed while being backed up, #1750

  • lrucache: regularly remove old FDs, #4427

  • borg command shall terminate with rc 2 for ImportErrors, #4424

  • make freebsd xattr platform code api compatible with linux, #3952

Other changes:

  • major setup code refactoring (especially how libraries like openssl, liblz4, libzstd, libb2 are discovered and how it falls back to code bundled with borg), new: uses pkg-config now (and needs python "pkgconfig" package installed), #1925

    if you are a borg package maintainer, please try packaging this (see comments in setup.py).

  • Vagrantfile: add zstd, reorder, build env vars, #4444

  • travis: install script improvements

  • update shell completions

  • docs:

    • add a sample logging.conf in docs/misc, #4380

    • fix spelling errors

    • update requirements / install docs, #4374

Version 1.2.0a4 (2019-03-11)

Fixes:

  • do not use O_NONBLOCK for special files, like FIFOs, block and char devices when using --read-special. fixes backing up FIFOs. fixes to test. #4394

  • more LibreSSL build fixes: LibreSSL has HMAC_CTX_free and HMAC_CTX_new

New features:

  • check: incremental repo check (only checks crc32 for segment entries), #1657 borg check --repository-only --max-duration SECONDS ...

  • delete: timestamp for borg delete --info added, #4359

Other changes:

  • redo stale lock handling, #3986 drop BORG_HOSTNAME_IS_UNIQUE (please use BORG_HOST_ID if needed). borg now always assumes it has a unique host id - either automatically from fqdn plus uuid.getnode() or overridden via BORG_HOST_ID.

  • docs:

    • added Alpine Linux to distribution list

    • elaborate on append-only mode docs

  • vagrant:

    • darwin: new 10.12 box

    • freebsd: new 12.0 box

    • openbsd: new 6.4 box

    • misc. updates / fixes

Version 1.2.0a3 (2019-02-26)

Fixes:

  • LibreSSL build fixes, #4403

  • dummy ACL/xattr code fixes (used by OpenBSD and others), #4403

  • create: fix openat/statat issues for root directory, #4405

Version 1.2.0a2 and earlier (2019-02-24)

New features:

  • compact: "borg compact" needs to be used to free repository space by compacting the segments (reading sparse segments, rewriting still needed data to new segments, deleting the sparse segments). Borg < 1.2 invoked compaction automatically at the end of each repository writing command. Borg >= 1.2 does not do that any more to give better speed, more control, more segment file stability (== less stuff moving to newer segments) and more robustness. See the docs about "borg compact" for more details.

  • "borg compact --cleanup-commits" is to cleanup the tons of 17byte long commit-only segment files caused by borg 1.1.x issue #2850. Invoke this once after upgrading (the server side) borg to 1.2. Compaction now automatically removes unneeded commit-only segment files.

  • prune: Show which rule was applied to keep archive, #2886

  • add fixed blocksize chunker (see --chunker-params docs), #1086

Fixes:

  • avoid stale filehandle issues, #3265

  • use more FDs, avoid race conditions on active fs, #906, #908, #1038

  • add O_NOFOLLOW to base flags, #908

  • compact:

    • require >10% freeable space in a segment, #2985

    • repository compaction now automatically removes unneeded 17byte commit-only segments, #2850

  • make swidth available on all posix platforms, #2667

Other changes:

  • repository: better speed and less stuff moving around by using separate segment files for manifest DELETEs and PUTs, #3947

  • use pyinstaller v3.3.1 to build binaries

  • update bundled zstd code to 1.3.8, #4210

  • update bundled lz4 code to 1.8.3, #4209

  • msgpack:

    • switch to recent "msgpack" pypi pkg name, #3890

    • wrap msgpack to avoid future compat complications, #3632, #2738

    • support msgpack 0.6.0 and 0.6.1, #4220, #4308

  • llfuse: modernize / simplify llfuse version requirements

  • code refactorings / internal improvements:

    • include size/csize/nfiles[_parts] stats into archive, #3241

    • calc_stats: use archive stats metadata, if available

    • crypto: refactored crypto to use an AEAD style API

    • crypto: new AES-OCB, CHACHA20-POLY1305

    • create: use less syscalls by not using a python file obj, #906, #3962

    • diff: refactor the diff functionality to new ItemDiff class, #2475

    • archive: create FilesystemObjectProcessors class

    • helpers: make a package, split into smaller modules

    • xattrs: move to platform package, use cython instead ctypes, #2495

    • xattrs/acls/bsdflags: misc. code/api optimizations

    • FUSE: separate creation of filesystem from implementation of llfuse funcs, #3042

    • FUSE: use unpacker.tell() instead of deprecated write_bytes, #3899

    • setup.py: move build_man / build_usage code to setup_docs.py

    • setup.py: update to use a newer Cython/setuptools API for compiling .pyx -> .c, #3788

    • use python 3.5's os.scandir / os.set_blocking

    • multithreading preparations (not used yet):

      • item.to_optr(), Item.from_optr()

      • fix chunker holding the GIL during blocking I/O

    • C code portability / basic MSC compatibility, #4147, #2677

  • testing:

    • vagrant: new VMs for linux/bsd/darwin, most with OpenSSL 1.1 and py36

Version 1.1.18 (2022-06-05)

Compatibility notes:

  • When upgrading from borg 1.0.x to 1.1.x, please note:

    • read all the compatibility notes for 1.1.0*, starting from 1.1.0b1.

    • borg upgrade: you do not need to and you also should not run it. There is one exception though: If you upgrade from an unpatched borg < 1.0.9, please read that section above: "Pre-1.0.9 manifest spoofing vulnerability (CVE-2016-10099)"

    • borg might ask some security-related questions once after upgrading. You can answer them either manually or via environment variable. One known case is if you use unencrypted repositories, then it will ask about a unknown unencrypted repository one time.

    • your first backup with 1.1.x might be significantly slower (it might completely read, chunk, hash a lot files) - this is due to the --files-cache mode change (and happens every time you change mode). You can avoid the one-time slowdown by using the pre-1.1.0rc4-compatible mode (but that is less safe for detecting changed files than the default). See the --files-cache docs for details.

  • 1.1.11 removes WSL autodetection (Windows 10 Subsystem for Linux). If WSL still has a problem with sync_file_range, you need to set BORG_WORKAROUNDS=basesyncfile in the borg process environment to work around the WSL issue.

  • 1.1.14 changes return codes due to a bug fix: In case you have scripts expecting rc == 2 for a signal exit, you need to update them to check for >= 128 (as documented since long).

  • 1.1.15 drops python 3.4 support, minimum requirement is 3.5 now.

  • 1.1.17 install_requires the "packaging" pypi package now.

New features:

  • check --repair: significantly speed up search for next valid object in segment, #6022

  • create: add retry_erofs workaround for O_NOATIME issue on volume shadow copies in WSL1, #6024

  • key export: display key if path is '-' or not given, #6092

  • list --format: add command_line to format keys, #6108

Fixes:

  • check: improve error handling for corrupt archive metadata block, make robust_iterator more robust, #4777

  • diff: support presence change for blkdev, chrdev and fifo items, #6483

  • diff: reduce memory consumption, fix is_hardlink_master

  • init: disallow overwriting of existing keyfiles

  • info: fix authenticated mode repo to show "Encrypted: No", #6462

  • info: emit repo info even if repo has 0 archives, #6120

  • list: remove placeholders for shake_* hashes, #6082

  • mount -o versions: give clear error msg instead of crashing

  • show_progress: add finished=true/false to archive_progress json, #6570

  • fix hardlinkable file type check, #6037

  • do not show archive name in error msgs referring to the repository, #6023

  • prettier error msg (no stacktrace) if exclude file is missing, #5734

  • do not require BORG_CONFIG_DIR if BORG_{SECURITY,KEYS}_DIR are set, #5979

  • atomically create the CACHE_TAG file, #6028

  • deal with the SaveFile/SyncFile race, docs, see #6176 5c5b59bc9

  • avoid expanding path into LHS of formatting operation + tests, #6064 #6063

  • repository: quota / compactable computation fixes, #6119. This is mainly to keep the repo code in sync with borg 1.2. As borg 1.1 compacts immediately, there was not really an issue with this in 1.1.

  • fix transaction rollback: use files cache filename as found in txn.active, #6353

  • do not load files cache for commands not using it, fixes #5673

  • fix scp repo url parsing for ip v6 addrs, #6526

  • repo::archive location placeholder expansion fixes, #5826, #5998

    • use expanded location for log output

    • support placeholder expansion for BORG_REPO env var

  • respect umask for created directory and file modes, #6400

  • safer truncate_and_unlink implementation

Other changes:

  • upgrade bundled xxhash code to 0.8.1

  • fix xxh64 related build (setup.py and post-0.8.1 patch for static_assert). The patch was required to build the bundled xxhash code on FreeBSD, see https://github.com/Cyan4973/xxHash/pull/670

  • msgpack build: remove endianness macro, #6105

  • update and fix shell completions

  • fuse: remove unneeded version check and compat code

  • delete --force: do not ask when deleting a repo, #5941

  • delete: don't commit if nothing was deleted, avoid cache sync, #6060

  • delete: add repository id and location to prompt

  • compact segments: improve freeable / freed space log output, #5679

  • if ensure_dir() fails, give more informative error message, #5952

  • load_key: no key is same as empty key, #6441

  • better error msg for defect or unsupported repo configs, #6566

  • use hmac.compare_digest instead of ==, #6470

  • implement more standard hashindex.setdefault behaviour

  • remove stray punctuation from secure-erase message

  • add development.lock.txt, use a real python 3.5 to generate frozen reqs

  • setuptools 60.7.0 breaks pyinstaller, #6246

  • setup.py clean2 was added to work around some setuptools customizability limitation.

  • allow extra compiler flags for every extension build

  • C code: make switch fallthrough explicit

  • Cython code: fix "useless trailing comma" cython warnings

  • requirements.lock.txt: use the latest cython 0.29.30

  • fix compilation warnings: ‘PyUnicode_AsUnicode’ is deprecated

  • docs:

    • ~/.config/borg/keys is not used for repokey keys, #6107

    • excluded parent dir's metadata can't restore, #6062

    • permissions note rewritten to make it less confusing, #5490

    • add note about grandfather-father-son backup retention policy / rotation scheme

    • clarify who starts the remote agent (borg serve)

    • test/improve pull backup docs, #5903

    • document the socat pull mode described in #900 #515ß

    • borg serve: improve ssh forced commands docs, #6083

    • improve docs for borg list --format, #6080

    • fix the broken link to .nix file

    • clarify pattern usage with commands, #5176

    • clarify user_id vs uid for fuse, #5723

    • fix binary build freebsd/macOS version, #5942

    • FAQ: fix manifest-timestamp path, #6016

    • remove duplicate faq entries, #5926

    • fix sphinx warnings, #5919

    • virtualisation speed tips

    • fix values of TAG bytes, #6515

    • recommend umask for passphrase file perms

    • update link to ubuntu packages, #6485

    • clarify on-disk order and size of log entry fields, #6357

    • do not transform --/--- to unicode dashes

    • improve linking inside docs, link to borg_placeholders, link to borg_patterns

    • use same phrasing in misc. help texts

    • borg init: explain the encryption modes better

    • explain the difference between a path that ends with or without a slash, #6297

    • clarify usage of patternfile roots, #6242

    • borg key export: add examples

    • updates about features not experimental any more: FUSE "versions" view, --pattern*, #6134

    • fix/update cygwin package requirements

    • impact of deleting path/to/repo/nonce, #5858

    • warn about tampered server nonce

    • mention BORG_FILES_CACHE_SUFFIX as alternative to BORG_FILES_CACHE_TTL, #5602

    • add a troubleshooting note about "is not a valid repository" to the FAQ

  • vagrant / CI / testing:

    • misc. fixes and updates, new python versions

    • macOS on github: re-enable fuse2 testing by downgrading to older macOS, #6099

    • fix OpenBSD symlink mode test failure, #2055

    • use the generic/openbsd6 box

    • strengthen the test: we can read data w/o nonces

    • add tests for path/to/repo/nonce deletion

    • darwin64: backport some tunings from master

    • darwin64: remove fakeroot, #6314

    • darwin64: fix vagrant scp, #5921

    • darwin64: use macfuse instead of osxfuse

    • add ubuntu "jammy" 22.04 LTS VM

    • adapt memory for openindiana64 and darwin64

Version 1.1.17 (2021-07-12)

Compatibility notes:

  • When upgrading from borg 1.0.x to 1.1.x, please note:

    • read all the compatibility notes for 1.1.0*, starting from 1.1.0b1.

    • borg upgrade: you do not need to and you also should not run it. There is one exception though: If you upgrade from an unpatched borg < 1.0.9, please read that section above: "Pre-1.0.9 manifest spoofing vulnerability (CVE-2016-10099)"

    • borg might ask some security-related questions once after upgrading. You can answer them either manually or via environment variable. One known case is if you use unencrypted repositories, then it will ask about a unknown unencrypted repository one time.

    • your first backup with 1.1.x might be significantly slower (it might completely read, chunk, hash a lot files) - this is due to the --files-cache mode change (and happens every time you change mode). You can avoid the one-time slowdown by using the pre-1.1.0rc4-compatible mode (but that is less safe for detecting changed files than the default). See the --files-cache docs for details.

  • 1.1.11 removes WSL autodetection (Windows 10 Subsystem for Linux). If WSL still has a problem with sync_file_range, you need to set BORG_WORKAROUNDS=basesyncfile in the borg process environment to work around the WSL issue.

  • 1.1.14 changes return codes due to a bug fix: In case you have scripts expecting rc == 2 for a signal exit, you need to update them to check for >= 128 (as documented since long).

  • 1.1.15 drops python 3.4 support, minimum requirement is 3.5 now.

  • 1.1.17 install_requires the "packaging" pypi package now.

Fixes:

  • pyinstaller dir-mode: fix pyi detection / LIBPATH treatment, #5897

  • handle crash due to kill stale lock race, #5828

  • fix BORG_CACHE_DIR crashing borg if empty, #5216

  • create --dry-run: fix display of kept tagfile, #5834

  • fix missing parameter in "did not consistently fail" msg, #5822

  • missing / healed chunks: always tell chunk ID, #5704

  • benchmark: make sure cleanup happens even on exceptions, #5630

New features:

  • implement BORG_SELFTEST env variable, #5871. this can be used to accelerate borg startup a bit. not recommended for normal usage, but borg mass hosters with a lot of borg invocations can save some resources with this. on my laptop, this saved ~100ms cpu time (sys+user) per borg command invocation.

  • implement BORG_LIBC env variable to give the libc filename, #5870. you can use this if a borg does not find your libc.

  • check: add progress indicator for archive check.

  • allow --files-cache=size (not recommended, make sure you know what you do)

Other changes:

  • Python 3.10 now officially supported! we test on py310-dev on github CI since a while and now also on the vagrant machines, so it should work ok.

  • github CI: test on py310 (again)

  • get rid of distutils, use packaging and setuptools. distutils is deprecated and gives warnings on py 3.10.

  • setup.py: rename "clean" to "clean2" to avoid shadowing the "clean" command.

  • remove libc filename fallback for the BSDs (there is no "usual" name)

  • cleanup flake8 checks, fix some pep8 violations.

  • docs building: replace deprecated function ".add_stylesheet()" for Sphinx 4 compatibility

  • docs:

    • add a hint on sleeping computer and ssh connections, #5301

    • update the documentation on hacked backup client, #5480

    • improve docs/FAQ about append-only remote repos, #5497

    • complement the documentation for pattern files and exclude files, #5520

    • "filename with spaces" example added to exclude file, #5236 note: no whitespace escaping needed, processed by borg.

    • add info on renaming repositories, #5240

    • clarify borg check --verify-data, #5808

    • add notice about defective hardware to check documentation, #5753

    • add paragraph added in #5855 to utility documentation source

    • add missing leading slashes in help patterns, #5857

    • clarify "you will need key and passphrase" borg init warning, #4622

    • pull mode: add some warnings, #5827

    • mention tar --compare (compare archive to fs files), #5880

    • fix typos, backport of #5597

  • vagrant:

    • add py3.7.11 for binary build, also add 3.10-dev.

    • use latest Cython 0.29.23 for py310 compat fixes.

    • more RAM for openindiana upgrade plan resolver, it just hangs (swaps?) if there is too little RAM.

    • fix install_pyenv to adapt to recent changes in pyenv (same as in master now).

    • use generic/netbsd9 box, copied from master branch.

Version 1.1.16 (2021-03-23)

Fixes:

  • setup.py: add special openssl prefix for Apple M1 compatibility

  • do not recurse into duplicate roots, #5603

  • remove empty shadowed_segments lists, #5275, #5614

  • fix libpython load error when borg fat binary / dir-based binary is invoked via a symlink by upgrading pyinstaller to v4.2, #5688

  • config: accept non-int value (like 500M or 100G) for max_segment_size or storage_quota, #5639. please note: when setting a non-int value for this in a repo config, using the repo will require borg >= 1.1.16.

New features:

  • bundled msgpack: drop support for old buffer protocol to support Python 3.10

  • verbose files cache logging via --debug-topic=files_cache, #5659. Use this if you suspect that borg does not detect unmodified files as expected.

  • create/extract: add --noxattrs and --noacls option, #3955. when given with borg create, borg will not get xattrs / ACLs from input files (and thus, it will not archive xattrs / ACLs). when given with borg extract, borg will not read xattrs / ACLs from archive and will not set xattrs / ACLs on extracted files.

  • diff: add --json-lines option, #3765

  • check: debug log segment filename

  • borg debug dump-hints

Other changes:

  • Tab completion support for additional archives for 'borg delete'

  • repository: deduplicate code of put and delete, no functional change

  • tests: fix result order issue (sporadic test failure on openindiana)

  • vagrant:

    • upgrade pyinstaller to v4.2, #5671

    • avoid grub-install asking interactively for device

    • remove the xenial box

    • update freebsd box to 12.1

  • docs:

    • update macOS install instructions, #5677

    • use macFUSE (not osxfuse) for Apple M1 compatibility

    • update docs for dev environment installation instructions, #5643

    • fix grammar in faq

    • recommend running tests only on installed versions for setup

    • add link back to git-installation

    • remove /var/cache exclusion in example commands, #5625. This is generally a poor idea and shouldn't be promoted through examples.

    • add repology.org badge with current packaging status

    • explain hash collision

    • add unsafe workaround to use an old repo copy, #5722

Version 1.1.15 (2020-12-25)

Fixes:

  • extract:

    • improve exception handling when setting xattrs, #5092.

    • emit a warning message giving the path, xattr key and error message.

    • continue trying to restore other xattrs and bsdflags of the same file after an exception with xattr-setting happened.

  • export-tar:

    • set tar format to GNU_FORMAT explicitly, #5274

    • fix memory leak with ssh: remote repository, #5568

    • fix potential memory leak with ssh: remote repository with partial extraction

  • create: fix --dry-run and --stats coexistence, #5415

  • use --timestamp for {utcnow} and {now} if given, #5189

New features:

  • create: implement --stdin-mode, --stdin-user and --stdin-group, #5333

  • allow appending the files cache filename with BORG_FILES_CACHE_SUFFIX env var

Other changes:

  • drop python 3.4 support, minimum requirement is 3.5 now.

  • enable using libxxhash instead of bundled xxh64 code

  • update llfuse requirements (1.3.8)

  • set cython language_level in some files to fix warnings

  • allow EIO with warning when trying to hardlink

  • PropDict: fail early if internal_dict is not a dict

  • update shell completions

  • tests / CI

    • add a test for the hashindex corruption bug, #5531 #4829

    • fix spurious failure in test_cache_files, #5438

    • added a github ci workflow

    • reduce testing on travis, no macOS, no py3x-dev, #5467

    • travis: use newer dists, native py on dist

  • vagrant:

    • remove jessie and trusty boxes, #5348 #5383

    • pyinstaller 4.0, build on py379

    • binary build on stretch64, #5348

    • remove easy_install based pip installation

  • docs:

    • clarify '--one-file-system' for btrfs, #5391

    • add example for excluding content using the --pattern cmd line arg

    • complement the documentation for pattern files and exclude files, #5524

    • made ansible playbook more generic, use package instead of pacman. also change state from "latest" to "present".

    • complete documentation on append-only remote repos, #5497

    • internals: rather talk about target size than statistics, #5336

    • new compression algorithm policy, #1633 #5505

    • faq: add a hint on sleeping computer, #5301

    • note requirements for full disk access on macOS Catalina, #5303

    • fix/improve description of borg upgrade hardlink usage, #5518

  • modernize 1.1 code:

    • drop code/workarounds only needed to support Python 3.4

    • remove workaround for pre-release py37 argparse bug

    • removed some outdated comments/docstrings

    • requirements: remove some restrictions, lock on current versions

Version 1.1.14 (2020-10-07)

Fixes:

  • check --repair: fix potential data loss when interrupting it, #5325

  • exit with 128 + signal number (as documented) when borg is killed by a signal, #5161

  • fix hardlinked CACHEDIR.TAG processing, #4911

  • create --read-special: .part files also should be regular files, #5217

  • llfuse dependency: choose least broken 1.3.6/1.3.7. 1.3.6 is broken on python 3.9, 1.3.7 is broken on FreeBSD.

Other changes:

  • upgrade bundled xxhash to 0.7.4

  • self test: if it fails, also point to OS and hardware, #5334

  • pyinstaller: compute basepath from spec file location

  • prettier error message when archive gets too big, #5307

  • check/recreate are not "experimental" any more (but still potentially dangerous):

    • recreate: remove extra confirmation

    • rephrase some warnings, update docs, #5164

  • shell completions:

    • misc. updates / fixes

    • support repositories in fish tab completion, #5256

    • complete $BORG_RECREATE_I_KNOW_WHAT_I_AM_DOING

    • rewrite zsh completion:

      • completion for almost all optional and positional arguments

      • completion for Borg environment variables (parameters)

  • use "allow/deny list" instead of "white/black list" wording

  • declare "allow_cache_wipe" marker in setup.cfg to avoid pytest warning

  • vagrant / tests:

    • misc. fixes / updates

    • use python 3.5.10 for binary build

    • build directory-based binaries additionally to the single file binaries

    • add libffi-dev, required to build python

    • use cryptography<3.0, more recent versions break the jessie box

    • test on python 3.9

    • do brew update with /dev/null redirect to avoid "too much log output" on travis-ci

  • docs:

    • add ssh-agent pull backup method docs, #5288

    • how to approach borg speed issues, #5371

    • mention double --force in prune docs

    • update Homebrew install instructions, #5185

    • better description of how cache and rebuilds of it work

    • point to borg create --list item flags in recreate usage, #5165

    • add security faq explaining AES-CTR crypto issues, #5254

    • add a note to create from stdin regarding files cache, #5180

    • fix borg.1 manpage generation regression, #5211

    • clarify how exclude options work in recreate, #5193

    • add section for retired contributors

    • hint about not misusing private email addresses of contributors for borg support

Version 1.1.13 (2020-06-06)

Compatibility notes:

  • When upgrading from borg 1.0.x to 1.1.x, please note:

    • read all the compatibility notes for 1.1.0*, starting from 1.1.0b1.

    • borg upgrade: you do not need to and you also should not run it. There is one exception though: If you upgrade from an unpatched borg < 1.0.9, please read that section above: "Pre-1.0.9 manifest spoofing vulnerability (CVE-2016-10099)"

    • borg might ask some security-related questions once after upgrading. You can answer them either manually or via environment variable. One known case is if you use unencrypted repositories, then it will ask about a unknown unencrypted repository one time.

    • your first backup with 1.1.x might be significantly slower (it might completely read, chunk, hash a lot files) - this is due to the --files-cache mode change (and happens every time you change mode). You can avoid the one-time slowdown by using the pre-1.1.0rc4-compatible mode (but that is less safe for detecting changed files than the default). See the --files-cache docs for details.

  • 1.1.11 removes WSL autodetection (Windows 10 Subsystem for Linux). If WSL still has a problem with sync_file_range, you need to set BORG_WORKAROUNDS=basesyncfile in the borg process environment to work around the WSL issue.

Fixes:

  • rebuilt using a current Cython version, compatible with python 3.8, #5214

Version 1.1.12 (2020-06-06)

Fixes:

  • fix preload-related memory leak, #5202.

  • mount / borgfs (FUSE filesystem):

    • fix FUSE low linear read speed on large files, #5067

    • fix crash on old llfuse without birthtime attrs, #5064 - accidentally we required llfuse >= 1.3. Now also old llfuse works again.

    • set f_namemax in statfs result, #2684

  • update precedence of env vars to set config and cache paths, #4894

  • correctly calculate compression ratio, taking header size into account, too

New features:

  • --bypass-lock option to bypass locking with read-only repositories

Other changes:

  • upgrade bundled zstd to 1.4.5

  • travis: adding comments and explanations to Travis config / install script, improve macOS builds.

  • tests: test_delete_force: avoid sporadic test setup issues, #5196

  • misc. vagrant fixes

  • the binary for macOS is now built on macOS 10.12

  • the binaries for Linux are now built on Debian 8 "Jessie", #3761

  • docs:

    • PlaceholderError not printed as JSON, #4073

    • "How important is Borg config?", #4941

    • make Sphinx warnings break docs build, #4587

    • some markup / warning fixes

    • add "updating borgbackup.org/releases" to release checklist, #4999

    • add "rendering docs" to release checklist, #5000

    • clarify borg init's encryption modes

    • add note about patterns and stored paths, #4160

    • add upgrade of tools to pip installation how-to

    • document one cause of orphaned chunks in check command, #2295

    • linked recommended restrictions to ssh public keys on borg servers in faq, #4946

Version 1.1.11 (2020-03-08)

Compatibility notes:

  • When upgrading from borg 1.0.x to 1.1.x, please note:

    • read all the compatibility notes for 1.1.0*, starting from 1.1.0b1.

    • borg upgrade: you do not need to and you also should not run it. There is one exception though: If you upgrade from an unpatched borg < 1.0.9, please read that section above: "Pre-1.0.9 manifest spoofing vulnerability (CVE-2016-10099)"

    • borg might ask some security-related questions once after upgrading. You can answer them either manually or via environment variable. One known case is if you use unencrypted repositories, then it will ask about a unknown unencrypted repository one time.

    • your first backup with 1.1.x might be significantly slower (it might completely read, chunk, hash a lot files) - this is due to the --files-cache mode change (and happens every time you change mode). You can avoid the one-time slowdown by using the pre-1.1.0rc4-compatible mode (but that is less safe for detecting changed files than the default). See the --files-cache docs for details.

  • 1.1.11 removes WSL autodetection (Windows 10 Subsystem for Linux). If WSL still has a problem with sync_file_range, you need to set BORG_WORKAROUNDS=basesyncfile in the borg process environment to work around the WSL issue.

Fixes:

  • fixed potential index corruption / data loss issue due to bug in hashindex_set, #4829. Please read and follow the more detailed notes close to the top of this document.

  • upgrade bundled xxhash to 0.7.3, #4891. 0.7.2 is the minimum requirement for correct operations on ARMv6 in non-fixup mode, where unaligned memory accesses cause bus errors. 0.7.3 adds some speedups and libxxhash 0.7.3 even has a pkg-config file now.

  • upgrade bundled lz4 to 1.9.2

  • upgrade bundled zstd to 1.4.4

  • fix crash when upgrading erroneous hints file, #4922

  • extract:

    • fix KeyError for "partial" extraction, #4607

    • fix "partial" extract for hardlinked contentless file types, #4725

    • fix preloading for old (0.xx) remote servers, #4652

    • fix confusing output of borg extract --list --strip-components, #4934

  • delete: after double-force delete, warn about necessary repair, #4704

  • create: give invalid repo error msg if repo config not found, #4411

  • mount: fix FUSE mount missing st_birthtime, #4763 #4767

  • check: do not stumble over invalid item key, #4845

  • info: if the archive doesn't exist, print a pretty message, #4793

  • SecurityManager.known(): check all files, #4614

  • Repository.open: use stat() to check for repo dir, #4695

  • Repository.check_can_create_repository: use stat() to check, #4695

  • fix invalid archive error message

  • fix optional/non-optional location arg, #4541

  • commit-time free space calc: ignore bad compact map entries, #4796

  • ignore EACCES (errno 13) when hardlinking the old config, #4730

  • --prefix / -P: fix processing, avoid argparse issue, #4769

New features:

  • enable placeholder usage in all extra archive arguments

  • new BORG_WORKAROUNDS mechanism, basesyncfile, #4710

  • recreate: support --timestamp option, #4745

  • support platforms without os.link (e.g. Android with Termux), #4901. if we don't have os.link, we just extract another copy instead of making a hardlink.

  • support linux platforms without sync_file_range (e.g. Android 7 with Termux), #4905

Other:

  • ignore --stats when given with --dry-run, but continue, #4373

  • add some ProgressIndicator msgids to code / fix docs, #4935

  • elaborate on "Calculating size" message

  • argparser: always use REPOSITORY in metavar, also use more consistent help phrasing.

  • check: improve error output for matching index size, see #4829

  • docs:

    • changelog: add advisory about hashindex_set bug #4829

    • better describe BORG_SECURITY_DIR, BORG_CACHE_DIR, #4919

    • infos about cache security assumptions, #4900

    • add FAQ describing difference between a local repo vs. repo on a server.

    • document how to test exclusion patterns without performing an actual backup

    • timestamps in the files cache are now usually ctime, #4583

    • fix bad reference to borg compact (does not exist in 1.1), #4660

    • create: borg 1.1 is not future any more

    • extract: document limitation "needs empty destination", #4598

    • how to supply a passphrase, use crypto devices, #4549

    • fix osxfuse github link in installation docs

    • add example of exclude-norecurse rule in help patterns

    • update macOS Brew link

    • add note about software for automating backups, #4581

    • AUTHORS: mention copyright+license for bundled msgpack

    • fix various code blocks in the docs, #4708

    • updated docs to cover use of temp directory on remote, #4545

    • add restore docs, #4670

    • add a pull backup / push restore how-to, #1552

    • add FAQ how to retain original paths, #4532

    • explain difference between --exclude and --pattern, #4118

    • add FAQs for SSH connection issues, #3866

    • improve password FAQ, #4591

    • reiterate that 'file cache names are absolute' in FAQ

  • tests:

    • cope with ANY error when importing pytest into borg.testsuite, #4652

    • fix broken test that relied on improper zlib assumptions

    • test_fuse: filter out selinux xattrs, #4574

  • travis / vagrant:

    • misc python versions removed / changed (due to openssl 1.1 compatibility) or added (3.7 and 3.8, for better borg compatibility testing)

    • binary building is on python 3.5.9 now

  • vagrant:

    • add new boxes: ubuntu 18.04 and 20.04, debian 10

    • update boxes: openindiana, darwin, netbsd

    • remove old boxes: centos 6

    • darwin: updated osxfuse to 3.10.4

    • use debian/ubuntu pip/virtualenv packages

    • rather use python 3.6.2 than 3.6.0, fixes coverage/sqlite3 issue

    • use requirements.d/development.lock.txt to avoid compat issues

  • travis:

    • darwin: backport some install code / order from master

    • remove deprecated keyword "sudo" from travis config

    • allow osx builds to fail, #4955 this is due to travis-ci frequently being so slow that the OS X builds just fail because they exceed 50 minutes and get killed by travis.

Version 1.1.10 (2019-05-16)

Fixes:

  • extract: hang on partial extraction with ssh: repo, when hardlink master is not matched/extracted and borg hangs on related slave hardlink, #4350

  • lrucache: regularly remove old FDs, #4427

  • avoid stale filehandle issues, #3265

  • freebsd: make xattr platform code api compatible with linux, #3952

  • use whitelist approach for borg serve, #4097

  • borg command shall terminate with rc 2 for ImportErrors, #4424

  • create: only run stat_simple_attrs() once, this increases backup with lots of unchanged files performance by ~ 5%.

  • prune: fix incorrect borg prune --stats output with --dry-run, #4373

  • key export: emit user-friendly error if repo key is exported to a directory, #4348

New features:

  • bundle latest supported msgpack-python release (0.5.6), remove msgpack-python from setup.py install_requires - by default we use the bundled code now. optionally, we still support using an external msgpack (see hints in setup.py), but this requires solid requirements management within distributions and is not recommended. borgbackup will break if you upgrade msgpack to an unsupported version.

  • display msgpack version as part of sysinfo (e.g. in tracebacks)

  • timestamp for borg delete --info added, #4359

  • enable placeholder usage in --comment and --glob-archives, #4559, #4495

Other:

  • serve: do not check python/libc for borg serve, #4483

  • shell completions: borg diff second archive

  • release scripts: signing binaries with Qubes OS support

  • testing:

    • vagrant: upgrade openbsd box to 6.4

    • travis-ci: lock test env to py 3.4 compatible versions, #4343

    • get rid of confusing coverage warning, #2069

    • rename test_mount_hardlinks to test_fuse_mount_hardlinks, so both can be excluded by "not test_fuse".

    • pure-py msgpack warning shall not make a lot of tests fail, #4558

  • docs:

    • add "SSH Configuration" section to "borg serve", #3988, #636, #4485

    • README: new URL for funding options

    • add a sample logging.conf in docs/misc, #4380

    • elaborate on append-only mode docs, #3504

    • installation: added Alpine Linux to distribution list, #4415

    • usage.html: only modify window.location when redirecting, #4133

    • add msgpack license to docs/3rd_party/msgpack

  • vagrant / binary builds:

    • use python 3.5.7 for builds

    • use osxfuse 3.8.3

Version 1.1.9 (2019-02-10)

Compatibility notes:

  • When upgrading from borg 1.0.x to 1.1.x, please note:

    • read all the compatibility notes for 1.1.0*, starting from 1.1.0b1.

    • borg upgrade: you do not need to and you also should not run it. There is one exception though: If you upgrade from an unpatched borg < 1.0.9, please read that section above: "Pre-1.0.9 manifest spoofing vulnerability (CVE-2016-10099)"

    • borg might ask some security-related questions once after upgrading. You can answer them either manually or via environment variable. One known case is if you use unencrypted repositories, then it will ask about a unknown unencrypted repository one time.

    • your first backup with 1.1.x might be significantly slower (it might completely read, chunk, hash a lot files) - this is due to the --files-cache mode change (and happens every time you change mode). You can avoid the one-time slowdown by using the pre-1.1.0rc4-compatible mode (but that is less safe for detecting changed files than the default). See the --files-cache docs for details.

Fixes:

  • security fix: configure FUSE with "default_permissions", #3903 "default_permissions" is now enforced by borg by default to let the kernel check uid/gid/mode based permissions. "ignore_permissions" can be given to not enforce "default_permissions".

  • make "hostname" short, even on misconfigured systems, #4262

  • fix free space calculation on macOS (and others?), #4289

  • config: quit with error message when no key is provided, #4223

  • recover_segment: handle too small segment files correctly, #4272

  • correctly release memoryview, #4243

  • avoid diaper pattern in configparser by opening files, #4263

  • add "# cython: language_level=3" directive to .pyx files, #4214

  • info: consider part files for "This archive" stats, #3522

  • work around Microsoft WSL issue #645 (sync_file_range), #1961

New features:

  • add --rsh command line option to complement BORG_RSH env var, #1701

  • init: --make-parent-dirs parent1/parent2/repo_dir, #4235

Other:

  • add archive name to check --repair output, #3447

  • check for unsupported msgpack versions

  • shell completions:

    • new shell completions for borg 1.1.9

    • more complete shell completions for borg mount -o

    • added shell completions for borg help

    • option arguments for zsh tab completion

  • docs:

    • add FAQ regarding free disk space check, #3905

    • update BORG_PASSCOMMAND example and clarify variable expansion, #4249

    • FAQ regarding change of compression settings, #4222

    • add note about BSD flags to changelog, #4246

    • improve logging in example automation script

    • add note about files changing during backup, #4081

    • work around the backslash issue, #4280

    • update release workflow using twine (docs, scripts), #4213

    • add warnings on repository copies to avoid future problems, #4272

  • tests:

    • fix the homebrew 1.9 issues on travis-ci, #4254

    • fix duplicate test method name, #4311

Version 1.1.8 (2018-12-09)

Fixes:

  • enforce storage quota if set by serve-command, #4093

  • invalid locations: give err msg containing parsed location, #4179

  • list repo: add placeholders for hostname and username, #4130

  • on linux, symlinks can't have ACLs, so don't try to set any, #4044

New features:

  • create: added PATH::archive output on INFO log level

  • read a passphrase from a file descriptor specified in the BORG_PASSPHRASE_FD environment variable.

Other:

  • docs:

    • option --format is required for some expensive-to-compute values for json

      borg list by default does not compute expensive values except when they are needed. whether they are needed is determined by the format, in standard mode as well as in --json mode.

    • tell that our binaries are x86/x64 amd/intel, bauerj has ARM

    • fixed wrong archive name pattern in CRUD benchmark help

    • fixed link to cachedir spec in docs, #4140

  • tests:

    • stop using fakeroot on travis, avoids sporadic EISDIR errors, #2482

    • xattr key names must start with "user." on linux

    • fix code so flake8 3.6 does not complain

    • explicitly convert environment variable to str, #4136

    • fix DeprecationWarning: Flags not at the start of the expression, #4137

    • support pytest4, #4172

  • vagrant:

    • use python 3.5.6 for builds

Version 1.1.7 (2018-08-11)

Compatibility notes:

  • added support for Python 3.7

Fixes:

  • cache lock: use lock_wait everywhere to fix infinite wait, see #3968

  • don't archive tagged dir when recursing an excluded dir, #3991

  • py37 argparse: work around bad default in py 3.7.0a/b/rc, #3996

  • py37 remove loggerDict.clear() from tearDown method, #3805

  • some fixes for bugs which likely did not result in problems in practice:

    • fixed logic bug in platform module API version check

    • fixed xattr/acl function prototypes, added missing ones

New features:

  • init: add warning to store both key and passphrase at safe place(s)

  • BORG_HOST_ID env var to work around all-zero MAC address issue, #3985

  • borg debug dump-repo-objs --ghost (dump everything from segment files, including deleted or superseded objects or commit tags)

  • borg debug search-repo-objs (search in repo objects for hex bytes or strings)

Other changes:

  • add Python 3.7 support

  • updated shell completions

  • call socket.gethostname only once

  • locking: better logging, add some asserts

  • borg debug dump-repo-objs:

    • filename layout improvements

    • use repository.scan() to get on-disk order

  • docs:

    • update installation instructions for macOS

    • added instructions to install fuse via homebrew

    • improve diff docs

    • added note that checkpoints inside files requires 1.1+

    • add link to tempfile module

    • remove row/column-spanning from docs source, #4000 #3990

  • tests:

    • fetch less data via os.urandom

    • add py37 env for tox

    • travis: add 3.7, remove 3.6-dev (we test with -dev in master)

  • vagrant / binary builds:

    • use osxfuse 3.8.2

    • use own (uptodate) openindiana box

Version 1.1.6 (2018-06-11)

Compatibility notes:

  • 1.1.6 changes:

    • also allow msgpack-python 0.5.6.

Fixes:

  • fix borg exception handling on ENOSPC error with xattrs, #3808

  • prune: fix/improve overall progress display

  • borg config repo ... does not need cache/manifest/key, #3802

  • debug dump-repo-objs should not depend on a manifest obj

  • pypi package:

    • include .coveragerc, needed by tox.ini

    • fix package long description, #3854

New features:

  • mount: add uid, gid, umask mount options

  • delete:

    • only commit once, #3823

    • implement --dry-run, #3822

  • check:

    • show progress while rebuilding missing manifest, #3787

    • more --repair output

  • borg config --list <repo>, #3612

Other changes:

  • update msgpack requirement, #3753

  • update bundled zstd to 1.3.4, #3745

  • update bundled lz4 code to 1.8.2, #3870

  • docs:

    • describe what BORG_LIBZSTD_PREFIX does

    • fix and deduplicate encryption quickstart docs, #3776

  • vagrant:

    • FUSE for macOS: upgrade 3.7.1 to 3.8.0

    • exclude macOS High Sierra upgrade on the darwin64 machine

    • remove borgbackup.egg-info dir in fs_init (after rsync)

    • use pyenv-based build/test on jessie32/62

    • use local 32 and 64bit debian jessie boxes

    • use "vagrant" as username for new xenial box

  • travis OS X: use xcode 8.3 (not broken)

Version 1.1.5 (2018-04-01)

Compatibility notes:

  • 1.1.5 changes:

    • require msgpack-python >= 0.4.6 and < 0.5.0. 0.5.0+ dropped python 3.4 testing and also caused some other issues because the python package was renamed to msgpack and emitted some FutureWarning.

Fixes:

  • create --list: fix that it was never showing M status, #3492

  • create: fix timing for first checkpoint (read files cache early, init checkpoint timer after that), see #3394

  • extract: set rc=1 when extracting damaged files with all-zero replacement chunks or with size inconsistencies, #3448

  • diff: consider an empty file as different to a non-existing file, #3688

  • files cache: improve exception handling, #3553

  • ignore exceptions in scandir_inorder() caused by an implicit stat(), also remove unneeded sort, #3545

  • fixed tab completion problem where a space is always added after path even when it shouldn't

  • build: do .h file content checks in binary mode, fixes build issue for non-ascii header files on pure-ascii locale platforms, #3544 #3639

  • borgfs: fix patterns/paths processing, #3551

  • config: add some validation, #3566

  • repository config: add validation for max_segment_size, #3592

  • set cache previous_location on load instead of save

  • remove platform.uname() call which caused library mismatch issues, #3732

  • add exception handler around deprecated platform.linux_distribution() call

  • use same datetime object for {now} and {utcnow}, #3548

New features:

  • create: implement --stdin-name, #3533

  • add chunker_params to borg archive info (--json)

  • BORG_SHOW_SYSINFO=no to hide system information from exceptions

Other changes:

  • updated zsh completions for borg 1.1.4

  • files cache related code cleanups

  • be more helpful when parsing invalid --pattern values, #3575

  • be more clear in secure-erase warning message, #3591

  • improve getpass user experience, #3689

  • docs build: unicode problem fixed when using a py27-based sphinx

  • docs:

    • security: explicitly note what happens OUTSIDE the attack model

    • security: add note about combining compression and encryption

    • security: describe chunk size / proximity issue, #3687

    • quickstart: add note about permissions, borg@localhost, #3452

    • quickstart: add introduction to repositories & archives, #3620

    • recreate --recompress: add missing metavar, clarify description, #3617

    • improve logging docs, #3549

    • add an example for --pattern usage, #3661

    • clarify path semantics when matching, #3598

    • link to offline documentation from README, #3502

    • add docs on how to verify a signed release with GPG, #3634

    • chunk seed is generated per repository (not: archive)

    • better formatting of CPU usage documentation, #3554

    • extend append-only repo rollback docs, #3579

  • tests:

    • fix erroneously skipped zstd compressor tests, #3606

    • skip a test if argparse is broken, #3705

  • vagrant:

    • xenial64 box now uses username 'vagrant', #3707

    • move cleanup steps to fs_init, #3706

    • the boxcutter wheezy boxes are 404, use local ones

    • update to Python 3.5.5 (for binary builds)

Version 1.1.4 (2017-12-31)

Compatibility notes:

  • When upgrading from borg 1.0.x to 1.1.x, please note:

    • read all the compatibility notes for 1.1.0*, starting from 1.1.0b1.

    • borg upgrade: you do not need to and you also should not run it. There is one exception though: If you upgrade from an unpatched borg < 1.0.9, please read that section above: "Pre-1.0.9 manifest spoofing vulnerability (CVE-2016-10099)"

    • borg might ask some security-related questions once after upgrading. You can answer them either manually or via environment variable. One known case is if you use unencrypted repositories, then it will ask about a unknown unencrypted repository one time.

    • your first backup with 1.1.x might be significantly slower (it might completely read, chunk, hash a lot files) - this is due to the --files-cache mode change (and happens every time you change mode). You can avoid the one-time slowdown by using the pre-1.1.0rc4-compatible mode (but that is less safe for detecting changed files than the default). See the --files-cache docs for details.

  • borg 1.1.4 changes:

    • zstd compression is new in borg 1.1.4, older borg can't handle it.

    • new minimum requirements for the compression libraries - if the required versions (header and lib) can't be found at build time, bundled code will be used:

      • added requirement: libzstd >= 1.3.0 (bundled: 1.3.2)

      • updated requirement: liblz4 >= 1.7.0 / r129 (bundled: 1.8.0)

Fixes:

  • check: data corruption fix: fix for borg check --repair malfunction, #3444. See the more detailed notes close to the top of this document.

  • delete: also delete security dir when deleting a repo, #3427

  • prune: fix building the "borg prune" man page, #3398

  • init: use given --storage-quota for local repo, #3470

  • init: properly quote repo path in output

  • fix startup delay with dns-only own fqdn resolving, #3471

New features:

  • added zstd compression. try it!

  • added placeholder {reverse-fqdn} for fqdn in reverse notation

  • added BORG_BASE_DIR environment variable, #3338

Other changes:

  • list help topics when invalid topic is requested

  • fix lz4 deprecation warning, requires lz4 >= 1.7.0 (r129)

  • add parens for C preprocessor macro argument usages (did not cause malfunction)

  • exclude broken pytest 3.3.0 release

  • updated fish/bash completions

  • init: more clear exception messages for borg create, #3465

  • docs:

    • add auto-generated docs for borg config

    • don't generate HTML docs page for borgfs, #3404

    • docs update for lz4 b2 zstd changes

    • add zstd to compression help, readme, docs

    • update requirements and install docs about bundled lz4 and zstd

  • refactored build of the compress and crypto.low_level extensions, #3415:

    • move some lib/build related code to setup_{zstd,lz4,b2}.py

    • bundle lz4 1.8.0 (requirement: >= 1.7.0 / r129)

    • bundle zstd 1.3.2 (requirement: >= 1.3.0)

    • blake2 was already bundled

    • rename BORG_LZ4_PREFIX env var to BORG_LIBLZ4_PREFIX for better consistency: we also have BORG_LIBB2_PREFIX and BORG_LIBZSTD_PREFIX now.

    • add prefer_system_lib* = True settings to setup.py - by default the build will prefer a shared library over the bundled code, if library and headers can be found and meet the minimum requirements.

Version 1.1.3 (2017-11-27)

Fixes:

  • Security Fix for CVE-2017-15914: Incorrect implementation of access controls allows remote users to override repository restrictions in Borg servers. A user able to access a remote Borg SSH server is able to circumvent access controls post-authentication. Affected releases: 1.1.0, 1.1.1, 1.1.2. Releases 1.0.x are NOT affected.

  • crc32: deal with unaligned buffer, add tests - this broke borg on older ARM CPUs that can not deal with unaligned 32bit memory accesses and raise a bus error in such cases. the fix might also improve performance on some CPUs as all 32bit memory accesses by the crc32 code are properly aligned now. #3317

  • mount: fixed support of --consider-part-files and do not show .borg_part_N files by default in the mounted FUSE filesystem. #3347

  • fixed cache/repo timestamp inconsistency message, highlight that information is obtained from security dir (deleting the cache will not bypass this error in case the user knows this is a legitimate repo).

  • borgfs: don't show sub-command in borgfs help, #3287

  • create: show an error when --dry-run and --stats are used together, #3298

New features:

  • mount: added exclusion group options and paths, #2138

    Reused some code to support similar options/paths as borg extract offers - making good use of these to only mount a smaller subset of dirs/files can speed up mounting a lot and also will consume way less memory.

    borg mount [options] repo_or_archive mountpoint path [paths...]

    paths: you can just give some "root paths" (like for borg extract) to only partially populate the FUSE filesystem.

    new options: --exclude[-from], --pattern[s-from], --strip-components

  • create/extract: support st_birthtime on platforms supporting it, #3272

  • add "borg config" command for querying/setting/deleting config values, #3304

Other changes:

  • clean up and simplify packaging (only package committed files, do not install .c/.h/.pyx files)

  • docs:

    • point out tuning options for borg create, #3239

    • add instructions for using ntfsclone, zerofree, #81

    • move image backup-related FAQ entries to a new page

    • clarify key aliases for borg list --format, #3111

    • mention break-lock in checkpointing FAQ entry, #3328

    • document sshfs rename workaround, #3315

    • add FAQ about removing files from existing archives

    • add FAQ about different prune policies

    • usage and man page for borgfs, #3216

    • clarify create --stats duration vs. wall time, #3301

    • clarify encrypted key format for borg key export, #3296

    • update release checklist about security fixes

    • document good and problematic option placements, fix examples, #3356

    • add note about using --nobsdflags to avoid speed penalty related to bsdflags, #3239

    • move most of support section to www.borgbackup.org

Version 1.1.2 (2017-11-05)

Fixes:

  • fix KeyError crash when talking to borg server < 1.0.7, #3244

  • extract: set bsdflags last (include immutable flag), #3263

  • create: don't do stat() call on excluded-norecurse directory, fix exception handling for stat() call, #3209

  • create --stats: do not count data volume twice when checkpointing, #3224

  • recreate: move chunks_healthy when excluding hardlink master, #3228

  • recreate: get rid of chunks_healthy when rechunking (does not match), #3218

  • check: get rid of already existing not matching chunks_healthy metadata, #3218

  • list: fix stdout broken pipe handling, #3245

  • list/diff: remove tag-file options (not used), #3226

New features:

  • bash, zsh and fish shell auto-completions, see scripts/shell_completions/

  • added BORG_CONFIG_DIR env var, #3083

Other changes:

  • docs:

    • clarify using a blank passphrase in keyfile mode

    • mention "!" (exclude-norecurse) type in "patterns" help

    • document to first heal before running borg recreate to re-chunk stuff, because that will have to get rid of chunks_healthy metadata.

    • more than 23 is not supported for CHUNK_MAX_EXP, #3115

    • borg does not respect nodump flag by default any more

    • clarify same-filesystem requirement for borg upgrade, #2083

    • update / rephrase cygwin / WSL status, #3174

    • improve docs about --stats, #3260

  • vagrant: openindiana new clang package

Already contained in 1.1.1 (last minute fix):

  • arg parsing: fix fallback function, refactor, #3205. This is a fixup for #3155, which was broken on at least python <= 3.4.2.

Version 1.1.1 (2017-10-22)

Compatibility notes:

  • The deprecated --no-files-cache is not a global/common option any more, but only available for borg create (it is not needed for anything else). Use --files-cache=disabled instead of --no-files-cache.

  • The nodump flag ("do not backup this file") is not honoured any more by default because this functionality (esp. if it happened by error or unexpected) was rather confusing and unexplainable at first to users. If you want that "do not backup NODUMP-flagged files" behaviour, use: borg create --exclude-nodump ...

  • If you are on Linux and do not need bsdflags archived, consider using --nobsdflags with borg create to avoid additional syscalls and speed up backup creation.

Fixes:

  • borg recreate: correctly compute part file sizes. fixes cosmetic, but annoying issue as borg check complains about size inconsistencies of part files in affected archives. you can solve that by running borg recreate on these archives, see also #3157.

  • bsdflags support: do not open BLK/CHR/LNK files, avoid crashes and slowness, #3130

  • recreate: don't crash on attic archives w/o time_end, #3109

  • don't crash on repository filesystems w/o hardlink support, #3107

  • don't crash in first part of truncate_and_unlink, #3117

  • fix server-side IndexError crash with clients < 1.0.7, #3192

  • don't show traceback if only a global option is given, show help, #3142

  • cache: use SaveFile for more safety, #3158

  • init: fix wrong encryption choices in command line parser, fix missing "authenticated-blake2", #3103

  • move --no-files-cache from common to borg create options, #3146

  • fix detection of non-local path (failed on ..filename), #3108

  • logging with fileConfig: set json attr on "borg" logger, #3114

  • fix crash with relative BORG_KEY_FILE, #3197

  • show excluded dir with "x" for tagged dirs / caches, #3189

New features:

  • create: --nobsdflags and --exclude-nodump options, #3160

  • extract: --nobsdflags option, #3160

Other changes:

  • remove annoying hardlinked symlinks warning, #3175

  • vagrant: use self-made FreeBSD 10.3 box, #3022

  • travis: don't brew update, hopefully fixes #2532

  • docs:

    • readme: -e option is required in borg 1.1

    • add example showing --show-version --show-rc

    • use --format rather than --list-format (deprecated) in example

    • update docs about hardlinked symlinks limitation

Version 1.1.0 (2017-10-07)

Compatibility notes:

  • borg command line: do not put options in between positional arguments

    This sometimes works (e.g. it worked in borg 1.0.x), but can easily stop working if we make positional arguments optional (like it happened for borg create's "paths" argument in 1.1). There are also places in borg 1.0 where we do that, so it doesn't work there in general either. #3356

    Good: borg create -v --stats repo::archive path Good: borg create repo::archive path -v --stats Bad: borg create repo::archive -v --stats path

Fixes:

  • fix LD_LIBRARY_PATH restoration for subprocesses, #3077

  • "auto" compression: make sure expensive compression is actually better, otherwise store lz4 compressed data we already computed.

Other changes:

  • docs:

    • FAQ: we do not implement futile attempts of ETA / progress displays

    • manpage: fix typos, update homepage

    • implement simple "issue" role for manpage generation, #3075

Version 1.1.0rc4 (2017-10-01)

Compatibility notes:

  • A borg server >= 1.1.0rc4 does not support borg clients 1.1.0b3-b5. #3033

  • The files cache is now controlled differently and has a new default mode:

    • the files cache now uses ctime by default for improved file change detection safety. You can still use mtime for more speed and less safety.

    • --ignore-inode is deprecated (use --files-cache=... without "inode")

    • --no-files-cache is deprecated (use --files-cache=disabled)

New features:

  • --files-cache - implement files cache mode control, #911 You can now control the files cache mode using this option: --files-cache={ctime,mtime,size,inode,rechunk,disabled} (only some combinations are supported). See the docs for details.

Fixes:

  • remote progress/logging: deal with partial lines, #2637

  • remote progress: flush json mode output

  • fix subprocess environments, #3050 (and more)

Other changes:

  • remove client_supports_log_v3 flag, #3033

  • exclude broken Cython 0.27(.0) in requirements, #3066

  • vagrant:

    • upgrade to FUSE for macOS 3.7.1

    • use Python 3.5.4 to build the binaries

  • docs:

    • security: change-passphrase only changes the passphrase, #2990

    • fixed/improved borg create --compression examples, #3034

    • add note about metadata dedup and --no[ac]time, #2518

    • twitter account @borgbackup now, better visible, #2948

    • simplified rate limiting wrapper in FAQ

Version 1.1.0rc3 (2017-09-10)

New features:

  • delete: support naming multiple archives, #2958

Fixes:

  • repo cleanup/write: invalidate cached FDs, #2982

  • fix datetime.isoformat() microseconds issues, #2994

  • recover_segment: use mmap(), lower memory needs, #2987

Other changes:

  • with-lock: close segment file before invoking subprocess

  • keymanager: don't depend on optional readline module, #2976

  • docs:

    • fix macOS keychain integration command

    • show/link new screencasts in README, #2936

    • document utf-8 locale requirement for json mode, #2273

  • vagrant: clean up shell profile init, user name, #2977

  • test_detect_attic_repo: don't test mount, #2975

  • add debug logging for repository cleanup

Version 1.1.0rc2 (2017-08-28)

Compatibility notes:

  • list: corrected mix-up of "isomtime" and "mtime" formats. Previously, "isomtime" was the default but produced a verbose human format, while "mtime" produced a ISO-8601-like format. The behaviours have been swapped (so "mtime" is human, "isomtime" is ISO-like), and the default is now "mtime". "isomtime" is now a real ISO-8601 format ("T" between date and time, not a space).

New features:

  • None.

Fixes:

  • list: fix weird mixup of mtime/isomtime

  • create --timestamp: set start time, #2957

  • ignore corrupt files cache, #2939

  • migrate locks to child PID when daemonize is used

  • fix exitcode of borg serve, #2910

  • only compare contents when chunker params match, #2899

  • umount: try fusermount, then try umount, #2863

Other changes:

  • JSON: use a more standard ISO 8601 datetime format, #2376

  • cache: write_archive_index: truncate_and_unlink on error, #2628

  • detect non-upgraded Attic repositories, #1933

  • delete various nogil and threading related lines

  • coala / pylint related improvements

  • docs:

    • renew asciinema/screencasts, #669

    • create: document exclusion through nodump, #2949

    • minor formatting fixes

    • tar: tarpipe example

    • improve "with-lock" and "info" docs, #2869

    • detail how to use macOS/GNOME/KDE keyrings for repo passwords, #392

  • travis: only short-circuit docs-only changes for pull requests

  • vagrant:

    • netbsd: bash is already installed

    • fix netbsd version in PKG_PATH

    • add exe location to PATH when we build an exe

Version 1.1.0rc1 (2017-07-24)

Compatibility notes:

  • delete: removed short option for --cache-only

New features:

  • support borg list repo --format {comment} {bcomment} {end}, #2081

  • key import: allow reading from stdin, #2760

Fixes:

  • with-lock: avoid creating segment files that might be overwritten later, #1867

  • prune: fix checkpoints processing with --glob-archives

  • FUSE: versions view: keep original file extension at end, #2769

  • fix --last, --first: do not accept values <= 0, fix reversed archive ordering with --last

  • include testsuite data (attic.tar.gz) when installing the package

  • use limited unpacker for outer key, for manifest (both security precautions), #2174 #2175

  • fix bashism in shell scripts, #2820, #2816

  • cleanup endianness detection, create _endian.h, fixes build on alpine linux, #2809

  • fix crash with --no-cache-sync (give known chunk size to chunk_incref), #2853

Other changes:

  • FUSE: versions view: linear numbering by archive time

  • split up interval parsing from filtering for --keep-within, #2610

  • add a basic .editorconfig, #2734

  • use archive creation time as mtime for FUSE mount, #2834

  • upgrade FUSE for macOS (osxfuse) from 3.5.8 to 3.6.3, #2706

  • hashindex: speed up by replacing modulo with "if" to check for wraparound

  • coala checker / pylint: fixed requirements and .coafile, more ignores

  • borg upgrade: name backup directories as 'before-upgrade', #2811

  • add .mailmap

  • some minor changes suggested by lgtm.com

  • docs:

    • better explanation of the --ignore-inode option relevance, #2800

    • fix openSUSE command and add openSUSE section

    • simplify ssh authorized_keys file using "restrict", add legacy note, #2121

    • mount: show usage of archive filters

    • mount: add repository example, #2462

    • info: update and add examples, #2765

    • prune: include example

    • improved style / formatting

    • improved/fixed segments_per_dir docs

    • recreate: fix wrong "remove unwanted files" example

    • reference list of status chars in borg recreate --filter description

    • update source-install docs about doc build dependencies, #2795

    • cleanup installation docs

    • file system requirements, update segs per dir

    • fix checkpoints/parts reference in FAQ, #2859

  • code:

    • hashindex: don't pass side effect into macro

    • crypto low_level: don't mutate local bytes()

    • use dash_open function to open file or "-" for stdin/stdout

    • archiver: argparse cleanup / refactoring

    • shellpattern: add match_end arg

  • tests: added some additional unit tests, some fixes, #2700 #2710

  • vagrant: fix setup of cygwin, add Debian 9 "stretch"

  • travis: don't perform full travis build on docs-only changes, #2531

Version 1.1.0b6 (2017-06-18)

Compatibility notes:

  • Running "borg init" via a "borg serve --append-only" server will not create an append-only repository anymore. Use "borg init --append-only" to initialize an append-only repository.

  • Repositories in the "repokey" and "repokey-blake2" modes with an empty passphrase are now treated as unencrypted repositories for security checks (e.g. BORG_UNKNOWN_UNENCRYPTED_REPO_ACCESS_IS_OK).

    Previously there would be no prompts nor messages if an unknown repository in one of these modes with an empty passphrase was encountered. This would allow an attacker to swap a repository, if one assumed that the lack of password prompts was due to a set BORG_PASSPHRASE.

    Since the "trick" does not work if BORG_PASSPHRASE is set, this does generally not affect scripts.

  • Repositories in the "authenticated" mode are now treated as the unencrypted repositories they are.

  • The client-side temporary repository cache now holds unencrypted data for better speed.

  • borg init: removed the short form of --append-only (-a).

  • borg upgrade: removed the short form of --inplace (-i).

New features:

  • reimplemented the RepositoryCache, size-limited caching of decrypted repo contents, integrity checked via xxh64. #2515

  • reduced space usage of chunks.archive.d. Existing caches are migrated during a cache sync. #235 #2638

  • integrity checking using xxh64 for important files used by borg, #1101:

    • repository: index and hints files

    • cache: chunks and files caches, chunks.archive.d

  • improve cache sync speed, #1729

  • create: new --no-cache-sync option

  • add repository mandatory feature flags infrastructure, #1806

  • Verify most operations against SecurityManager. Location, manifest timestamp and key types are now checked for almost all non-debug commands. #2487

  • implement storage quotas, #2517

  • serve: add --restrict-to-repository, #2589

  • BORG_PASSCOMMAND: use external tool providing the key passphrase, #2573

  • borg export-tar, #2519

  • list: --json-lines instead of --json for archive contents, #2439

  • add --debug-profile option (and also "borg debug convert-profile"), #2473

  • implement --glob-archives/-a, #2448

  • normalize authenticated key modes for better naming consistency:

    • rename "authenticated" to "authenticated-blake2" (uses blake2b)

    • implement "authenticated" mode (uses hmac-sha256)

Fixes:

  • hashindex: read/write indices >2 GiB on 32bit systems, better error reporting, #2496

  • repository URLs: implement IPv6 address support and also more informative error message when parsing fails.

  • mount: check whether llfuse is installed before asking for passphrase, #2540

  • mount: do pre-mount checks before opening repository, #2541

  • FUSE:

    • fix crash if empty (None) xattr is read, #2534

    • fix read(2) caching data in metadata cache

    • fix negative uid/gid crash (fix crash when mounting archives of external drives made on cygwin), #2674

    • redo ItemCache, on top of object cache

    • use decrypted cache

    • remove unnecessary normpaths

  • serve: ignore --append-only when initializing a repository (borg init), #2501

  • serve: fix incorrect type of exception_short for Errors, #2513

  • fix --exclude and --exclude-from recursing into directories, #2469

  • init: don't allow creating nested repositories, #2563

  • --json: fix encryption[mode] not being the cmdline name

  • remote: propagate Error.traceback correctly

  • fix remote logging and progress, #2241

    • implement --debug-topic for remote servers

    • remote: restore "Remote:" prefix (as used in 1.0.x)

    • rpc negotiate: enable v3 log protocol only for supported clients

    • fix --progress and logging in general for remote

  • fix parse_version, add tests, #2556

  • repository: truncate segments (and also some other files) before unlinking, #2557

  • recreate: keep timestamps as in original archive, #2384

  • recreate: if single archive is not processed, exit 2

  • patterns: don't recurse with ! / --exclude for pf:, #2509

  • cache sync: fix n^2 behaviour in lookup_name

  • extract: don't write to disk with --stdout (affected non-regular-file items), #2645

  • hashindex: implement KeyError, more tests

Other changes:

  • remote: show path in PathNotAllowed

  • consider repokey w/o passphrase == unencrypted, #2169

  • consider authenticated mode == unencrypted, #2503

  • restrict key file names, #2560

  • document follow_symlinks requirements, check libc, use stat and chown with follow_symlinks=False, #2507

  • support common options on the main command, #2508

  • support common options on mid-level commands (e.g. borg key export)

  • make --progress a common option

  • increase DEFAULT_SEGMENTS_PER_DIR to 1000

  • chunker: fix invalid use of types (function only used by tests)

  • chunker: don't do uint32_t >> 32

  • FUSE:

    • add instrumentation (--debug and SIGUSR1/SIGINFO)

    • reduced memory usage for repository mounts by lazily instantiating archives

    • improved archive load times

  • info: use CacheSynchronizer & HashIndex.stats_against (better performance)

  • docs:

    • init: document --encryption as required

    • security: OpenSSL usage

    • security: used implementations; note python libraries

    • security: security track record of OpenSSL and msgpack

    • patterns: document denial of service (regex, wildcards)

    • init: note possible denial of service with "none" mode

    • init: document SHA extension is supported in OpenSSL and thus SHA is faster on AMD Ryzen than blake2b.

    • book: use A4 format, new builder option format.

    • book: create appendices

    • data structures: explain repository compaction

    • data structures: add chunk layout diagram

    • data structures: integrity checking

    • data structures: demingle cache and repo index

    • Attic FAQ: separate section for attic stuff

    • FAQ: I get an IntegrityError or similar - what now?

    • FAQ: Can I use Borg on SMR hard drives?, #2252

    • FAQ: specify "using inline shell scripts"

    • add systemd warning regarding placeholders, #2543

    • xattr: document API

    • add docs/misc/borg-data-flow data flow chart

    • debugging facilities

    • README: how to help the project, #2550

    • README: add bountysource badge, #2558

    • fresh new theme + tweaking

    • logo: vectorized (PDF and SVG) versions

    • frontends: use headlines - you can link to them

    • mark --pattern, --patterns-from as experimental

    • highlight experimental features in online docs

    • remove regex based pattern examples, #2458

    • nanorst for "borg help TOPIC" and --help

    • split deployment

    • deployment: hosting repositories

    • deployment: automated backups to a local hard drive

    • development: vagrant, windows10 requirements

    • development: update docs remarks

    • split usage docs, #2627

    • usage: avoid bash highlight, [options] instead of <options>

    • usage: add benchmark page

    • helpers: truncate_and_unlink doc

    • don't suggest to leak BORG_PASSPHRASE

    • internals: columnize rather long ToC [webkit fixup] internals: manifest & feature flags

    • internals: more HashIndex details

    • internals: fix ASCII art equations

    • internals: edited obj graph related sections a bit

    • internals: layers image + description

    • fix way too small figures in pdf

    • index: disable syntax highlight (bash)

    • improve options formatting, fix accidental block quotes

  • testing / checking:

    • add support for using coala, #1366

    • testsuite: add ArchiverCorruptionTestCase

    • do not test logger name, #2504

    • call setup_logging after destroying logging config

    • testsuite.archiver: normalise pytest.raises vs. assert_raises

    • add test for preserved intermediate folder permissions, #2477

    • key: add round-trip test

    • remove attic dependency of the tests, #2505

    • enable remote tests on cygwin

    • tests: suppress tar's future timestamp warning

    • cache sync: add more refcount tests

    • repository: add tests, including corruption tests

  • vagrant:

    • control VM cpus and pytest workers via env vars VMCPUS and XDISTN

    • update cleaning workdir

    • fix openbsd shell

    • add OpenIndiana

  • packaging:

    • binaries: don't bundle libssl

    • setup.py clean to remove compiled files

    • fail in borg package if version metadata is very broken (setuptools_scm)

  • repo / code structure:

    • create borg.algorithms and borg.crypto packages

    • algorithms: rename crc32 to checksums

    • move patterns to module, #2469

    • gitignore: complete paths for src/ excludes

    • cache: extract CacheConfig class

    • implement IntegrityCheckedFile + Detached variant, #2502 #1688

    • introduce popen_with_error_handling to handle common user errors

Version 1.1.0b5 (2017-04-30)

Compatibility notes:

  • BORG_HOSTNAME_IS_UNIQUE is now on by default.

  • removed --compression-from feature

  • recreate: add --recompress flag, unify --always-recompress and --recompress

Fixes:

  • catch exception for os.link when hardlinks are not supported, #2405

  • borg rename / recreate: expand placeholders, #2386

  • generic support for hardlinks (files, devices, FIFOs), #2324

  • extract: also create parent dir for device files, if needed, #2358

  • extract: if a hardlink master is not in the to-be-extracted subset, the "x" status was not displayed for it, #2351

  • embrace y2038 issue to support 32bit platforms: clamp timestamps to int32, #2347

  • verify_data: fix IntegrityError handling for defect chunks, #2442

  • allow excluding parent and including child, #2314

Other changes:

  • refactor compression decision stuff

  • change global compression default to lz4 as well, to be consistent with --compression defaults.

  • placeholders: deny access to internals and other unspecified stuff

  • clearer error message for unrecognized placeholder

  • more clear exception if borg check does not help, #2427

  • vagrant: upgrade FUSE for macOS to 3.5.8, #2346

  • linux binary builds: get rid of glibc 2.13 dependency, #2430

  • docs:

    • placeholders: document escaping

    • serve: env vars in original commands are ignored

    • tell what kind of hardlinks we support

    • more docs about compression

    • LICENSE: use canonical formulation ("copyright holders and contributors" instead of "author")

    • document borg init behaviour via append-only borg serve, #2440

    • be clear about what buzhash is used for, #2390

    • add hint about chunker params, #2421

    • clarify borg upgrade docs, #2436

    • FAQ to explain warning when running borg check --repair, #2341

    • repository file system requirements, #2080

    • pre-install considerations

    • misc. formatting / crossref fixes

  • tests:

    • enhance travis setuptools_scm situation

    • add extra test for the hashindex

    • fix invalid param issue in benchmarks

These belong to 1.1.0b4 release, but did not make it into changelog by then:

  • vagrant: increase memory for parallel testing

  • lz4 compress: lower max. buffer size, exception handling

  • add docstring to do_benchmark_crud

  • patterns help: mention path full-match in intro

Version 1.1.0b4 (2017-03-27)

Compatibility notes:

  • init: the --encryption argument is mandatory now (there are several choices)

  • moved "borg migrate-to-repokey" to "borg key migrate-to-repokey".

  • "borg change-passphrase" is deprecated, use "borg key change-passphrase" instead.

  • the --exclude-if-present option now supports tagging a folder with any filesystem object type (file, folder, etc), instead of expecting only files as tags, #1999

  • the --keep-tag-files option has been deprecated in favor of the new --keep-exclude-tags, to account for the change mentioned above.

  • use lz4 compression by default, #2179

New features:

  • JSON API to make developing frontends and automation easier (see All about JSON: How to develop frontends)

    • add JSON output to commands: borg create/list/info --json ....

    • add --log-json option for structured logging output.

    • add JSON progress information, JSON support for confirmations (yes()).

  • add two new options --pattern and --patterns-from as discussed in #1406

  • new path full match pattern style (pf:) for very fast matching, #2334

  • add 'debug dump-manifest' and 'debug dump-archive' commands

  • add 'borg benchmark crud' command, #1788

  • new 'borg delete --force --force' to delete severely corrupted archives, #1975

  • info: show utilization of maximum archive size, #1452

  • list: add dsize and dcsize keys, #2164

  • paperkey.html: Add interactive html template for printing key backups.

  • key export: add qr html export mode

  • securely erase config file (which might have old encryption key), #2257

  • archived file items: add size to metadata, 'borg extract' and 'borg check' do check the file size for consistency, FUSE uses precomputed size from Item.

Fixes:

  • fix remote speed regression introduced in 1.1.0b3, #2185

  • fix regression handling timestamps beyond 2262 (revert bigint removal), introduced in 1.1.0b3, #2321

  • clamp (nano)second values to unproblematic range, #2304

  • hashindex: rebuild hashtable if we have too little empty buckets (performance fix), #2246

  • Location regex: fix bad parsing of wrong syntax

  • ignore posix_fadvise errors in repository.py, #2095

  • borg rpc: use limited msgpack.Unpacker (security precaution), #2139

  • Manifest: Make sure manifest timestamp is strictly monotonically increasing.

  • create: handle BackupOSError on a per-path level in one spot

  • create: clarify -x option / meaning of "same filesystem"

  • create: don't create hard link refs to failed files

  • archive check: detect and fix missing all-zero replacement chunks, #2180

  • files cache: update inode number when --ignore-inode is used, #2226

  • fix decompression exceptions crashing check --verify-data and others instead of reporting integrity error, #2224 #2221

  • extract: warning for unextracted big extended attributes, #2258, #2161

  • mount: umount on SIGINT/^C when in foreground

  • mount: handle invalid hard link refs

  • mount: fix huge RAM consumption when mounting a repository (saves number of archives * 8 MiB), #2308

  • hashindex: detect mingw byte order #2073

  • hashindex: fix wrong skip_hint on hashindex_set when encountering tombstones, the regression was introduced in #1748

  • fix ChunkIndex.__contains__ assertion for big-endian archs

  • fix borg key/debug/benchmark crashing without subcommand, #2240

  • Location: accept //servername/share/path

  • correct/refactor calculation of unique/non-unique chunks

  • extract: fix missing call to ProgressIndicator.finish

  • prune: fix error msg, it is --keep-within, not --within

  • fix "auto" compression mode bug (not compressing), #2331

  • fix symlink item fs size computation, #2344

Other changes:

  • remote repository: improved async exception processing, #2255 #2225

  • with --compression auto,C, only use C if lz4 achieves at least 3% compression

  • PatternMatcher: only normalize path once, #2338

  • hashindex: separate endian-dependent defs from endian detection

  • migrate-to-repokey: ask using canonical_path() as we do everywhere else.

  • SyncFile: fix use of fd object after close

  • make LoggedIO.close_segment reentrant

  • creating a new segment: use "xb" mode, #2099

  • redo key_creator, key_factory, centralise key knowledge, #2272

  • add return code functions, #2199

  • list: only load cache if needed

  • list: files->items, clarifications

  • list: add "name" key for consistency with info cmd

  • ArchiveFormatter: add "start" key for compatibility with "info"

  • RemoteRepository: account rx/tx bytes

  • setup.py build_usage/build_man/build_api fixes

  • Manifest.in: simplify, exclude .so, .dll and .orig, #2066

  • FUSE: get rid of chunk accounting, st_blocks = ceil(size / blocksize).

  • tests:

    • help python development by testing 3.6-dev

    • test for borg delete --force

  • vagrant:

    • freebsd: some fixes, #2067

    • darwin64: use osxfuse 3.5.4 for tests / to build binaries

    • darwin64: improve VM settings

    • use python 3.5.3 to build binaries, #2078

    • upgrade pyinstaller from 3.1.1+ to 3.2.1

    • pyinstaller: use fixed AND freshly compiled bootloader, #2002

    • pyinstaller: automatically builds bootloader if missing

  • docs:

    • create really nice man pages

    • faq: mention --remote-ratelimit in bandwidth limit question

    • fix caskroom link, #2299

    • docs/security: reiterate that RPC in Borg does no networking

    • docs/security: counter tracking, #2266

    • docs/development: update merge remarks

    • address SSH batch mode in docs, #2202 #2270

    • add warning about running build_usage on Python >3.4, #2123

    • one link per distro in the installation page

    • improve --exclude-if-present and --keep-exclude-tags, #2268

    • improve automated backup script in doc, #2214

    • improve remote-path description

    • update docs for create -C default change (lz4)

    • document relative path usage, #1868

    • document snapshot usage, #2178

    • corrected some stuff in internals+security

    • internals: move toctree to after the introduction text

    • clarify metadata kind, manifest ops

    • key enc: correct / clarify some stuff, link to internals/security

    • datas: enc: 1.1.x mas different MACs

    • datas: enc: correct factual error -- no nonce involved there.

    • make internals.rst an index page and edit it a bit

    • add "Cryptography in Borg" and "Remote RPC protocol security" sections

    • document BORG_HOSTNAME_IS_UNIQUE, #2087

    • FAQ by categories as proposed by @anarcat in #1802

    • FAQ: update Which file types, attributes, etc. are not preserved?

    • development: new branching model for git repository

    • development: define "ours" merge strategy for auto-generated files

    • create: move --exclude note to main doc

    • create: move item flags to main doc

    • fix examples using borg init without -e/--encryption

    • list: don't print key listings in fat (html + man)

    • remove Python API docs (were very incomplete, build problems on RTFD)

    • added FAQ section about backing up root partition

Version 1.1.0b3 (2017-01-15)

Compatibility notes:

  • borg init: removed the default of "--encryption/-e", #1979 This was done so users do a informed decision about -e mode.

Bug fixes:

  • borg recreate: don't rechunkify unless explicitly told so

  • borg info: fixed bug when called without arguments, #1914

  • borg init: fix free space check crashing if disk is full, #1821

  • borg debug delete/get obj: fix wrong reference to exception

  • fix processing of remote ~/ and ~user/ paths (regressed since 1.1.0b1), #1759

  • posix platform module: only build / import on non-win32 platforms, #2041

New features:

  • new CRC32 implementations that are much faster than the zlib one used previously, #1970

  • add blake2b key modes (use blake2b as MAC). This links against system libb2, if possible, otherwise uses bundled code

  • automatically remove stale locks - set BORG_HOSTNAME_IS_UNIQUE env var to enable stale lock killing. If set, stale locks in both cache and repository are deleted. #562 #1253

  • borg info <repo>: print general repo information, #1680

  • borg check --first / --last / --sort / --prefix, #1663

  • borg mount --first / --last / --sort / --prefix, #1542

  • implement "health" item formatter key, #1749

  • BORG_SECURITY_DIR to remember security related infos outside the cache. Key type, location and manifest timestamp checks now survive cache deletion. This also means that you can now delete your cache and avoid previous warnings, since Borg can still tell it's safe.

  • implement BORG_NEW_PASSPHRASE, #1768

Other changes:

  • borg recreate:

    • remove special-cased --dry-run

    • update --help

    • remove bloat: interruption blah, autocommit blah, resuming blah

    • re-use existing checkpoint functionality

    • archiver tests: add check_cache tool - lints refcounts

  • fixed cache sync performance regression from 1.1.0b1 onwards, #1940

  • syncing the cache without chunks.archive.d (see The borg cache eats way too much disk space, what can I do?) now avoids any merges and is thus faster, #1940

  • borg check --verify-data: faster due to linear on-disk-order scan

  • borg debug-xxx commands removed, we use "debug xxx" subcommands now, #1627

  • improve metadata handling speed

  • shortcut hashindex_set by having hashindex_lookup hint about address

  • improve / add progress displays, #1721

  • check for index vs. segment files object count mismatch

  • make RPC protocol more extensible: use named parameters.

  • RemoteRepository: misc. code cleanups / refactors

  • clarify cache/repository README file

  • docs:

    • quickstart: add a comment about other (remote) filesystems

    • quickstart: only give one possible ssh url syntax, all others are documented in usage chapter.

    • mention file://

    • document repo URLs / archive location

    • clarify borg diff help, #980

    • deployment: synthesize alternative --restrict-to-path example

    • improve cache / index docs, esp. files cache docs, #1825

    • document using "git merge 1.0-maint -s recursive -X rename-threshold=20%" for avoiding troubles when merging the 1.0-maint branch into master.

  • tests:

    • FUSE tests: catch ENOTSUP on freebsd

    • FUSE tests: test troublesome xattrs last

    • fix byte range error in test, #1740

    • use monkeypatch to set env vars, but only on pytest based tests.

    • point XDG_*_HOME to temp dirs for tests, #1714

    • remove all BORG_* env vars from the outer environment

Version 1.1.0b2 (2016-10-01)

Bug fixes:

  • fix incorrect preservation of delete tags, leading to "object count mismatch" on borg check, #1598. This only occurred with 1.1.0b1 (not with 1.0.x) and is normally fixed by running another borg create/delete/prune.

  • fix broken --progress for double-cell paths (e.g. CJK), #1624

  • borg recreate: also catch SIGHUP

  • FUSE:

    • fix hardlinks in versions view, #1599

    • add parameter check to ItemCache.get to make potential failures more clear

New features:

  • Archiver, RemoteRepository: add --remote-ratelimit (send data)

  • borg help compression, #1582

  • borg check: delete chunks with integrity errors, #1575, so they can be "repaired" immediately and maybe healed later.

  • archives filters concept (refactoring/unifying older code)

    • covers --first/--last/--prefix/--sort-by options

    • currently used for borg list/info/delete

Other changes:

  • borg check --verify-data slightly tuned (use get_many())

  • change {utcnow} and {now} to ISO-8601 format ("T" date/time separator)

  • repo check: log transaction IDs, improve object count mismatch diagnostic

  • Vagrantfile: use TW's fresh-bootloader pyinstaller branch

  • fix module names in api.rst

  • hashindex: bump api_version

Version 1.1.0b1 (2016-08-28)

New features:

  • new commands:

    • borg recreate: re-create existing archives, #787 #686 #630 #70, also see #757, #770.

      • selectively remove files/dirs from old archives

      • re-compress data

      • re-chunkify data, e.g. to have upgraded Attic / Borg 0.xx archives deduplicate with Borg 1.x archives or to experiment with chunker-params.

    • borg diff: show differences between archives

    • borg with-lock: execute a command with the repository locked, #990

  • borg create:

    • Flexible compression with pattern matching on path/filename, and LZ4 heuristic for deciding compressibility, #810, #1007

    • visit files in inode order (better speed, esp. for large directories and rotating disks)

    • in-file checkpoints, #1217

    • increased default checkpoint interval to 30 minutes (was 5 minutes), #896

    • added uuid archive format tag, #1151

    • save mountpoint directories with --one-file-system, makes system restore easier, #1033

    • Linux: added support for some BSD flags, #1050

    • add 'x' status for excluded paths, #814

      • also means files excluded via UF_NODUMP, #1080

  • borg check:

    • will not produce the "Checking segments" output unless new --progress option is passed, #824.

    • --verify-data to verify data cryptographically on the client, #975

  • borg list, #751, #1179

    • removed {formatkeys}, see "borg list --help"

    • --list-format is deprecated, use --format instead

    • --format now also applies to listing archives, not only archive contents, #1179

    • now supports the usual [PATH [PATHS…]] syntax and excludes

    • new keys: csize, num_chunks, unique_chunks, NUL

    • supports guaranteed_available hashlib hashes (to avoid varying functionality depending on environment), which includes the SHA1 and SHA2 family as well as MD5

  • borg prune:

    • to better visualize the "thinning out", we now list all archives in reverse time order. rephrase and reorder help text.

    • implement --keep-last N via --keep-secondly N, also --keep-minutely. assuming that there is not more than 1 backup archive made in 1s, --keep-last N and --keep-secondly N are equivalent, #537

    • cleanup checkpoints except the latest, #1008

  • borg extract:

    • added --progress, #1449

    • Linux: limited support for BSD flags, #1050

  • borg info:

    • output is now more similar to borg create --stats, #977

  • borg mount:

    • provide "borgfs" wrapper for borg mount, enables usage via fstab, #743

    • "versions" mount option - when used with a repository mount, this gives a merged, versioned view of the files in all archives, #729

  • repository:

    • added progress information to commit/compaction phase (often takes some time when deleting/pruning), #1519

    • automatic recovery for some forms of repository inconsistency, #858

    • check free space before going forward with a commit, #1336

    • improved write performance (esp. for rotating media), #985

      • new IO code for Linux

      • raised default segment size to approx 512 MiB

    • improved compaction performance, #1041

    • reduced client CPU load and improved performance for remote repositories, #940

  • options that imply output (--show-rc, --show-version, --list, --stats, --progress) don't need -v/--info to have that output displayed, #865

  • add archive comments (via borg (re)create --comment), #842

  • borg list/prune/delete: also output archive id, #731

  • --show-version: shows/logs the borg version, #725

  • added --debug-topic for granular debug logging, #1447

  • use atomic file writing/updating for configuration and key files, #1377

  • BORG_KEY_FILE environment variable, #1001

  • self-testing module, #970

Bug fixes:

  • list: fixed default output being produced if --format is given with empty parameter, #1489

  • create: fixed overflowing progress line with CJK and similar characters, #1051

  • prune: fixed crash if --prefix resulted in no matches, #1029

  • init: clean up partial repo if passphrase input is aborted, #850

  • info: quote cmdline arguments that have spaces in them

  • fix hardlinks failing in some cases for extracting subtrees, #761

Other changes:

  • replace stdlib hmac with OpenSSL, zero-copy decrypt (10-15% increase in performance of hash-lists and extract).

  • improved chunker performance, #1021

  • open repository segment files in exclusive mode (fail-safe), #1134

  • improved error logging, #1440

  • Source:

    • pass meta-data around, #765

    • move some constants to new constants module

    • better readability and fewer errors with namedtuples, #823

    • moved source tree into src/ subdirectory, #1016

    • made borg.platform a package, #1113

    • removed dead crypto code, #1032

    • improved and ported parts of the test suite to py.test, #912

    • created data classes instead of passing dictionaries around, #981, #1158, #1161

    • cleaned up imports, #1112

  • Docs:

    • better help texts and sphinx reproduction of usage help:

      • Group options

      • Nicer list of options in Sphinx

      • Deduplicate 'Common options' (including --help)

    • chunker: added some insights by "Voltara", #903

    • clarify what "deduplicated size" means

    • fix / update / add package list entries

    • added a SaltStack usage example, #956

    • expanded FAQ

    • new contributors in AUTHORS!

  • Tests:

    • vagrant: add ubuntu/xenial 64bit - this box has still some issues

    • ChunkBuffer: add test for leaving partial chunk in buffer, fixes #945

Version 1.0.13 (2019-02-15)

Please note: this is very likely the last 1.0.x release, please upgrade to 1.1.x.

Bug fixes:

  • security fix: configure FUSE with "default_permissions", #3903. "default_permissions" is now enforced by borg by default to let the kernel check uid/gid/mode based permissions. "ignore_permissions" can be given to not enforce "default_permissions".

  • xattrs: fix borg exception handling on ENOSPC error, #3808.

New features:

  • Read a passphrase from a file descriptor specified in the BORG_PASSPHRASE_FD environment variable.

Other changes:

  • acl platform code: fix acl set return type

  • xattr:

    • add linux {list,get,set}xattr ctypes prototypes

    • fix darwin flistxattr ctypes prototype

  • testing / travis-ci:

    • fix the homebrew 1.9 issues on travis-ci, #4254

    • travis OS X: use xcode 8.3 (not broken)

    • tox.ini: lock requirements

    • unbreak 1.0-maint on travis, fixes #4123

  • vagrant:

    • misc. fixes

    • FUSE for macOS: upgrade 3.7.1 to 3.8.3

    • Python: upgrade 3.5.5 to 3.5.6

  • docs:

    • Update installation instructions for macOS

    • update release workflow using twine (docs, scripts), #4213

Version 1.0.12 (2018-04-08)

Bug fixes:

  • repository: cleanup/write: invalidate cached FDs, tests

  • serve: fix exitcode, #2910

  • extract: set bsdflags last (include immutable flag), #3263

  • create --timestamp: set start time, #2957

  • create: show excluded dir with "x" for tagged dirs / caches, #3189

  • migrate locks to child PID when daemonize is used

  • Buffer: fix wrong thread-local storage use, #2951

  • fix detection of non-local path, #3108

  • fix LDLP restoration for subprocesses, #3077

  • fix subprocess environments (xattr module's fakeroot version check, borg umount, BORG_PASSCOMMAND), #3050

  • remote: deal with partial lines, #2637

  • get rid of datetime.isoformat, use safe parse_timestamp to parse timestamps, #2994

  • build: do .h file content checks in binary mode, fixes build issue for non-ascii header files on pure-ascii locale platforms, #3544 #3639

  • remove platform.uname() call which caused library mismatch issues, #3732

  • add exception handler around deprecated platform.linux_distribution() call

Other changes:

  • require msgpack-python >= 0.4.6 and < 0.5.0, see #3753

  • add parens for C preprocessor macro argument usages (did not cause malfunction)

  • ignore corrupt files cache, #2939

  • replace "modulo" with "if" to check for wraparound in hashmap

  • keymanager: don't depend on optional readline module, #2980

  • exclude broken pytest 3.3.0 release

  • exclude broken Cython 0.27(.0) release, #3066

  • flake8: add some ignores

  • docs:

    • create: document exclusion through nodump

    • document good and problematic option placements, fix examples, #3356

    • update docs about hardlinked symlinks limitation

    • faq: we do not implement futile attempts of ETA / progress displays

    • simplified rate limiting wrapper in FAQ

    • twitter account @borgbackup, #2948

    • add note about metadata dedup and --no[ac]time, #2518

    • change-passphrase only changes the passphrase, #2990

    • clarify encrypted key format for borg key export, #3296

    • document sshfs rename workaround, #3315

    • update release checklist about security fixes

    • docs about how to verify a signed release, #3634

    • chunk seed is generated per /repository/

  • vagrant:

    • use FUSE for macOS 3.7.1 to build the macOS binary

    • use python 3.5.5 to build the binaries

    • add exe location to PATH when we build an exe

    • use https pypi url for wheezy

    • netbsd: bash is already installed

    • netbsd: fix netbsd version in PKG_PATH

    • use self-made FreeBSD 10.3 box, #3022

    • backport fs_init (including related updates) from 1.1

    • the boxcutter wheezy boxes are 404, use local ones

  • travis:

    • don't perform full Travis build on docs-only changes, #2531

    • only short-circuit docs-only changes for pull requests

Version 1.0.11 (2017-07-21)

Bug fixes:

  • use limited unpacker for outer key (security precaution), #2174

  • fix paperkey import bug

Other changes:

  • change --checkpoint-interval default from 600s to 1800s, #2841. this improves efficiency for big repositories a lot.

  • docs: fix OpenSUSE command and add OpenSUSE section

  • tests: add tests for split_lstring and paperkey

  • vagrant:

    • fix openbsd shell

    • backport cpu/ram setup from master

    • add stretch64 VM

Version 1.0.11rc1 (2017-06-27)

Bug fixes:

  • performance: rebuild hashtable if we have too few empty buckets, #2246. this fixes some sporadic, but severe performance breakdowns.

  • Archive: allocate zeros when needed, #2308 fixes huge memory usage of mount (8 MiB × number of archives)

  • IPv6 address support also: Location: more informative exception when parsing fails

  • borg single-file binary: use pyinstaller v3.2.1, #2396 this fixes that the prelink cronjob on some distros kills the borg binary by stripping away parts of it.

  • extract:

    • warning for unextracted big extended attributes, #2258

    • also create parent dir for device files, if needed.

    • don't write to disk with --stdout, #2645

  • archive check: detect and fix missing all-zero replacement chunks, #2180

  • fix (de)compression exceptions, #2224 #2221

  • files cache: update inode number, #2226

  • borg rpc: use limited msgpack.Unpacker (security precaution), #2139

  • Manifest: use limited msgpack.Unpacker (security precaution), #2175

  • Location: accept //servername/share/path

  • fix ChunkIndex.__contains__ assertion for big-endian archs (harmless)

  • create: handle BackupOSError on a per-path level in one spot

  • fix error msg, there is no --keep-last in borg 1.0.x, #2282

  • clamp (nano)second values to unproblematic range, #2304

  • fuse / borg mount:

    • fix st_blocks to be an integer (not float) value

    • fix negative uid/gid crash (they could come into archives e.g. when backing up external drives under cygwin), #2674

    • fix crash if empty (None) xattr is read

    • do pre-mount checks before opening repository

    • check llfuse is installed before asking for passphrase

  • borg rename: expand placeholders, #2386

  • borg serve: fix forced command lines containing BORG_* env vars

  • fix error msg, it is --keep-within, not --within

  • fix borg key/debug/benchmark crashing without subcommand, #2240

  • chunker: fix invalid use of types, don't do uint32_t >> 32

  • document follow_symlinks requirements, check libc, #2507

New features:

  • added BORG_PASSCOMMAND environment variable, #2573

  • add minimal version of in repository mandatory feature flags, #2134

    This should allow us to make sure older borg versions can be cleanly prevented from doing operations that are no longer safe because of repository format evolution. This allows more fine grained control than just incrementing the manifest version. So for example a change that still allows new archives to be created but would corrupt the repository when an old version tries to delete an archive or check the repository would add the new feature to the check and delete set but leave it out of the write set.

  • borg delete --force --force to delete severely corrupted archives, #1975

Other changes:

  • embrace y2038 issue to support 32bit platforms

  • be more clear that this is a "beyond repair" case, #2427

  • key file names: limit to 100 characters and remove colons from host name

  • upgrade FUSE for macOS to 3.5.8, #2346

  • split up parsing and filtering for --keep-within, better error message, #2610

  • docs:

    • fix caskroom link, #2299

    • address SSH batch mode, #2202 #2270

    • improve remote-path description

    • document snapshot usage, #2178

    • document relative path usage, #1868

    • one link per distro in the installation page

    • development: new branching model in git repository

    • kill api page

    • added FAQ section about backing up root partition

    • add bountysource badge, #2558

    • create empty docs.txt reequirements, #2694

    • README: how to help the project

    • note -v/--verbose requirement on affected options, #2542

    • document borg init behaviour via append-only borg serve, #2440

    • be clear about what buzhash is used for (chunking) and want it is not used for (deduplication)- also say already in the readme that we use a cryptohash for dedupe, so people don't worry, #2390

    • add hint about chunker params to borg upgrade docs, #2421

    • clarify borg upgrade docs, #2436

    • quickstart: delete problematic BORG_PASSPHRASE use, #2623

    • faq: specify "using inline shell scripts"

    • document pattern denial of service, #2624

  • tests:

    • remove attic dependency of the tests, #2505

    • travis:

      • enhance travis setuptools_scm situation

      • install fakeroot for Linux

    • add test for borg delete --force

    • enable remote tests on cygwin (the cygwin issue that caused these tests to break was fixed in cygwin at least since cygwin 2.8, maybe even since 2.7.0).

    • remove skipping the noatime tests on GNU/Hurd, #2710

    • fix borg import issue, add comment, #2718

    • include attic.tar.gz when installing the package also: add include_package_data=True

Version 1.0.10 (2017-02-13)

Bug fixes:

  • Manifest timestamps are now monotonically increasing, this fixes issues when the system clock jumps backwards or is set inconsistently across computers accessing the same repository, #2115

  • Fixed testing regression in 1.0.10rc1 that lead to a hard dependency on py.test >= 3.0, #2112

New features:

  • "key export" can now generate a printable HTML page with both a QR code and a human-readable "paperkey" representation (and custom text) through the --qr-html option.

    The same functionality is also available through paperkey.html, which is the same HTML page generated by --qr-html. It works with existing "key export" files and key files.

Other changes:

  • docs:

    • language clarification - "borg create --one-file-system" option does not respect mount points, but considers different file systems instead, #2141

  • setup.py: build_api: sort file list for determinism

Version 1.0.10rc1 (2017-01-29)

Bug fixes:

  • borg serve: fix transmission data loss of pipe writes, #1268 This affects only the cygwin platform (not Linux, BSD, OS X).

  • Avoid triggering an ObjectiveFS bug in xattr retrieval, #1992

  • When running out of buffer memory when reading xattrs, only skip the current file, #1993

  • Fixed "borg upgrade --tam" crashing with unencrypted repositories. Since the issue is not relevant for unencrypted repositories, it now does nothing and prints an error, #1981.

  • Fixed change-passphrase crashing with unencrypted repositories, #1978

  • Fixed "borg check repo::archive" indicating success if "archive" does not exist, #1997

  • borg check: print non-exit-code warning if --last or --prefix aren't fulfilled

  • fix bad parsing of wrong repo location syntax

  • create: don't create hard link refs to failed files, mount: handle invalid hard link refs, #2092

  • detect mingw byte order, #2073

  • creating a new segment: use "xb" mode, #2099

  • mount: umount on SIGINT/^C when in foreground, #2082

Other changes:

  • binary: use fixed AND freshly compiled pyinstaller bootloader, #2002

  • xattr: ignore empty names returned by llistxattr(2) et al

  • Enable the fault handler: install handlers for the SIGSEGV, SIGFPE, SIGABRT, SIGBUS and SIGILL signals to dump the Python traceback.

  • Also print a traceback on SIGUSR2.

  • borg change-passphrase: print key location (simplify making a backup of it)

  • officially support Python 3.6 (setup.py: add Python 3.6 qualifier)

  • tests:

    • vagrant / travis / tox: add Python 3.6 based testing

    • vagrant: fix openbsd repo, #2042

    • vagrant: fix the freebsd64 machine, #2037 #2067

    • vagrant: use python 3.5.3 to build binaries, #2078

    • vagrant: use osxfuse 3.5.4 for tests / to build binaries vagrant: improve darwin64 VM settings

    • travis: fix osxfuse install (fixes OS X testing on Travis CI)

    • travis: require succeeding OS X tests, #2028

    • travis: use latest pythons for OS X based testing

    • use pytest-xdist to parallelize testing

    • fix xattr test race condition, #2047

    • setup.cfg: fix pytest deprecation warning, #2050

  • docs:

    • language clarification - VM backup FAQ

    • borg create: document how to backup stdin, #2013

    • borg upgrade: fix incorrect title levels

    • add CVE numbers for issues fixed in 1.0.9, #2106

  • fix typos (taken from Debian package patch)

  • remote: include data hexdump in "unexpected RPC data" error message

  • remote: log SSH command line at debug level

  • API_VERSION: use numberspaces, #2023

  • remove .github from pypi package, #2051

  • add pip and setuptools to requirements file, #2030

  • SyncFile: fix use of fd object after close (cosmetic)

  • Manifest.in: simplify, exclude *.{so,dll,orig}, #2066

  • ignore posix_fadvise errors in repository.py, #2095 (works around issues with docker on ARM)

  • make LoggedIO.close_segment reentrant, avoid reentrance

Version 1.0.9 (2016-12-20)

Security fixes:

  • A flaw in the cryptographic authentication scheme in Borg allowed an attacker to spoof the manifest. See Pre-1.0.9 manifest spoofing vulnerability (CVE-2016-10099) above for the steps you should take.

    CVE-2016-10099 was assigned to this vulnerability.

  • borg check: When rebuilding the manifest (which should only be needed very rarely) duplicate archive names would be handled on a "first come first serve" basis, allowing an attacker to apparently replace archives.

    CVE-2016-10100 was assigned to this vulnerability.

Bug fixes:

  • borg check:

    • rebuild manifest if it's corrupted

    • skip corrupted chunks during manifest rebuild

  • fix TypeError in integrity error handler, #1903, #1894

  • fix location parser for archives with @ char (regression introduced in 1.0.8), #1930

  • fix wrong duration/timestamps if system clock jumped during a create

  • fix progress display not updating if system clock jumps backwards

  • fix checkpoint interval being incorrect if system clock jumps

Other changes:

  • docs:

    • add python3-devel as a dependency for cygwin-based installation

    • clarify extract is relative to current directory

    • FAQ: fix link to changelog

    • markup fixes

  • tests:

    • test_get_(cache|keys)_dir: clean env state, #1897

    • get back pytest's pretty assertion failures, #1938

  • setup.py build_usage:

    • fixed build_usage not processing all commands

    • fixed build_usage not generating includes for debug commands

Version 1.0.9rc1 (2016-11-27)

Bug fixes:

  • files cache: fix determination of newest mtime in backup set (which is used in cache cleanup and led to wrong "A" [added] status for unchanged files in next backup), #1860.

  • borg check:

    • fix incorrectly reporting attic 0.13 and earlier archives as corrupt

    • handle repo w/o objects gracefully and also bail out early if repo is completely empty, #1815.

  • fix tox/pybuild in 1.0-maint

  • at xattr module import time, loggers are not initialized yet

New features:

  • borg umount <mountpoint> exposed already existing umount code via the CLI api, so users can use it, which is more consistent than using borg to mount and fusermount -u (or umount) to un-mount, #1855.

  • implement borg create --noatime --noctime, fixes #1853

Other changes:

  • docs:

    • display README correctly on PyPI

    • improve cache / index docs, esp. files cache docs, fixes #1825

    • different pattern matching for --exclude, #1779

    • datetime formatting examples for {now} placeholder, #1822

    • clarify passphrase mode attic repo upgrade, #1854

    • clarify --umask usage, #1859

    • clarify how to choose PR target branch

    • clarify prune behavior for different archive contents, #1824

    • fix PDF issues, add logo, fix authors, headings, TOC

    • move security verification to support section

    • fix links in standalone README (:ref: tags)

    • add link to security contact in README

    • add FAQ about security

    • move fork differences to FAQ

    • add more details about resource usage

  • tests: skip remote tests on cygwin, #1268

  • travis:

    • allow OS X failures until the brew cask osxfuse issue is fixed

    • caskroom osxfuse-beta gone, it's osxfuse now (3.5.3)

  • vagrant:

    • upgrade OSXfuse / FUSE for macOS to 3.5.3

    • remove llfuse from tox.ini at a central place

    • do not try to install llfuse on centos6

    • fix FUSE test for darwin, #1546

    • add windows virtual machine with cygwin

    • Vagrantfile cleanup / code deduplication

Version 1.0.8 (2016-10-29)

Bug fixes:

  • RemoteRepository: Fix busy wait in call_many, #940

New features:

  • implement borgmajor/borgminor/borgpatch placeholders, #1694 {borgversion} was already there (full version string). With the new placeholders you can now also get e.g. 1 or 1.0 or 1.0.8.

Other changes:

  • avoid previous_location mismatch, #1741

    due to the changed canonicalization for relative paths in PR #1711 / #1655 (implement /./ relpath hack), there would be a changed repo location warning and the user would be asked if this is ok. this would break automation and require manual intervention, which is unwanted.

    thus, we automatically fix the previous_location config entry, if it only changed in the expected way, but still means the same location.

  • docs:

    • deployment.rst: do not use bare variables in ansible snippet

    • add clarification about append-only mode, #1689

    • setup.py: add comment about requiring llfuse, #1726

    • update usage.rst / api.rst

    • repo url / archive location docs + typo fix

    • quickstart: add a comment about other (remote) filesystems

  • vagrant / tests:

    • no chown when rsyncing (fixes boxes w/o vagrant group)

    • fix FUSE permission issues on linux/freebsd, #1544

    • skip FUSE test for borg binary + fakeroot

    • ignore security.selinux xattrs, fixes tests on centos, #1735

Version 1.0.8rc1 (2016-10-17)

Bug fixes:

  • fix signal handling (SIGINT, SIGTERM, SIGHUP), #1620 #1593 Fixes e.g. leftover lock files for quickly repeated signals (e.g. Ctrl-C Ctrl-C) or lost connections or systemd sending SIGHUP.

  • progress display: adapt formatting to narrow screens, do not crash, #1628

  • borg create --read-special - fix crash on broken symlink, #1584. also correctly processes broken symlinks. before this regressed to a crash (5b45385) a broken symlink would've been skipped.

  • process_symlink: fix missing backup_io() Fixes a chmod/chown/chgrp/unlink/rename/... crash race between getting dirents and dispatching to process_symlink.

  • yes(): abort on wrong answers, saying so, #1622

  • fixed exception borg serve raised when connection was closed before repository was opened. Add an error message for this.

  • fix read-from-closed-FD issue, #1551 (this seems not to get triggered in 1.0.x, but was discovered in master)

  • hashindex: fix iterators (always raise StopIteration when exhausted) (this seems not to get triggered in 1.0.x, but was discovered in master)

  • enable relative paths in ssh:// repo URLs, via /./relpath hack, #1655

  • allow repo paths with colons, #1705

  • update changed repo location immediately after acceptance, #1524

  • fix debug get-obj / delete-obj crash if object not found and remote repo, #1684

  • pyinstaller: use a spec file to build borg.exe binary, exclude osxfuse dylib on Mac OS X (avoids mismatch lib <-> driver), #1619

New features:

  • add "borg key export" / "borg key import" commands, #1555, so users are able to backup / restore their encryption keys more easily.

    Supported formats are the keyfile format used by borg internally and a special "paper" format with by line checksums for printed backups. For the paper format, the import is an interactive process which checks each line as soon as it is input.

  • add "borg debug-refcount-obj" to determine a repo objects' referrer counts, #1352

Other changes:

  • add "borg debug ..." subcommands (borg debug-* still works, but will be removed in borg 1.1)

  • setup.py: Add subcommand support to build_usage.

  • remote: change exception message for unexpected RPC data format to indicate dataflow direction.

  • improved messages / error reporting:

    • IntegrityError: add placeholder for message, so that the message we give appears not only in the traceback, but also in the (short) error message, #1572

    • borg.key: include chunk id in exception msgs, #1571

    • better messages for cache newer than repo, #1700

  • vagrant (testing/build VMs):

    • upgrade OSXfuse / FUSE for macOS to 3.5.2

    • update Debian Wheezy boxes, #1686

    • openbsd / netbsd: use own boxes, fixes misc rsync installation and FUSE/llfuse related testing issues, #1695 #1696 #1670 #1671 #1728

  • docs:

    • add docs for "key export" and "key import" commands, #1641

    • fix inconsistency in FAQ (pv-wrapper).

    • fix second block in "Easy to use" section not showing on GitHub, #1576

    • add bestpractices badge

    • link reference docs and faq about BORG_FILES_CACHE_TTL, #1561

    • improve borg info --help, explain size infos, #1532

    • add release signing key / security contact to README, #1560

    • add contribution guidelines for developers

    • development.rst: add sphinx_rtd_theme to the sphinx install command

    • adjust border color in borg.css

    • add debug-info usage help file

    • internals.rst: fix typos

    • setup.py: fix build_usage to always process all commands

    • added docs explaining multiple --restrict-to-path flags, #1602

    • add more specific warning about write-access debug commands, #1587

    • clarify FAQ regarding backup of virtual machines, #1672

  • tests:

    • work around FUSE xattr test issue with recent fakeroot

    • simplify repo/hashindex tests

    • travis: test FUSE-enabled borg, use trusty to have a recent FUSE

    • re-enable FUSE tests for RemoteArchiver (no deadlocks any more)

    • clean env for pytest based tests, #1714

    • fuse_mount contextmanager: accept any options

Version 1.0.7 (2016-08-19)

Security fixes:

  • borg serve: fix security issue with remote repository access, #1428 If you used e.g. --restrict-to-path /path/client1/ (with or without trailing slash does not make a difference), it acted like a path prefix match using /path/client1 (note the missing trailing slash) - the code then also allowed working in e.g. /path/client13 or /path/client1000.

    As this could accidentally lead to major security/privacy issues depending on the paths you use, the behaviour was changed to be a strict directory match. That means --restrict-to-path /path/client1 (with or without trailing slash does not make a difference) now uses /path/client1/ internally (note the trailing slash here!) for matching and allows precisely that path AND any path below it. So, /path/client1 is allowed, /path/client1/repo1 is allowed, but not /path/client13 or /path/client1000.

    If you willingly used the undocumented (dangerous) previous behaviour, you may need to rearrange your --restrict-to-path paths now. We are sorry if that causes work for you, but we did not want a potentially dangerous behaviour in the software (not even using a for-backwards-compat option).

Bug fixes:

  • fixed repeated LockTimeout exceptions when borg serve tried to write into a already write-locked repo (e.g. by a borg mount), #502 part b) This was solved by the fix for #1220 in 1.0.7rc1 already.

  • fix cosmetics + file leftover for "not a valid borg repository", #1490

  • Cache: release lock if cache is invalid, #1501

  • borg extract --strip-components: fix leak of preloaded chunk contents

  • Repository, when a InvalidRepository exception happens:

    • fix spurious, empty lock.roster

    • fix repo not closed cleanly

New features:

  • implement borg debug-info, fixes #1122 (just calls already existing code via cli, same output as below tracebacks)

Other changes:

  • skip the O_NOATIME test on GNU Hurd, fixes #1315 (this is a very minor issue and the GNU Hurd project knows the bug)

  • document using a clean repo to test / build the release

Version 1.0.7rc2 (2016-08-13)

Bug fixes:

  • do not write objects to repository that are bigger than the allowed size, borg will reject reading them, #1451.

    Important: if you created archives with many millions of files or directories, please verify if you can open them successfully, e.g. try a "borg list REPO::ARCHIVE".

  • lz4 compression: dynamically enlarge the (de)compression buffer, the static buffer was not big enough for archives with extremely many items, #1453

  • larger item metadata stream chunks, raise archive item limit by 8x, #1452

  • fix untracked segments made by moved DELETEs, #1442

    Impact: Previously (metadata) segments could become untracked when deleting data, these would never be cleaned up.

  • extended attributes (xattrs) related fixes:

    • fixed a race condition in xattrs querying that led to the entire file not being backed up (while logging the error, exit code = 1), #1469

    • fixed a race condition in xattrs querying that led to a crash, #1462

    • raise OSError including the error message derived from errno, deal with path being a integer FD

Other changes:

  • print active env var override by default, #1467

  • xattr module: refactor code, deduplicate, clean up

  • repository: split object size check into too small and too big

  • add a transaction_id assertion, so borg init on a broken (inconsistent) filesystem does not look like a coding error in borg, but points to the real problem.

  • explain confusing TypeError caused by compat support for old servers, #1456

  • add forgotten usage help file from build_usage

  • refactor/unify buffer code into helpers.Buffer class, add tests

  • docs:

    • document archive limitation, #1452

    • improve prune examples

Version 1.0.7rc1 (2016-08-05)

Bug fixes:

  • fix repo lock deadlocks (related to lock upgrade), #1220

  • catch unpacker exceptions, resync, #1351

  • fix borg break-lock ignoring BORG_REPO env var, #1324

  • files cache performance fixes (fixes unnecessary re-reading/chunking/ hashing of unmodified files for some use cases):

    • fix unintended file cache eviction, #1430

    • implement BORG_FILES_CACHE_TTL, update FAQ, raise default TTL from 10 to 20, #1338

  • FUSE:

    • cache partially read data chunks (performance), #965, #966

    • always create a root dir, #1125

  • use an OrderedDict for helptext, making the build reproducible, #1346

  • RemoteRepository init: always call close on exceptions, #1370 (cosmetic)

  • ignore stdout/stderr broken pipe errors (cosmetic), #1116

New features:

  • better borg versions management support (useful esp. for borg servers wanting to offer multiple borg versions and for clients wanting to choose a specific server borg version), #1392:

    • add BORG_VERSION environment variable before executing "borg serve" via ssh

    • add new placeholder {borgversion}

    • substitute placeholders in --remote-path

  • borg init --append-only option (makes using the more secure append-only mode more convenient. when used remotely, this requires 1.0.7+ also on the borg server), #1291.

Other changes:

  • Vagrantfile:

    • darwin64: upgrade to FUSE for macOS 3.4.1 (aka osxfuse), #1378

    • xenial64: use user "ubuntu", not "vagrant" (as usual), #1331

  • tests:

    • fix FUSE tests on OS X, #1433

  • docs:

    • FAQ: add backup using stable filesystem names recommendation

    • FAQ about glibc compatibility added, #491, glibc-check improved

    • FAQ: 'A' unchanged file; remove ambiguous entry age sentence.

    • OS X: install pkg-config to build with FUSE support, fixes #1400

    • add notes about shell/sudo pitfalls with env. vars, #1380

    • added platform feature matrix

  • implement borg debug-dump-repo-objs

Version 1.0.6 (2016-07-12)

Bug fixes:

  • Linux: handle multiple LD_PRELOAD entries correctly, #1314, #1111

  • Fix crash with unclear message if the libc is not found, #1314, #1111

Other changes:

  • tests:

    • Fixed O_NOATIME tests for Solaris and GNU Hurd, #1315

    • Fixed sparse file tests for (file) systems not supporting it, #1310

  • docs:

    • Fixed syntax highlighting, #1313

    • misc docs: added data processing overview picture

Version 1.0.6rc1 (2016-07-10)

New features:

  • borg check --repair: heal damaged files if missing chunks re-appear (e.g. if the previously missing chunk was added again in a later backup archive), #148. (*) Also improved logging.

Bug fixes:

  • sync_dir: silence fsync() failing with EINVAL, #1287 Some network filesystems (like smbfs) don't support this and we use this in repository code.

  • borg mount (FUSE):

    • fix directories being shadowed when contained paths were also specified, #1295

    • raise I/O Error (EIO) on damaged files (unless -o allow_damaged_files is used), #1302. (*)

  • borg extract: warn if a damaged file is extracted, #1299. (*)

  • Added some missing return code checks (ChunkIndex._add, hashindex_resize).

  • borg check: fix/optimize initial hash table size, avoids resize of the table.

Other changes:

  • tests:

    • add more FUSE tests, #1284

    • deduplicate FUSE (u)mount code

    • fix borg binary test issues, #862

  • docs:

    • changelog: added release dates to older borg releases

    • fix some sphinx (docs generator) warnings, #881

Notes:

(*) Some features depend on information (chunks_healthy list) added to item metadata when a file with missing chunks was "repaired" using all-zero replacement chunks. The chunks_healthy list is generated since borg 1.0.4, thus borg can't recognize such "repaired" (but content-damaged) files if the repair was done with an older borg version.

Version 1.0.5 (2016-07-07)

Bug fixes:

  • borg mount: fix FUSE crash in xattr code on Linux introduced in 1.0.4, #1282

Other changes:

  • backport some FAQ entries from master branch

  • add release helper scripts

  • Vagrantfile:

    • centos6: no FUSE, don't build binary

    • add xz for redhat-like dists

Version 1.0.4 (2016-07-07)

New features:

  • borg serve --append-only, #1168 This was included because it was a simple change (append-only functionality was already present via repository config file) and makes better security now practically usable.

  • BORG_REMOTE_PATH environment variable, #1258 This was included because it was a simple change (--remote-path cli option was already present) and makes borg much easier to use if you need it.

  • Repository: cleanup incomplete transaction on "no space left" condition. In many cases, this can avoid a 100% full repo filesystem (which is very problematic as borg always needs free space - even to delete archives).

Bug fixes:

  • Fix wrong handling and reporting of OSErrors in borg create, #1138. This was a serious issue: in the context of "borg create", errors like repository I/O errors (e.g. disk I/O errors, ssh repo connection errors) were handled badly and did not lead to a crash (which would be good for this case, because the repo transaction would be incomplete and trigger a transaction rollback to clean up). Now, error handling for source files is cleanly separated from every other error handling, so only problematic input files are logged and skipped.

  • Implement fail-safe error handling for borg extract. Note that this isn't nearly as critical as the borg create error handling bug, since nothing is written to the repo. So this was "merely" misleading error reporting.

  • Add missing error handler in directory attr restore loop.

  • repo: make sure write data hits disk before the commit tag (#1236) and also sync the containing directory.

  • FUSE: getxattr fail must use errno.ENOATTR, #1126 (fixes Mac OS X Finder malfunction: "zero bytes" file length, access denied)

  • borg check --repair: do not lose information about the good/original chunks. If we do not lose the original chunk IDs list when "repairing" a file (replacing missing chunks with all-zero chunks), we have a chance to "heal" the file back into its original state later, in case the chunks re-appear (e.g. in a fresh backup). Healing is not implemented yet, see #148.

  • fixes for --read-special mode:

    • ignore known files cache, #1241

    • fake regular file mode, #1214

    • improve symlinks handling, #1215

  • remove passphrase from subprocess environment, #1105

  • Ignore empty index file (will trigger index rebuild), #1195

  • add missing placeholder support for --prefix, #1027

  • improve exception handling for placeholder replacement

  • catch and format exceptions in arg parsing

  • helpers: fix "undefined name 'e'" in exception handler

  • better error handling for missing repo manifest, #1043

  • borg delete:

    • make it possible to delete a repo without manifest

    • borg delete --forced allows one to delete corrupted archives, #1139

  • borg check:

    • make borg check work for empty repo

    • fix resync and msgpacked item qualifier, #1135

    • rebuild_manifest: fix crash if 'name' or 'time' key were missing.

    • better validation of item metadata dicts, #1130

    • better validation of archive metadata dicts

  • close the repo on exit - even if rollback did not work, #1197. This is rather cosmetic, it avoids repo closing in the destructor.

  • tests:

    • fix sparse file test, #1170

    • flake8: ignore new F405, #1185

    • catch "invalid argument" on cygwin, #257

    • fix sparseness assertion in test prep, #1264

Other changes:

  • make borg build/work on OpenSSL 1.0 and 1.1, #1187

  • docs / help:

    • fix / clarify prune help, #1143

    • fix "patterns" help formatting

    • add missing docs / help about placeholders

    • resources: rename atticmatic to borgmatic

    • document sshd settings, #545

    • more details about checkpoints, add split trick, #1171

    • support docs: add freenode web chat link, #1175

    • add prune visualization / example, #723

    • add note that Fnmatch is default, #1247

    • make clear that lzma levels > 6 are a waste of cpu cycles

    • add a "do not edit" note to auto-generated files, #1250

    • update cygwin installation docs

  • repository interoperability with borg master (1.1dev) branch:

    • borg check: read item metadata keys from manifest, #1147

    • read v2 hints files, #1235

    • fix hints file "unknown version" error handling bug

  • tests: add tests for format_line

  • llfuse: update version requirement for freebsd

  • Vagrantfile:

    • use openbsd 5.9, #716

    • do not install llfuse on netbsd (broken)

    • update OSXfuse to version 3.3.3

    • use Python 3.5.2 to build the binaries

  • glibc compatibility checker: scripts/glibc_check.py

  • add .eggs to .gitignore

Version 1.0.3 (2016-05-20)

Bug fixes:

  • prune: avoid that checkpoints are kept and completed archives are deleted in a prune run), #997

  • prune: fix commandline argument validation - some valid command lines were considered invalid (annoying, but harmless), #942

  • fix capabilities extraction on Linux (set xattrs last, after chown()), #1069

  • repository: fix commit tags being seen in data

  • when probing key files, do binary reads. avoids crash when non-borg binary files are located in borg's key files directory.

  • handle SIGTERM and make a clean exit - avoids orphan lock files.

  • repository cache: don't cache large objects (avoid using lots of temp. disk space), #1063

Other changes:

  • Vagrantfile: OS X: update osxfuse / install lzma package, #933

  • setup.py: add check for platform_darwin.c

  • setup.py: on freebsd, use a llfuse release that builds ok

  • docs / help:

    • update readthedocs URLs, #991

    • add missing docs for "borg break-lock", #992

    • borg create help: add some words to about the archive name

    • borg create help: document format tags, #894

Version 1.0.2 (2016-04-16)

Bug fixes:

  • fix malfunction and potential corruption on (nowadays rather rare) big-endian architectures or bi-endian archs in (rare) BE mode. #886, #889

    cache resync / index merge was malfunctioning due to this, potentially leading to data loss. borg info had cosmetic issues (displayed wrong values).

    note: all (widespread) little-endian archs (like x86/x64) or bi-endian archs in (widespread) LE mode (like ARMEL, MIPSEL, ...) were NOT affected.

  • add overflow and range checks for 1st (special) uint32 of the hashindex values, switch from int32 to uint32.

  • fix so that refcount will never overflow, but just stick to max. value after a overflow would have occurred.

  • borg delete: fix --cache-only for broken caches, #874

    Makes --cache-only idempotent: it won't fail if the cache is already deleted.

  • fixed borg create --one-file-system erroneously traversing into other filesystems (if starting fs device number was 0), #873

  • workaround a bug in Linux fadvise FADV_DONTNEED, #907

Other changes:

  • better test coverage for hashindex, incl. overflow testing, checking correct computations so endianness issues would be discovered.

  • reproducible doc for ProgressIndicator*, make the build reproducible.

  • use latest llfuse for vagrant machines

  • docs:

    • use /path/to/repo in examples, fixes #901

    • fix confusing usage of "repo" as archive name (use "arch")

Version 1.0.1 (2016-04-08)

New features:

Usually there are no new features in a bugfix release, but these were added due to their high impact on security/safety/speed or because they are fixes also:

  • append-only mode for repositories, #809, #36 (see docs)

  • borg create: add --ignore-inode option to make borg detect unmodified files even if your filesystem does not have stable inode numbers (like sshfs and possibly CIFS).

  • add options --warning, --error, --critical for missing log levels, #826. it's not recommended to suppress warnings or errors, but the user may decide this on his own. note: --warning is not given to borg serve so a <= 1.0.0 borg will still work as server (it is not needed as it is the default). do not use --error or --critical when using a <= 1.0.0 borg server.

Bug fixes:

  • fix silently skipping EIO, #748

  • add context manager for Repository (avoid orphan repository locks), #285

  • do not sleep for >60s while waiting for lock, #773

  • unpack file stats before passing to FUSE

  • fix build on illumos

  • don't try to backup doors or event ports (Solaris and derivatives)

  • remove useless/misleading libc version display, #738

  • test suite: reset exit code of persistent archiver, #844

  • RemoteRepository: clean up pipe if remote open() fails

  • Remote: don't print tracebacks for Error exceptions handled downstream, #792

  • if BORG_PASSPHRASE is present but wrong, don't prompt for password, but fail instead, #791

  • ArchiveChecker: move "orphaned objects check skipped" to INFO log level, #826

  • fix capitalization, add ellipses, change log level to debug for 2 messages, #798

Other changes:

  • update llfuse requirement, llfuse 1.0 works

  • update OS / dist packages on build machines, #717

  • prefer showing --info over -v in usage help, #859

  • docs:

    • fix cygwin requirements (gcc-g++)

    • document how to debug / file filesystem issues, #664

    • fix reproducible build of api docs

    • RTD theme: CSS !important overwrite, #727

    • Document logo font. Recreate logo png. Remove GIMP logo file.

Version 1.0.0 (2016-03-05)

The major release number change (0.x -> 1.x) indicates bigger incompatible changes, please read the compatibility notes, adapt / test your scripts and check your backup logs.

Compatibility notes:

  • drop support for python 3.2 and 3.3, require 3.4 or 3.5, #221 #65 #490 note: we provide binaries that include python 3.5.1 and everything else needed. they are an option in case you are stuck with < 3.4 otherwise.

  • change encryption to be on by default (using "repokey" mode)

  • moved keyfile keys from ~/.borg/keys to ~/.config/borg/keys, you can either move them manually or run "borg upgrade <REPO>"

  • remove support for --encryption=passphrase, use borg migrate-to-repokey to switch to repokey mode, #97

  • remove deprecated --compression <number>, use --compression zlib,<number> instead in case of 0, you could also use --compression none

  • remove deprecated --hourly/daily/weekly/monthly/yearly use --keep-hourly/daily/weekly/monthly/yearly instead

  • remove deprecated --do-not-cross-mountpoints, use --one-file-system instead

  • disambiguate -p option, #563:

    • -p now is same as --progress

    • -P now is same as --prefix

  • remove deprecated "borg verify", use "borg extract --dry-run" instead

  • cleanup environment variable semantics, #355 the environment variables used to be "yes sayers" when set, this was conceptually generalized to "automatic answerers" and they just give their value as answer (as if you typed in that value when being asked). See the "usage" / "Environment Variables" section of the docs for details.

  • change the builtin default for --chunker-params, create 2MiB chunks, #343 --chunker-params new default: 19,23,21,4095 - old default: 10,23,16,4095

    one of the biggest issues with borg < 1.0 (and also attic) was that it had a default target chunk size of 64kiB, thus it created a lot of chunks and thus also a huge chunk management overhead (high RAM and disk usage).

    please note that the new default won't change the chunks that you already have in your repository. the new big chunks do not deduplicate with the old small chunks, so expect your repo to grow at least by the size of every changed file and in the worst case (e.g. if your files cache was lost / is not used) by the size of every file (minus any compression you might use).

    in case you want to immediately see a much lower resource usage (RAM / disk) for chunks management, it might be better to start with a new repo than continuing in the existing repo (with an existing repo, you'ld have to wait until all archives with small chunks got pruned to see a lower resource usage).

    if you used the old --chunker-params default value (or if you did not use --chunker-params option at all) and you'ld like to continue using small chunks (and you accept the huge resource usage that comes with that), just explicitly use borg create --chunker-params=10,23,16,4095.

  • archive timestamps: the 'time' timestamp now refers to archive creation start time (was: end time), the new 'time_end' timestamp refers to archive creation end time. This might affect prune if your backups take rather long. if you give a timestamp via cli this is stored into 'time', therefore it now needs to mean archive creation start time.

New features:

  • implement password roundtrip, #695

Bug fixes:

  • remote end does not need cache nor keys directories, do not create them, #701

  • added retry counter for passwords, #703

Other changes:

  • fix compiler warnings, #697

  • docs:

    • update README.rst to new changelog location in docs/changes.rst

    • add Teemu to AUTHORS

    • changes.rst: fix old chunker params, #698

    • FAQ: how to limit bandwidth

Version 1.0.0rc2 (2016-02-28)

New features:

  • format options for location: user, pid, fqdn, hostname, now, utcnow, user

  • borg list --list-format

  • borg prune -v --list enables the keep/prune list output, #658

Bug fixes:

  • fix _open_rb noatime handling, #657

  • add a simple archivename validator, #680

  • borg create --stats: show timestamps in localtime, use same labels/formatting as borg info, #651

  • llfuse compatibility fixes (now compatible with: 0.40, 0.41, 0.42)

Other changes:

  • it is now possible to use "pip install borgbackup[fuse]" to automatically install the llfuse dependency using the correct version requirement for it. you still need to care about having installed the FUSE / build related OS package first, though, so that building llfuse can succeed.

  • Vagrant: drop Ubuntu Precise (12.04) - does not have Python >= 3.4

  • Vagrant: use pyinstaller v3.1.1 to build binaries

  • docs:

    • borg upgrade: add to docs that only LOCAL repos are supported

    • borg upgrade also handles borg 0.xx -> 1.0

    • use pip extras or requirements file to install llfuse

    • fix order in release process

    • updated usage docs and other minor / cosmetic fixes

    • verified borg examples in docs, #644

    • freebsd dependency installation and FUSE configuration, #649

    • add example how to restore a raw device, #671

    • add a hint about the dev headers needed when installing from source

    • add examples for delete (and handle delete after list, before prune), #656

    • update example for borg create -v --stats (use iso datetime format), #663

    • added example to BORG_RSH docs

    • "connection closed by remote": add FAQ entry and point to issue #636

Version 1.0.0rc1 (2016-02-07)

New features:

  • borg migrate-to-repokey ("passphrase" -> "repokey" encryption key mode)

  • implement --short for borg list REPO, #611

  • implement --list for borg extract (consistency with borg create)

  • borg serve: overwrite client's --restrict-to-path with ssh forced command's option value (but keep everything else from the client commandline), #544

  • use $XDG_CONFIG_HOME/keys for keyfile keys (~/.config/borg/keys), #515

  • "borg upgrade" moves the keyfile keys to the new location

  • display both archive creation start and end time in "borg info", #627

Bug fixes:

  • normalize trailing slashes for the repository path, #606

  • Cache: fix exception handling in __init__, release lock, #610

Other changes:

  • suppress unneeded exception context (PEP 409), simpler tracebacks

  • removed special code needed to deal with imperfections / incompatibilities / missing stuff in py 3.2/3.3, simplify code that can be done simpler in 3.4

  • removed some version requirements that were kept on old versions because newer did not support py 3.2 any more

  • use some py 3.4+ stdlib code instead of own/openssl/pypi code:

    • use os.urandom instead of own cython openssl RAND_bytes wrapper, #493

    • use hashlib.pbkdf2_hmac from py stdlib instead of own openssl wrapper

    • use hmac.compare_digest instead of == operator (constant time comparison)

    • use stat.filemode instead of homegrown code

    • use "mock" library from stdlib, #145

    • remove borg.support (with non-broken argparse copy), it is ok in 3.4+, #358

  • Vagrant: copy CHANGES.rst as symlink, #592

  • cosmetic code cleanups, add flake8 to tox/travis, #4

  • docs / help:

    • make "borg -h" output prettier, #591

    • slightly rephrase prune help

    • add missing example for --list option of borg create

    • quote exclude line that includes an asterisk to prevent shell expansion

    • fix dead link to license

    • delete Ubuntu Vivid, it is not supported anymore (EOL)

    • OS X binary does not work for older OS X releases, #629

    • borg serve's special support for forced/original ssh commands, #544

    • misc. updates and fixes

Version 0.30.0 (2016-01-23)

Compatibility notes:

  • you may need to use -v (or --info) more often to actually see output emitted at INFO log level (because it is suppressed at the default WARNING log level). See the "general" section in the usage docs.

  • for borg create, you need --list (additionally to -v) to see the long file list (was needed so you can have e.g. --stats alone without the long list)

  • see below about BORG_DELETE_I_KNOW_WHAT_I_AM_DOING (was: BORG_CHECK_I_KNOW_WHAT_I_AM_DOING)

Bug fixes:

  • fix crash when using borg create --dry-run --keep-tag-files, #570

  • make sure teardown with cleanup happens for Cache and RepositoryCache, avoiding leftover locks and TEMP dir contents, #285 (partially), #548

  • fix locking KeyError, partial fix for #502

  • log stats consistently, #526

  • add abbreviated weekday to timestamp format, fixes #496

  • strip whitespace when loading exclusions from file

  • unset LD_LIBRARY_PATH before invoking ssh, fixes strange OpenSSL library version warning when using the borg binary, #514

  • add some error handling/fallback for C library loading, #494

  • added BORG_DELETE_I_KNOW_WHAT_I_AM_DOING for check in "borg delete", #503

  • remove unused "repair" rpc method name

New features:

  • borg create: implement exclusions using regular expression patterns.

  • borg create: implement inclusions using patterns.

  • borg extract: support patterns, #361

  • support different styles for patterns:

    • fnmatch (fm: prefix, default when omitted), like borg <= 0.29.

    • shell (sh: prefix) with * not matching directory separators and **/ matching 0..n directories

    • path prefix (pp: prefix, for unifying borg create pp1 pp2 into the patterns system), semantics like in borg <= 0.29

    • regular expression (re:), new!

  • --progress option for borg upgrade (#291) and borg delete <archive>

  • update progress indication more often (e.g. for borg create within big files or for borg check repo), #500

  • finer chunker granularity for items metadata stream, #547, #487

  • borg create --list now used (additionally to -v) to enable the verbose file list output

  • display borg version below tracebacks, #532

Other changes:

  • hashtable size (and thus: RAM and disk consumption) follows a growth policy: grows fast while small, grows slower when getting bigger, #527

  • Vagrantfile: use pyinstaller 3.1 to build binaries, freebsd sqlite3 fix, fixes #569

  • no separate binaries for centos6 any more because the generic linux binaries also work on centos6 (or in general: on systems with a slightly older glibc than debian7

  • dev environment: require virtualenv<14.0 so we get a py32 compatible pip

  • docs:

    • add space-saving chunks.archive.d trick to FAQ

    • important: clarify -v and log levels in usage -> general, please read!

    • sphinx configuration: create a simple man page from usage docs

    • add a repo server setup example

    • disable unneeded SSH features in authorized_keys examples for security.

    • borg prune only knows "--keep-within" and not "--within"

    • add gource video to resources docs, #507

    • add netbsd install instructions

    • authors: make it more clear what refers to borg and what to attic

    • document standalone binary requirements, #499

    • rephrase the mailing list section

    • development docs: run build_api and build_usage before tagging release

    • internals docs: hash table max. load factor is 0.75 now

    • markup, typo, grammar, phrasing, clarifications and other fixes.

    • add gcc gcc-c++ to redhat/fedora/corora install docs, fixes #583

Version 0.29.0 (2015-12-13)

Compatibility notes:

  • when upgrading to 0.29.0 you need to upgrade client as well as server installations due to the locking and commandline interface changes otherwise you'll get an error msg about a RPC protocol mismatch or a wrong commandline option. if you run a server that needs to support both old and new clients, it is suggested that you have a "borg-0.28.2" and a "borg-0.29.0" command. clients then can choose via e.g. "borg --remote-path=borg-0.29.0 ...".

  • the default waiting time for a lock changed from infinity to 1 second for a better interactive user experience. if the repo you want to access is currently locked, borg will now terminate after 1s with an error message. if you have scripts that shall wait for the lock for a longer time, use --lock-wait N (with N being the maximum wait time in seconds).

Bug fixes:

  • hash table tuning (better chosen hashtable load factor 0.75 and prime initial size of 1031 gave ~1000x speedup in some scenarios)

  • avoid creation of an orphan lock for one case, #285

  • --keep-tag-files: fix file mode and multiple tag files in one directory, #432

  • fixes for "borg upgrade" (attic repo converter), #466

  • remove --progress isatty magic (and also --no-progress option) again, #476

  • borg init: display proper repo URL

  • fix format of umask in help pages, #463

New features:

  • implement --lock-wait, support timeout for UpgradableLock, #210

  • implement borg break-lock command, #157

  • include system info below traceback, #324

  • sane remote logging, remote stderr, #461:

    • remote log output: intercept it and log it via local logging system, with "Remote: " prefixed to message. log remote tracebacks.

    • remote stderr: output it to local stderr with "Remote: " prefixed.

  • add --debug and --info (same as --verbose) to set the log level of the builtin logging configuration (which otherwise defaults to warning), #426 note: there are few messages emitted at DEBUG level currently.

  • optionally configure logging via env var BORG_LOGGING_CONF

  • add --filter option for status characters: e.g. to show only the added or modified files (and also errors), use "borg create -v --filter=AME ...".

  • more progress indicators, #394

  • use ISO-8601 date and time format, #375

  • "borg check --prefix" to restrict archive checking to that name prefix, #206

Other changes:

  • hashindex_add C implementation (speed up cache re-sync for new archives)

  • increase FUSE read_size to 1024 (speed up metadata operations)

  • check/delete/prune --save-space: free unused segments quickly, #239

  • increase rpc protocol version to 2 (see also Compatibility notes), #458

  • silence borg by default (via default log level WARNING)

  • get rid of C compiler warnings, #391

  • upgrade OS X FUSE to 3.0.9 on the OS X binary build system

  • use python 3.5.1 to build binaries

  • docs:

    • new mailing list borgbackup@python.org, #468

    • readthedocs: color and logo improvements

    • load coverage icons over SSL (avoids mixed content)

    • more precise binary installation steps

    • update release procedure docs about OS X FUSE

    • FAQ entry about unexpected 'A' status for unchanged file(s), #403

    • add docs about 'E' file status

    • add "borg upgrade" docs, #464

    • add developer docs about output and logging

    • clarify encryption, add note about client-side encryption

    • add resources section, with videos, talks, presentations, #149

    • Borg moved to Arch Linux [community]

    • fix wrong installation instructions for archlinux

Version 0.28.2 (2015-11-15)

New features:

  • borg create --exclude-if-present TAGFILE - exclude directories that have the given file from the backup. You can additionally give --keep-tag-files to preserve just the directory roots and the tag-files (but not backup other directory contents), #395, attic #128, attic #142

Other changes:

  • do not create docs sources at build time (just have them in the repo), completely remove have_cython() hack, do not use the "mock" library at build time, #384

  • avoid hidden import, make it easier for PyInstaller, easier fix for #218

  • docs:

    • add description of item flags / status output, fixes #402

    • explain how to regenerate usage and API files (build_api or build_usage) and when to commit usage files directly into git, #384

    • minor install docs improvements

Version 0.28.1 (2015-11-08)

Bug fixes:

  • do not try to build api / usage docs for production install, fixes unexpected "mock" build dependency, #384

Other changes:

  • avoid using msgpack.packb at import time

  • fix formatting issue in changes.rst

  • fix build on readthedocs

Version 0.28.0 (2015-11-08)

Compatibility notes:

  • changed return codes (exit codes), see docs. in short: old: 0 = ok, 1 = error. now: 0 = ok, 1 = warning, 2 = error

New features:

  • refactor return codes (exit codes), fixes #61

  • add --show-rc option enable "terminating with X status, rc N" output, fixes 58, #351

  • borg create backups atime and ctime additionally to mtime, fixes #317 - extract: support atime additionally to mtime - FUSE: support ctime and atime additionally to mtime

  • support borg --version

  • emit a warning if we have a slow msgpack installed

  • borg list --prefix=thishostname- REPO, fixes #205

  • Debug commands (do not use except if you know what you do: debug-get-obj, debug-put-obj, debug-delete-obj, debug-dump-archive-items.

Bug fixes:

  • setup.py: fix bug related to BORG_LZ4_PREFIX processing

  • fix "check" for repos that have incomplete chunks, fixes #364

  • borg mount: fix unlocking of repository at umount time, fixes #331

  • fix reading files without touching their atime, #334

  • non-ascii ACL fixes for Linux, FreeBSD and OS X, #277

  • fix acl_use_local_uid_gid() and add a test for it, attic #359

  • borg upgrade: do not upgrade repositories in place by default, #299

  • fix cascading failure with the index conversion code, #269

  • borg check: implement 'cmdline' archive metadata value decoding, #311

  • fix RobustUnpacker, it missed some metadata keys (new atime and ctime keys were missing, but also bsdflags). add check for unknown metadata keys.

  • create from stdin: also save atime, ctime (cosmetic)

  • use default_notty=False for confirmations, fixes #345

  • vagrant: fix msgpack installation on centos, fixes #342

  • deal with unicode errors for symlinks in same way as for regular files and have a helpful warning message about how to fix wrong locale setup, fixes #382

  • add ACL keys the RobustUnpacker must know about

Other changes:

  • improve file size displays, more flexible size formatters

  • explicitly commit to the units standard, #289

  • archiver: add E status (means that an error occurred when processing this (single) item

  • do binary releases via "github releases", closes #214

  • create: use -x and --one-file-system (was: --do-not-cross-mountpoints), #296

  • a lot of changes related to using "logging" module and screen output, #233

  • show progress display if on a tty, output more progress information, #303

  • factor out status output so it is consistent, fix surrogates removal, maybe fixes #309

  • move away from RawConfigParser to ConfigParser

  • archive checker: better error logging, give chunk_id and sequence numbers (can be used together with borg debug-dump-archive-items).

  • do not mention the deprecated passphrase mode

  • emit a deprecation warning for --compression N (giving a just a number)

  • misc .coverragerc fixes (and coverage measurement improvements), fixes #319

  • refactor confirmation code, reduce code duplication, add tests

  • prettier error messages, fixes #307, #57

  • tests:

    • add a test to find disk-full issues, #327

    • travis: also run tests on Python 3.5

    • travis: use tox -r so it rebuilds the tox environments

    • test the generated pyinstaller-based binary by archiver unit tests, #215

    • vagrant: tests: announce whether fakeroot is used or not

    • vagrant: add vagrant user to fuse group for debianoid systems also

    • vagrant: llfuse install on darwin needs pkgconfig installed

    • vagrant: use pyinstaller from develop branch, fixes #336

    • benchmarks: test create, extract, list, delete, info, check, help, fixes #146

    • benchmarks: test with both the binary and the python code

    • archiver tests: test with both the binary and the python code, fixes #215

    • make basic test more robust

  • docs:

    • moved docs to borgbackup.readthedocs.org, #155

    • a lot of fixes and improvements, use mobile-friendly RTD standard theme

    • use zlib,6 compression in some examples, fixes #275

    • add missing rename usage to docs, closes #279

    • include the help offered by borg help <topic> in the usage docs, fixes #293

    • include a list of major changes compared to attic into README, fixes #224

    • add OS X install instructions, #197

    • more details about the release process, #260

    • fix linux glibc requirement (binaries built on debian7 now)

    • build: move usage and API generation to setup.py

    • update docs about return codes, #61

    • remove api docs (too much breakage on rtd)

    • borgbackup install + basics presentation (asciinema)

    • describe the current style guide in documentation

    • add section about debug commands

    • warn about not running out of space

    • add example for rename

    • improve chunker params docs, fixes #362

    • minor development docs update

Version 0.27.0 (2015-10-07)

New features:

  • "borg upgrade" command - attic -> borg one time converter / migration, #21

  • temporary hack to avoid using lots of disk space for chunks.archive.d, #235: To use it: rm -rf chunks.archive.d ; touch chunks.archive.d

  • respect XDG_CACHE_HOME, attic #181

  • add support for arbitrary SSH commands, attic #99

  • borg delete --cache-only REPO (only delete cache, not REPO), attic #123

Bug fixes:

  • use Debian 7 (wheezy) to build pyinstaller borgbackup binaries, fixes slow down observed when running the Centos6-built binary on Ubuntu, #222

  • do not crash on empty lock.roster, fixes #232

  • fix multiple issues with the cache config version check, #234

  • fix segment entry header size check, attic #352 plus other error handling improvements / code deduplication there.

  • always give segment and offset in repo IntegrityErrors

Other changes:

  • stop producing binary wheels, remove docs about it, #147

  • docs: - add warning about prune - generate usage include files only as needed - development docs: add Vagrant section - update / improve / reformat FAQ - hint to single-file pyinstaller binaries from README

Version 0.26.1 (2015-09-28)

This is a minor update, just docs and new pyinstaller binaries.

  • docs update about python and binary requirements

  • better docs for --read-special, fix #220

  • re-built the binaries, fix #218 and #213 (glibc version issue)

  • update web site about single-file pyinstaller binaries

Note: if you did a python-based installation, there is no need to upgrade.

Version 0.26.0 (2015-09-19)

New features:

  • Faster cache sync (do all in one pass, remove tar/compression stuff), #163

  • BORG_REPO env var to specify the default repo, #168

  • read special files as if they were regular files, #79

  • implement borg create --dry-run, attic issue #267

  • Normalize paths before pattern matching on OS X, #143

  • support OpenBSD and NetBSD (except xattrs/ACLs)

  • support / run tests on Python 3.5

Bug fixes:

  • borg mount repo: use absolute path, attic #200, attic #137

  • chunker: use off_t to get 64bit on 32bit platform, #178

  • initialize chunker fd to -1, so it's not equal to STDIN_FILENO (0)

  • fix reaction to "no" answer at delete repo prompt, #182

  • setup.py: detect lz4.h header file location

  • to support python < 3.2.4, add less buggy argparse lib from 3.2.6 (#194)

  • fix for obtaining char * from temporary Python value (old code causes a compile error on Mint 17.2)

  • llfuse 0.41 install troubles on some platforms, require < 0.41 (UnicodeDecodeError exception due to non-ascii llfuse setup.py)

  • cython code: add some int types to get rid of unspecific python add / subtract operations (avoid undefined symbol FPE_... error on some platforms)

  • fix verbose mode display of stdin backup

  • extract: warn if a include pattern never matched, fixes #209, implement counters for Include/ExcludePatterns

  • archive names with slashes are invalid, attic issue #180

  • chunker: add a check whether the POSIX_FADV_DONTNEED constant is defined - fixes building on OpenBSD.

Other changes:

  • detect inconsistency / corruption / hash collision, #170

  • replace versioneer with setuptools_scm, #106

  • docs:

    • pkg-config is needed for llfuse installation

    • be more clear about pruning, attic issue #132

  • unit tests:

    • xattr: ignore security.selinux attribute showing up

    • ext3 seems to need a bit more space for a sparse file

    • do not test lzma level 9 compression (avoid MemoryError)

    • work around strange mtime granularity issue on netbsd, fixes #204

    • ignore st_rdev if file is not a block/char device, fixes #203

    • stay away from the setgid and sticky mode bits

  • use Vagrant to do easy cross-platform testing (#196), currently:

    • Debian 7 "wheezy" 32bit, Debian 8 "jessie" 64bit

    • Ubuntu 12.04 32bit, Ubuntu 14.04 64bit

    • Centos 7 64bit

    • FreeBSD 10.2 64bit

    • OpenBSD 5.7 64bit

    • NetBSD 6.1.5 64bit

    • Darwin (OS X Yosemite)

Version 0.25.0 (2015-08-29)

Compatibility notes:

  • lz4 compression library (liblz4) is a new requirement (#156)

  • the new compression code is very compatible: as long as you stay with zlib compression, older borg releases will still be able to read data from a repo/archive made with the new code (note: this is not the case for the default "none" compression, use "zlib,0" if you want a "no compression" mode that can be read by older borg). Also the new code is able to read repos and archives made with older borg versions (for all zlib levels 0..9).

Deprecations:

  • --compression N (with N being a number, as in 0.24) is deprecated. We keep the --compression 0..9 for now to not break scripts, but it is deprecated and will be removed later, so better fix your scripts now: --compression 0 (as in 0.24) is the same as --compression zlib,0 (now). BUT: if you do not want compression, you rather want --compression none (which is the default). --compression 1 (in 0.24) is the same as --compression zlib,1 (now) --compression 9 (in 0.24) is the same as --compression zlib,9 (now)

New features:

  • create --compression none (default, means: do not compress, just pass through data "as is". this is more efficient than zlib level 0 as used in borg 0.24)

  • create --compression lz4 (super-fast, but not very high compression)

  • create --compression zlib,N (slower, higher compression, default for N is 6)

  • create --compression lzma,N (slowest, highest compression, default N is 6)

  • honor the nodump flag (UF_NODUMP) and do not backup such items

  • list --short just outputs a simple list of the files/directories in an archive

Bug fixes:

  • fixed --chunker-params parameter order confusion / malfunction, fixes #154

  • close fds of segments we delete (during compaction)

  • close files which fell out the lrucache

  • fadvise DONTNEED now is only called for the byte range actually read, not for the whole file, fixes #158.

  • fix issue with negative "all archives" size, fixes #165

  • restore_xattrs: ignore if setxattr fails with EACCES, fixes #162

Other changes:

  • remove fakeroot requirement for tests, tests run faster without fakeroot (test setup does not fail any more without fakeroot, so you can run with or without fakeroot), fixes #151 and #91.

  • more tests for archiver

  • recover_segment(): don't assume we have an fd for segment

  • lrucache refactoring / cleanup, add dispose function, py.test tests

  • generalize hashindex code for any key length (less hardcoding)

  • lock roster: catch file not found in remove() method and ignore it

  • travis CI: use requirements file

  • improved docs:

    • replace hack for llfuse with proper solution (install libfuse-dev)

    • update docs about compression

    • update development docs about fakeroot

    • internals: add some words about lock files / locking system

    • support: mention BountySource and for what it can be used

    • theme: use a lighter green

    • add pypi, wheel, dist package based install docs

    • split install docs into system-specific preparations and generic instructions

Version 0.24.0 (2015-08-09)

Incompatible changes (compared to 0.23):

  • borg now always issues --umask NNN option when invoking another borg via ssh on the repository server. By that, it's making sure it uses the same umask for remote repos as for local ones. Because of this, you must upgrade both server and client(s) to 0.24.

  • the default umask is 077 now (if you do not specify via --umask) which might be a different one as you used previously. The default umask avoids that you accidentally give access permissions for group and/or others to files created by borg (e.g. the repository).

Deprecations:

  • "--encryption passphrase" mode is deprecated, see #85 and #97. See the new "--encryption repokey" mode for a replacement.

New features:

  • borg create --chunker-params ... to configure the chunker, fixes #16 (attic #302, attic #300, and somehow also #41). This can be used to reduce memory usage caused by chunk management overhead, so borg does not create a huge chunks index/repo index and eats all your RAM if you back up lots of data in huge files (like VM disk images). See docs/misc/create_chunker-params.txt for more information.

  • borg info now reports chunk counts in the chunk index.

  • borg create --compression 0..9 to select zlib compression level, fixes #66 (attic #295).

  • borg init --encryption repokey (to store the encryption key into the repo), fixes #85

  • improve at-end error logging, always log exceptions and set exit_code=1

  • LoggedIO: better error checks / exceptions / exception handling

  • implement --remote-path to allow non-default-path borg locations, #125

  • implement --umask M and use 077 as default umask for better security, #117

  • borg check: give a named single archive to it, fixes #139

  • cache sync: show progress indication

  • cache sync: reimplement the chunk index merging in C

Bug fixes:

  • fix segfault that happened for unreadable files (chunker: n needs to be a signed size_t), #116

  • fix the repair mode, #144

  • repo delete: add destroy to allowed rpc methods, fixes issue #114

  • more compatible repository locking code (based on mkdir), maybe fixes #92 (attic #317, attic #201).

  • better Exception msg if no Borg is installed on the remote repo server, #56

  • create a RepositoryCache implementation that can cope with >2GiB, fixes attic #326.

  • fix Traceback when running check --repair, attic #232

  • clarify help text, fixes #73.

  • add help string for --no-files-cache, fixes #140

Other changes:

  • improved docs:

    • added docs/misc directory for misc. writeups that won't be included "as is" into the html docs.

    • document environment variables and return codes (attic #324, attic #52)

    • web site: add related projects, fix web site url, IRC #borgbackup

    • Fedora/Fedora-based install instructions added to docs

    • Cygwin-based install instructions added to docs

    • updated AUTHORS

    • add FAQ entries about redundancy / integrity

    • clarify that borg extract uses the cwd as extraction target

    • update internals doc about chunker params, memory usage and compression

    • added docs about development

    • add some words about resource usage in general

    • document how to backup a raw disk

    • add note about how to run borg from virtual env

    • add solutions for (ll)fuse installation problems

    • document what borg check does, fixes #138

    • reorganize borgbackup.github.io sidebar, prev/next at top

    • deduplicate and refactor the docs / README.rst

  • use borg-tmp as prefix for temporary files / directories

  • short prune options without "keep-" are deprecated, do not suggest them

  • improved tox configuration

  • remove usage of unittest.mock, always use mock from pypi

  • use entrypoints instead of scripts, for better use of the wheel format and modern installs

  • add requirements.d/development.txt and modify tox.ini

  • use travis-ci for testing based on Linux and (new) OS X

  • use coverage.py, pytest-cov and codecov.io for test coverage support

I forgot to list some stuff already implemented in 0.23.0, here they are:

New features:

  • efficient archive list from manifest, meaning a big speedup for slow repo connections and "list <repo>", "delete <repo>", "prune" (attic #242, attic #167)

  • big speedup for chunks cache sync (esp. for slow repo connections), fixes #18

  • hashindex: improve error messages

Other changes:

  • explicitly specify binary mode to open binary files

  • some easy micro optimizations

Version 0.23.0 (2015-06-11)

Incompatible changes (compared to attic, fork related):

  • changed sw name and cli command to "borg", updated docs

  • package name (and name in urls) uses "borgbackup" to have fewer collisions

  • changed repo / cache internal magic strings from ATTIC* to BORG*, changed cache location to .cache/borg/ - this means that it currently won't accept attic repos (see issue #21 about improving that)

Bug fixes:

  • avoid defect python-msgpack releases, fixes attic #171, fixes attic #185

  • fix traceback when trying to do unsupported passphrase change, fixes attic #189

  • datetime does not like the year 10.000, fixes attic #139

  • fix "info" all archives stats, fixes attic #183

  • fix parsing with missing microseconds, fixes attic #282

  • fix misleading hint the fuse ImportError handler gave, fixes attic #237

  • check unpacked data from RPC for tuple type and correct length, fixes attic #127

  • fix Repository._active_txn state when lock upgrade fails

  • give specific path to xattr.is_enabled(), disable symlink setattr call that always fails

  • fix test setup for 32bit platforms, partial fix for attic #196

  • upgraded versioneer, PEP440 compliance, fixes attic #257

New features:

  • less memory usage: add global option --no-cache-files

  • check --last N (only check the last N archives)

  • check: sort archives in reverse time order

  • rename repo::oldname newname (rename repository)

  • create -v output more informative

  • create --progress (backup progress indicator)

  • create --timestamp (utc string or reference file/dir)

  • create: if "-" is given as path, read binary from stdin

  • extract: if --stdout is given, write all extracted binary data to stdout

  • extract --sparse (simple sparse file support)

  • extra debug information for 'fread failed'

  • delete <repo> (deletes whole repo + local cache)

  • FUSE: reflect deduplication in allocated blocks

  • only allow whitelisted RPC calls in server mode

  • normalize source/exclude paths before matching

  • use posix_fadvise to not spoil the OS cache, fixes attic #252

  • toplevel error handler: show tracebacks for better error analysis

  • sigusr1 / sigint handler to print current file infos - attic PR #286

  • RPCError: include the exception args we get from remote

Other changes:

  • source: misc. cleanups, pep8, style

  • docs and faq improvements, fixes, updates

  • cleanup crypto.pyx, make it easier to adapt to other AES modes

  • do os.fsync like recommended in the python docs

  • source: Let chunker optionally work with os-level file descriptor.

  • source: Linux: remove duplicate os.fsencode calls

  • source: refactor _open_rb code a bit, so it is more consistent / regular

  • source: refactor indicator (status) and item processing

  • source: use py.test for better testing, flake8 for code style checks

  • source: fix tox >=2.0 compatibility (test runner)

  • pypi package: add python version classifiers, add FreeBSD to platforms

Attic Changelog

Here you can see the full list of changes between each Attic release until Borg forked from Attic:

Version 0.17

(bugfix release, released on X)

  • Fix hashindex ARM memory alignment issue (#309)

  • Improve hashindex error messages (#298)

Version 0.16

(bugfix release, released on May 16, 2015)

  • Fix typo preventing the security confirmation prompt from working (#303)

  • Improve handling of systems with improperly configured file system encoding (#289)

  • Fix "All archives" output for attic info. (#183)

  • More user friendly error message when repository key file is not found (#236)

  • Fix parsing of iso 8601 timestamps with zero microseconds (#282)

Version 0.15

(bugfix release, released on Apr 15, 2015)

  • xattr: Be less strict about unknown/unsupported platforms (#239)

  • Reduce repository listing memory usage (#163).

  • Fix BrokenPipeError for remote repositories (#233)

  • Fix incorrect behavior with two character directory names (#265, #268)

  • Require approval before accessing relocated/moved repository (#271)

  • Require approval before accessing previously unknown unencrypted repositories (#271)

  • Fix issue with hash index files larger than 2GB.

  • Fix Python 3.2 compatibility issue with noatime open() (#164)

  • Include missing pyx files in dist files (#168)

Version 0.14

(feature release, released on Dec 17, 2014)

  • Added support for stripping leading path segments (#95) "attic extract --strip-segments X"

  • Add workaround for old Linux systems without acl_extended_file_no_follow (#96)

  • Add MacPorts' path to the default openssl search path (#101)

  • HashIndex improvements, eliminates unnecessary IO on low memory systems.

  • Fix "Number of files" output for attic info. (#124)

  • limit create file permissions so files aren't read while restoring

  • Fix issue with empty xattr values (#106)

Version 0.13

(feature release, released on Jun 29, 2014)

  • Fix sporadic "Resource temporarily unavailable" when using remote repositories

  • Reduce file cache memory usage (#90)

  • Faster AES encryption (utilizing AES-NI when available)

  • Experimental Linux, OS X and FreeBSD ACL support (#66)

  • Added support for backup and restore of BSDFlags (OSX, FreeBSD) (#56)

  • Fix bug where xattrs on symlinks were not correctly restored

  • Added cachedir support. CACHEDIR.TAG compatible cache directories can now be excluded using --exclude-caches (#74)

  • Fix crash on extreme mtime timestamps (year 2400+) (#81)

  • Fix Python 3.2 specific lockf issue (EDEADLK)

Version 0.12

(feature release, released on April 7, 2014)

  • Python 3.4 support (#62)

  • Various documentation improvements a new style

  • attic mount now supports mounting an entire repository not only individual archives (#59)

  • Added option to restrict remote repository access to specific path(s): attic serve --restrict-to-path X (#51)

  • Include "all archives" size information in "--stats" output. (#54)

  • Added --stats option to attic delete and attic prune

  • Fixed bug where attic prune used UTC instead of the local time zone when determining which archives to keep.

  • Switch to SI units (Power of 1000 instead 1024) when printing file sizes

Version 0.11

(feature release, released on March 7, 2014)

  • New "check" command for repository consistency checking (#24)

  • Documentation improvements

  • Fix exception during "attic create" with repeated files (#39)

  • New "--exclude-from" option for attic create/extract/verify.

  • Improved archive metadata deduplication.

  • "attic verify" has been deprecated. Use "attic extract --dry-run" instead.

  • "attic prune --hourly|daily|..." has been deprecated. Use "attic prune --keep-hourly|daily|..." instead.

  • Ignore xattr errors during "extract" if not supported by the filesystem. (#46)

Version 0.10

(bugfix release, released on Jan 30, 2014)

  • Fix deadlock when extracting 0 sized files from remote repositories

  • "--exclude" wildcard patterns are now properly applied to the full path not just the file name part (#5).

  • Make source code endianness agnostic (#1)

Version 0.9

(feature release, released on Jan 23, 2014)

  • Remote repository speed and reliability improvements.

  • Fix sorting of segment names to ignore NFS left over files. (#17)

  • Fix incorrect display of time (#13)

  • Improved error handling / reporting. (#12)

  • Use fcntl() instead of flock() when locking repository/cache. (#15)

  • Let ssh figure out port/user if not specified so we don't override .ssh/config (#9)

  • Improved libcrypto path detection (#23).

Version 0.8.1

(bugfix release, released on Oct 4, 2013)

  • Fix segmentation fault issue.

Version 0.8

(feature release, released on Oct 3, 2013)

  • Fix xattr issue when backing up sshfs filesystems (#4)

  • Fix issue with excessive index file size (#6)

  • Support access of read only repositories.

  • New syntax to enable repository encryption:

    attic init --encryption="none|passphrase|keyfile".

  • Detect and abort if repository is older than the cache.

Version 0.7

(feature release, released on Aug 5, 2013)

  • Ported to FreeBSD

  • Improved documentation

  • Experimental: Archives mountable as FUSE filesystems.

  • The "user." prefix is no longer stripped from xattrs on Linux

Version 0.6.1

(bugfix release, released on July 19, 2013)

  • Fixed an issue where mtime was not always correctly restored.

Version 0.6

First public release on July 9, 2013

Internals

The internals chapter describes and analyses most of the inner workings of Borg.

Borg uses a low-level, key-value store, the Repository, and implements a more complex data structure on top of it, which is made up of the manifest, archives, items and data Chunks.

Each repository can hold multiple archives, which represent individual backups that contain a full archive of the files specified when the backup was performed.

Deduplication is performed globally across all data in the repository (multiple backups and even multiple hosts), both on data and file metadata, using Chunks created by the chunker using the Buzhash algorithm ("buzhash" chunker) or a simpler fixed blocksize algorithm ("fixed" chunker).

To actually perform the repository-wide deduplication, a hash of each chunk is checked against the chunks cache, which is a hash-table of all chunks that already exist.

_images/structure.png

Layers in Borg. On the very top commands are implemented, using a data access layer provided by the Archive and Item classes. The "key" object provides both compression and authenticated encryption used by the data access layer. The "key" object represents the sole trust boundary in Borg. The lowest layer is the repository, either accessed directly (Repository) or remotely (RemoteRepository).

Security

Cryptography in Borg
Attack model

The attack model of Borg is that the environment of the client process (e.g. borg create) is trusted and the repository (server) is not. The attacker has any and all access to the repository, including interactive manipulation (man-in-the-middle) for remote repositories.

Furthermore the client environment is assumed to be persistent across attacks (practically this means that the security database cannot be deleted between attacks).

Under these circumstances Borg guarantees that the attacker cannot

  1. modify the data of any archive without the client detecting the change

  2. rename, remove or add an archive without the client detecting the change

  3. recover plain-text data

  4. recover definite (heuristics based on access patterns are possible) structural information such as the object graph (which archives refer to what chunks)

The attacker can always impose a denial of service per definition (he could forbid connections to the repository, or delete it entirely).

When the above attack model is extended to include multiple clients independently updating the same repository, then Borg fails to provide confidentiality (i.e. guarantees 3) and 4) do not apply any more).

Structural Authentication

Borg is fundamentally based on an object graph structure (see Internals), where the root object is called the manifest.

Borg follows the Horton principle, which states that not only the message must be authenticated, but also its meaning (often expressed through context), because every object used is referenced by a parent object through its object ID up to the manifest. The object ID in Borg is a MAC of the object's plaintext, therefore this ensures that an attacker cannot change the context of an object without forging the MAC.

In other words, the object ID itself only authenticates the plaintext of the object and not its context or meaning. The latter is established by a different object referring to an object ID, thereby assigning a particular meaning to an object. For example, an archive item contains a list of object IDs that represent packed file metadata. On their own it's not clear that these objects would represent what they do, but by the archive item referring to them in a particular part of its own data structure assigns this meaning.

This results in a directed acyclic graph of authentication from the manifest to the data chunks of individual files.

Authenticating the manifest

Since the manifest has a fixed ID (000...000) the aforementioned authentication does not apply to it, indeed, cannot apply to it; it is impossible to authenticate the root node of a DAG through its edges, since the root node has no incoming edges.

With the scheme as described so far an attacker could easily replace the manifest, therefore Borg includes a tertiary authentication mechanism (TAM) that is applied to the manifest since version 1.0.9 (see Pre-1.0.9 manifest spoofing vulnerability (CVE-2016-10099)).

TAM works by deriving a separate key through HKDF from the other encryption and authentication keys and calculating the HMAC of the metadata to authenticate [1]:

# RANDOM(n) returns n random bytes
salt = RANDOM(64)

ikm = id_key || enc_key || enc_hmac_key
# *context* depends on the operation, for manifest authentication it is
# the ASCII string "borg-metadata-authentication-manifest".
tam_key = HKDF-SHA-512(ikm, salt, context)

# *data* is a dict-like structure
data[hmac] = zeroes
packed = pack(data)
data[hmac] = HMAC(tam_key, packed)
packed_authenticated = pack(data)

Since an attacker cannot gain access to this key and also cannot make the client authenticate arbitrary data using this mechanism, the attacker is unable to forge the authentication.

This effectively 'anchors' the manifest to the key, which is controlled by the client, thereby anchoring the entire DAG, making it impossible for an attacker to add, remove or modify any part of the DAG without Borg being able to detect the tampering.

Note that when using BORG_PASSPHRASE the attacker cannot swap the entire repository against a new repository with e.g. repokey mode and no passphrase, because Borg will abort access when BORG_PASSPHRASE is incorrect.

However, interactively a user might not notice this kind of attack immediately, if she assumes that the reason for the absent passphrase prompt is a set BORG_PASSPHRASE. See issue #2169 for details.

Encryption

Encryption is currently based on the Encrypt-then-MAC construction, which is generally seen as the most robust way to create an authenticated encryption scheme from encryption and message authentication primitives.

Every operation (encryption, MAC / authentication, chunk ID derivation) uses independent, random keys generated by os.urandom [2].

Borg does not support unauthenticated encryption -- only authenticated encryption schemes are supported. No unauthenticated encryption schemes will be added in the future.

Depending on the chosen mode (see borg init) different primitives are used:

  • The actual encryption is currently always AES-256 in CTR mode. The counter is added in plaintext, since it is needed for decryption, and is also tracked locally on the client to avoid counter reuse.

  • The authentication primitive is either HMAC-SHA-256 or BLAKE2b-256 in a keyed mode.

    Both HMAC-SHA-256 and BLAKE2b have undergone extensive cryptanalysis and have proven secure against known attacks. The known vulnerability of SHA-256 against length extension attacks does not apply to HMAC-SHA-256.

    The authentication primitive should be chosen based upon SHA hardware support: all AMD Ryzen, Intel 10th+ generation mobile and Intel 11th+ generation desktop processors, Apple M1+ and most current ARM64 architectures support SHA extensions and are likely to perform best with HMAC-SHA-256. 64-bit CPUs without SHA extensions are likely to perform best with BLAKE2b.

  • The primitive used for authentication is always the same primitive that is used for deriving the chunk ID, but they are always used with independent keys.

Encryption:

id = AUTHENTICATOR(id_key, data)
compressed = compress(data)

iv = reserve_iv()
encrypted = AES-256-CTR(enc_key, 8-null-bytes || iv, compressed)
authenticated = type-byte || AUTHENTICATOR(enc_hmac_key, encrypted) || iv || encrypted

Decryption:

# Given: input *authenticated* data, possibly a *chunk-id* to assert
type-byte, mac, iv, encrypted = SPLIT(authenticated)

ASSERT(type-byte is correct)
ASSERT( CONSTANT-TIME-COMPARISON( mac, AUTHENTICATOR(enc_hmac_key, encrypted) ) )

decrypted = AES-256-CTR(enc_key, 8-null-bytes || iv, encrypted)
decompressed = decompress(decrypted)

ASSERT( CONSTANT-TIME-COMPARISON( chunk-id, AUTHENTICATOR(id_key, decompressed) ) )

The client needs to track which counter values have been used, since encrypting a chunk requires a starting counter value and no two chunks may have overlapping counter ranges (otherwise the bitwise XOR of the overlapping plaintexts is revealed).

The client does not directly track the counter value, because it changes often (with each encrypted chunk), instead it commits a "reservation" to the security database and the repository by taking the current counter value and adding 4 GiB / 16 bytes (the block size) to the counter. Thus the client only needs to commit a new reservation every few gigabytes of encrypted data.

This mechanism also avoids reusing counter values in case the client crashes or the connection to the repository is severed, since any reservation would have been committed to both the security database and the repository before any data is encrypted. Borg uses its standard mechanism (SaveFile) to ensure that reservations are durable (on most hardware / storage systems), therefore a crash of the client's host would not impact tracking of reservations.

However, this design is not infallible, and requires synchronization between clients, which is handled through the repository. Therefore in a multiple-client scenario a repository can trick a client into reusing counter values by ignoring counter reservations and replaying the manifest (which will fail if the client has seen a more recent manifest or has a more recent nonce reservation). If the repository is untrusted, but a trusted synchronization channel exists between clients, the security database could be synchronized between them over said trusted channel. This is not part of Borg's functionality.

Offline key security

Borg cannot secure the key material while it is running, because the keys are needed in plain to decrypt/encrypt repository objects.

For offline storage of the encryption keys they are encrypted with a user-chosen passphrase.

A 256 bit key encryption key (KEK) is derived from the passphrase using PBKDF2-HMAC-SHA256 with a random 256 bit salt which is then used to Encrypt-and-MAC (unlike the Encrypt-then-MAC approach used otherwise) a packed representation of the keys with AES-256-CTR with a constant initialization vector of 0. A HMAC-SHA256 of the plaintext is generated using the same KEK and is stored alongside the ciphertext, which is converted to base64 in its entirety.

This base64 blob (commonly referred to as keyblob) is then stored in the key file or in the repository config (keyfile and repokey modes respectively).

This scheme, and specifically the use of a constant IV with the CTR mode, is secure because an identical passphrase will result in a different derived KEK for every key encryption due to the salt.

The use of Encrypt-and-MAC instead of Encrypt-then-MAC is seen as uncritical (but not ideal) here, since it is combined with AES-CTR mode, which is not vulnerable to padding attacks.

See also

Refer to the Key files section for details on the format.

Refer to issue #747 for suggested improvements of the encryption scheme and password-based key derivation.

Implementations used

We do not implement cryptographic primitives ourselves, but rely on widely used libraries providing them:

  • AES-CTR and HMAC-SHA-256 from OpenSSL 1.0 / 1.1 are used, which is also linked into the static binaries we provide. We think this is not an additional risk, since we don't ever use OpenSSL's networking, TLS or X.509 code, but only their primitives implemented in libcrypto.

  • SHA-256, SHA-512 and BLAKE2b from Python's hashlib standard library module are used. Borg requires a Python built with OpenSSL support (due to PBKDF2), therefore these functions are delegated to OpenSSL by Python.

  • HMAC, PBKDF2 and a constant-time comparison from Python's hmac standard library module is used. While the HMAC implementation is written in Python, the PBKDF2 implementation is provided by OpenSSL. The constant-time comparison (compare_digest) is written in C and part of Python.

Implemented cryptographic constructions are:

  • Encrypt-then-MAC based on AES-256-CTR and either HMAC-SHA-256 or keyed BLAKE2b256 as described above under Encryption.

  • Encrypt-and-MAC based on AES-256-CTR and HMAC-SHA-256 as described above under Offline key security.

  • HKDF-SHA-512

Remote RPC protocol security

Note

This section could be further expanded / detailed.

The RPC protocol is fundamentally based on msgpack'd messages exchanged over an encrypted SSH channel (the system's SSH client is used for this by piping data from/to it).

This means that the authorization and transport security properties are inherited from SSH and the configuration of the SSH client and the SSH server -- Borg RPC does not contain any networking code. Networking is done by the SSH client running in a separate process, Borg only communicates over the standard pipes (stdout, stderr and stdin) with this process. This also means that Borg doesn't have to directly use a SSH client (or SSH at all). For example, sudo or qrexec could be used as an intermediary.

By using the system's SSH client and not implementing a (cryptographic) network protocol Borg sidesteps many security issues that would normally impact distributing statically linked / standalone binaries.

The remainder of this section will focus on the security of the RPC protocol within Borg.

The assumed worst-case a server can inflict to a client is a denial of repository service.

The situation where a server can create a general DoS on the client should be avoided, but might be possible by e.g. forcing the client to allocate large amounts of memory to decode large messages (or messages that merely indicate a large amount of data follows). The RPC protocol code uses a limited msgpack Unpacker to prohibit this.

We believe that other kinds of attacks, especially critical vulnerabilities like remote code execution are inhibited by the design of the protocol:

  1. The server cannot send requests to the client on its own accord, it only can send responses. This avoids "unexpected inversion of control" issues.

  2. msgpack serialization does not allow embedding or referencing code that is automatically executed. Incoming messages are unpacked by the msgpack unpacker into native Python data structures (like tuples and dictionaries), which are then passed to the rest of the program.

    Additional verification of the correct form of the responses could be implemented.

  3. Remote errors are presented in two forms:

    1. A simple plain-text stderr channel. A prefix string indicates the kind of message (e.g. WARNING, INFO, ERROR), which is used to suppress it according to the log level selected in the client.

      A server can send arbitrary log messages, which may confuse a user. However, log messages are only processed when server requests are in progress, therefore the server cannot interfere / confuse with security critical dialogue like the password prompt.

    2. Server-side exceptions passed over the main data channel. These follow the general pattern of server-sent responses and are sent instead of response data for a request.

The msgpack implementation used (msgpack-python) has a good security track record, a large test suite and no issues found by fuzzing. It is based on the msgpack-c implementation, sharing the unpacking engine and some support code. msgpack-c has a good track record as well. Some issues [3] in the past were located in code not included in msgpack-python. Borg does not use msgpack-c.

Using OpenSSL

Borg uses the OpenSSL library for most cryptography (see Implementations used above). OpenSSL is bundled with static releases, thus the bundled copy is not updated with system updates.

OpenSSL is a large and complex piece of software and has had its share of vulnerabilities, however, it is important to note that Borg links against libcrypto not libssl. libcrypto is the low-level cryptography part of OpenSSL, while libssl implements TLS and related protocols.

The latter is not used by Borg (cf. Remote RPC protocol security, Borg itself does not implement any network access) and historically contained most vulnerabilities, especially critical ones. The static binaries released by the project contain neither libssl nor the Python ssl/_ssl modules.

Compression and Encryption

Combining encryption with compression can be insecure in some contexts (e.g. online protocols).

There was some discussion about this in github issue #1040 and for Borg some developers concluded this is no problem at all, some concluded this is hard and extremely slow to exploit and thus no problem in practice.

No matter what, there is always the option not to use compression if you are worried about this.

Fingerprinting
Stored chunk sizes

A borg repository does not hide the size of the chunks it stores (size information is needed to operate the repository).

The chunks stored in the repo are the (compressed, encrypted and authenticated) output of the chunker. The sizes of these stored chunks are influenced by the compression, encryption and authentication.

buzhash chunker

The buzhash chunker chunks according to the input data, the chunker's parameters and the secret chunker seed (which all influence the chunk boundary positions).

Small files below some specific threshold (default: 512 KiB) result in only one chunk (identical content / size as the original file), bigger files result in multiple chunks.

fixed chunker

This chunker yields fixed sized chunks, with optional support of a differently sized header chunk. The last chunk is not required to have the full block size and is determined by the input file size.

Within our attack model, an attacker possessing a specific set of files which he assumes that the victim also possesses (and backups into the repository) could try a brute force fingerprinting attack based on the chunk sizes in the repository to prove his assumption.

To make this more difficult, borg has an obfuscate pseudo compressor, that will take the output of the normal compression step and tries to obfuscate the size of that output. Of course, it can only add to the size, not reduce it. Thus, the optional usage of this mechanism comes at a cost: it will make your repository larger (ranging from a few percent larger [cheap] to ridiculously larger [expensive], depending on the algorithm/params you wisely choose).

The output of the compressed-size obfuscation step will then be encrypted and authenticated, as usual. Of course, using that obfuscation would not make any sense without encryption. Thus, the additional data added by the obfuscator are just 0x00 bytes, which is good enough because after encryption it will look like random anyway.

To summarize, this is making size-based fingerprinting difficult:

  • user-selectable chunker algorithm (and parametrization)

  • for the buzhash chunker: secret, random per-repo chunker seed

  • user-selectable compression algorithm (and level)

  • optional obfuscate pseudo compressor with different choices of algorithm and parameters

Stored chunk proximity

Borg does not try to obfuscate order / proximity of files it discovers by recursing through the filesystem. For performance reasons, we sort directory contents in file inode order (not in file name alphabetical order), so order fingerprinting is not useful for an attacker.

But, when new files are close to each other (when looking at recursion / scanning order), the resulting chunks will be also stored close to each other in the resulting repository segment file(s).

This might leak additional information for the chunk size fingerprinting attack (see above).

Data structures and file formats

This page documents the internal data structures and storage mechanisms of Borg. It is partly based on mailing list discussion about internals and also on static code analysis.

Repository

Borg stores its data in a Repository, which is a file system based transactional key-value store. Thus the repository does not know about the concept of archives or items.

Each repository has the following file structure:

README

simple text file telling that this is a Borg repository

config

repository configuration

data/

directory where the actual data is stored

hints.%d

hints for repository compaction

index.%d

repository index

lock.roster and lock.exclusive/*

used by the locking system to manage shared and exclusive locks

Transactionality is achieved by using a log (aka journal) to record changes. The log is a series of numbered files called segments. Each segment is a series of log entries. The segment number together with the offset of each entry relative to its segment start establishes an ordering of the log entries. This is the "definition" of time for the purposes of the log.

Config file

Each repository has a config file which is a INI-style file and looks like this:

[repository]
version = 1
segments_per_dir = 1000
max_segment_size = 524288000
id = 57d6c1d52ce76a836b532b0e42e677dec6af9fca3673db511279358828a21ed6

This is where the repository.id is stored. It is a unique identifier for repositories. It will not change if you move the repository around so you can make a local transfer then decide to move the repository to another (even remote) location at a later time.

Keys

Repository keys are byte-strings of fixed length (32 bytes), they don't have a particular meaning (except for the Manifest).

Normally the keys are computed like this:

key = id = id_hash(unencrypted_data)

The id_hash function depends on the encryption mode.

As the id / key is used for deduplication, id_hash must be a cryptographically strong hash or MAC.

Segments

Objects referenced by a key are stored inline in files (segments) of approx. 500 MB size in numbered subdirectories of repo/data. The number of segments per directory is controlled by the value of segments_per_dir. If you change this value in a non-empty repository, you may also need to relocate the segment files manually.

A segment starts with a magic number (BORG_SEG as an eight byte ASCII string), followed by a number of log entries. Each log entry consists of: (in this order)

  • First, unsigned 32-bit number, the CRC32 of the entire entry (for a PUT including the DATA) excluding the CRC32 field

  • Second, unsigned 32-bit size of the entry (including the whole header)

  • Third, unsigned 8-bit entry tag: PUT(0), DELETE(1) or COMMIT(2)

  • Fourth, on PUT or DELETE, 32 byte key

  • Fifth, PUT only, (size - 41) bytes of data (length = size - sizeof(CRC32) - sizeof(size) - sizeof(entry tag) - sizeof(key))

Those files are strictly append-only and modified only once.

Tag is either PUT, DELETE, or COMMIT.

When an object is written to the repository a PUT entry is written to the file containing the object id and data. If an object is deleted a DELETE entry is appended with the object id.

A COMMIT tag is written when a repository transaction is committed. The segment number of the segment containing a commit is the transaction ID.

When a repository is opened any PUT or DELETE operations not followed by a COMMIT tag are discarded since they are part of a partial/uncommitted transaction.

The size of individual segments is limited to 4 GiB, since the offset of entries within segments is stored in a 32-bit unsigned integer in the repository index.

Objects

All objects (the manifest, archives, archive item streams chunks and file data chunks) are encrypted and/or compressed. See Encryption for a graphic outlining the anatomy of an object in Borg. The type for compression is explained in Compression.

Index, hints and integrity

The repository index is stored in index.<TRANSACTION_ID> and is used to determine an object's location in the repository. It is a HashIndex, a hash table using open addressing. It maps object keys to two unsigned 32-bit integers; the first integer gives the segment number, the second indicates the offset of the object's entry within the segment.

The hints file is a msgpacked file named hints.<TRANSACTION_ID>. It contains:

  • version

  • list of segments

  • compact

The integrity file is a msgpacked file named integrity.<TRANSACTION_ID>. It contains checksums of the index and hints files and is described in the Checksumming data structures section below.

If the index or hints are corrupted, they are re-generated automatically. If they are outdated, segments are replayed from the index state to the currently committed transaction.

Compaction

For a given key only the last entry regarding the key, which is called current (all other entries are called superseded), is relevant: If there is no entry or the last entry is a DELETE then the key does not exist. Otherwise the last PUT defines the value of the key.

By superseding a PUT (with either another PUT or a DELETE) the log entry becomes obsolete. A segment containing such obsolete entries is called sparse, while a segment containing no such entries is called compact.

Since writing a DELETE tag does not actually delete any data and thus does not free disk space any log-based data store will need a compaction strategy (somewhat analogous to a garbage collector).

Borg uses a simple forward compacting algorithm, which avoids modifying existing segments. Compaction runs when a commit is issued with compact=True parameter, e.g. by the borg compact command (unless the Append-only mode (forbid compaction) is active). One client transaction can manifest as multiple physical transactions, since compaction is transacted, too, and Borg does not distinguish between the two:

Perspective| Time -->
-----------+--------------
Client     | Begin transaction - Modify Data - Commit | <client waits for repository> (done)
Repository | Begin transaction - Modify Data - Commit | Compact segments - Commit   | (done)

The compaction algorithm requires two inputs in addition to the segments themselves:

  1. Which segments are sparse, to avoid scanning all segments (impractical). Further, Borg uses a conditional compaction strategy: Only those segments that exceed a threshold sparsity are compacted.

    To implement the threshold condition efficiently, the sparsity has to be stored as well. Therefore, Borg stores a mapping (segment id,) -> (number of sparse bytes,).

    The 1.0.x series used a simpler non-conditional algorithm, which only required the list of sparse segments. Thus, it only stored a list, not the mapping described above.

  2. Each segment's reference count, which indicates how many live objects are in a segment. This is not strictly required to perform the algorithm. Rather, it is used to validate that a segment is unused before deleting it. If the algorithm is incorrect, or the reference count was not accounted correctly, then an assertion failure occurs.

These two pieces of information are stored in the hints file (hints.N) next to the index (index.N).

When loading a hints file, Borg checks the version contained in the file. The 1.0.x series writes version 1 of the format (with the segments list instead of the mapping, mentioned above). Since Borg 1.0.4, version 2 is read as well. The 1.1.x series writes version 2 of the format and reads either version. When reading a version 1 hints file, Borg 1.1.x will read all sparse segments to determine their sparsity.

This process may take some time if a repository has been kept in append-only mode or borg compact has not been used for a longer time, which both has caused the number of sparse segments to grow.

Compaction processes sparse segments from oldest to newest; sparse segments which don't contain enough deleted data to justify compaction are skipped. This avoids doing e.g. 500 MB of writing current data to a new segment when only a couple kB were deleted in a segment.

Segments that are compacted are read in entirety. Current entries are written to a new segment, while superseded entries are omitted. After each segment an intermediary commit is written to the new segment. Then, the old segment is deleted (asserting that the reference count diminished to zero), freeing disk space.

A simplified example (excluding conditional compaction and with simpler commit logic) showing the principal operation of compaction:

_images/compaction.png

(The actual algorithm is more complex to avoid various consistency issues, refer to the borg.repository module for more comments and documentation on these issues.)

Storage quotas

Quotas are implemented at the Repository level. The active quota of a repository is determined by the storage_quota config entry or a run-time override (via borg serve). The currently used quota is stored in the hints file. Operations (PUT and DELETE) during a transaction modify the currently used quota:

  • A PUT adds the size of the log entry to the quota, i.e. the length of the data plus the 41 byte header.

  • A DELETE subtracts the size of the deleted log entry from the quota, which includes the header.

Thus, PUT and DELETE are symmetric and cancel each other out precisely.

The quota does not track on-disk size overheads (due to conditional compaction or append-only mode). In normal operation the inclusion of the log entry headers in the quota act as a faithful proxy for index and hints overheads.

By tracking effective content size, the client can always recover from a full quota by deleting archives. This would not be possible if the quota tracked on-disk size, since journaling DELETEs requires extra disk space before space is freed. Tracking effective size on the other hand accounts DELETEs immediately as freeing quota.

Enforcing the quota

The storage quota is meant as a robust mechanism for service providers, therefore borg serve has to enforce it without loopholes (e.g. modified clients). The following sections refer to using quotas on remotely accessed repositories. For local access, consider client and serve the same. Accordingly, quotas cannot be enforced with local access, since the quota can be changed in the repository config.

The quota is enforcible only if all borg serve versions accessible to clients support quotas (see next section). Further, quota is per repository. Therefore, ensure clients can only access a defined set of repositories with their quotas set, using --restrict-to-repository.

If the client exceeds the storage quota the StorageQuotaExceeded exception is raised. Normally a client could ignore such an exception and just send a commit() command anyway, circumventing the quota. However, when StorageQuotaExceeded is raised, it is stored in the transaction_doomed attribute of the repository. If the transaction is doomed, then commit will re-raise this exception, aborting the commit.

The transaction_doomed indicator is reset on a rollback (which erases the quota-exceeding state).

Compatibility with older servers and enabling quota after-the-fact

If no quota data is stored in the hints file, Borg assumes zero quota is used. Thus, if a repository with an enabled quota is written to with an older borg serve version that does not understand quotas, then the quota usage will be erased.

The client version is irrelevant to the storage quota and has no part in it. The form of error messages due to exceeding quota varies with client versions.

A similar situation arises when upgrading from a Borg release that did not have quotas. Borg will start tracking quota use from the time of the upgrade, starting at zero.

If the quota shall be enforced accurately in these cases, either

  • delete the index.N and hints.N files, forcing Borg to rebuild both, re-acquiring quota data in the process, or

  • edit the msgpacked hints.N file (not recommended and thus not documented further).

The object graph

On top of the simple key-value store offered by the Repository, Borg builds a much more sophisticated data structure that is essentially a completely encrypted object graph. Objects, such as archives, are referenced by their chunk ID, which is cryptographically derived from their contents. More on how this helps security in Structural Authentication.

_images/object-graph.png
The manifest

The manifest is the root of the object hierarchy. It references all archives in a repository, and thus all data in it. Since no object references it, it cannot be stored under its ID key. Instead, the manifest has a fixed all-zero key.

The manifest is rewritten each time an archive is created, deleted, or modified. It looks like this:

{
    b'version': 1,
    b'timestamp': b'2017-05-05T12:42:23.042864',
    b'item_keys': [b'acl_access', b'acl_default', ...],
    b'config': {},
    b'archives': {
        b'2017-05-05-system-backup': {
            b'id': b'<32 byte binary object ID>',
            b'time': b'2017-05-05T12:42:22.942864',
        },
    },
    b'tam': ...,
}

The version field can be either 1 or 2. The versions differ in the way feature flags are handled, described below.

The timestamp field is used to avoid logical replay attacks where the server just resets the repository to a previous state.

item_keys is a list containing all Item keys that may be encountered in the repository. It is used by borg check, which verifies that all keys in all items are a subset of these keys. Thus, an older version of borg check supporting this mechanism can correctly detect keys introduced in later versions.

The tam key is part of the tertiary authentication mechanism (formerly known as "tertiary authentication for metadata") and authenticates the manifest, since an ID check is not possible.

config is a general-purpose location for additional metadata. All versions of Borg preserve its contents (it may have been a better place for item_keys, which is not preserved by unaware Borg versions, releases predating 1.0.4).

Feature flags

Feature flags are used to add features to data structures without causing corruption if older versions are used to access or modify them. The main issues to consider for a feature flag oriented design are flag granularity, flag storage, and cache invalidation.

Feature flags are divided in approximately three categories, detailed below. Due to the nature of ID-based deduplication, write (i.e. creating archives) and read access are not symmetric; it is possible to create archives referencing chunks that are not readable with the current feature set. The third category are operations that require accurate reference counts, for example archive deletion and check.

As the manifest is always updated and always read, it is the ideal place to store feature flags, comparable to the super-block of a file system. The only problem is to recover from a lost manifest, i.e. how is it possible to detect which feature flags are enabled, if there is no manifest to tell. This issue is left open at this time, but is not expected to be a major hurdle; it doesn't have to be handled efficiently, it just needs to be handled.

Lastly, cache invalidation is handled by noting which feature flags were and which were not understood while manipulating a cache. This allows borg to detect whether the cache needs to be invalidated, i.e. rebuilt from scratch. See Cache feature flags below.

The config key stores the feature flags enabled on a repository:

config = {
    b'feature_flags': {
        b'read': {
            b'mandatory': [b'some_feature'],
        },
        b'check': {
            b'mandatory': [b'other_feature'],
        }
        b'write': ...,
        b'delete': ...
    },
}

The top-level distinction for feature flags is the operation the client intends to perform,

the read operation includes extraction and listing of archives,
the write operation includes creating new archives,
the delete (archives) operation,
the check operation requires full understanding of everything in the repository.

These are weakly set-ordered; check will include everything required for delete, delete will likely include write and read. However, read may require more features than write (due to ID-based deduplication, write does not necessarily require reading/understanding repository contents).

Each operation can contain several sets of feature flags. Only one set, the mandatory set is currently defined.

Upon reading the manifest, the Borg client has already determined which operation should be performed. If feature flags are found in the manifest, the set of feature flags supported by the client is compared to the mandatory set found in the manifest. If any unsupported flags are found (i.e. the mandatory set is not a subset of the features supported by the Borg client used), the operation is aborted with a MandatoryFeatureUnsupported error:

Unsupported repository feature(s) {'some_feature'}. A newer version of borg is required to access this repository.

Older Borg releases do not have this concept and do not perform feature flags checks. These can be locked out with manifest version 2. Thus, the only difference between manifest versions 1 and 2 is that the latter is only accepted by Borg releases implementing feature flags.

Therefore, as soon as any mandatory feature flag is enabled in a repository, the manifest version must be switched to version 2 in order to lock out all Borg releases unaware of feature flags.

Cache feature flags

The cache does not have its separate set of feature flags. Instead, Borg stores which flags were used to create or modify a cache.

All mandatory manifest features from all operations are gathered in one set. Then, two sets of features are computed;

  • those features that are supported by the client and mandated by the manifest are added to the mandatory_features set,

  • the ignored_features set comprised of those features mandated by the manifest, but not supported by the client.

Because the client previously checked compliance with the mandatory set of features required for the particular operation it is executing, the mandatory_features set will contain all necessary features required for using the cache safely.

Conversely, the ignored_features set contains only those features which were not relevant to operating the cache. Otherwise, the client would not pass the feature set test against the manifest.

When opening a cache and the mandatory_features set is not a subset of the features supported by the client, the cache is wiped out and rebuilt, since a client not supporting a mandatory feature that the cache was built with would be unable to update it correctly. The assumption behind this behaviour is that any of the unsupported features could have been reflected in the cache and there is no way for the client to discern whether that is the case. Meanwhile, it may not be practical for every feature to have clients using it track whether the feature had an impact on the cache. Therefore, the cache is wiped.

When opening a cache and the intersection of ignored_features and the features supported by the client contains any elements, i.e. the client possesses features that the previous client did not have and those new features are enabled in the repository, the cache is wiped out and rebuilt.

While the former condition likely requires no tweaks, the latter condition is formulated in an especially conservative way to play it safe. It seems likely that specific features might be exempted from the latter condition.

Defined feature flags

Currently no feature flags are defined.

From currently planned features, some examples follow, these may/may not be implemented and purely serve as examples.

  • A mandatory read feature could be using a different encryption scheme (e.g. session keys). This may not be mandatory for the write operation - reading data is not strictly required for creating an archive.

  • Any additions to the way chunks are referenced (e.g. to support larger archives) would become a mandatory delete and check feature; delete implies knowing correct reference counts, so all object references need to be understood. check must discover the entire object graph as well, otherwise the "orphan chunks check" could delete data still in use.

Archives

Each archive is an object referenced by the manifest. The archive object itself does not store any of the data contained in the archive it describes.

Instead, it contains a list of chunks which form a msgpacked stream of items. The archive object itself further contains some metadata:

  • version

  • name, which might differ from the name set in the manifest. When borg check rebuilds the manifest (e.g. if it was corrupted) and finds more than one archive object with the same name, it adds a counter to the name in the manifest, but leaves the name field of the archives as it was.

  • items, a list of chunk IDs containing item metadata (size: count * ~34B)

  • cmdline, the command line which was used to create the archive

  • hostname

  • username

  • time and time_end are the start and end timestamps, respectively

  • comment, a user-specified archive comment

  • chunker_params are the chunker-params used for creating the archive. This is used by borg recreate to determine whether a given archive needs rechunking.

  • Some other pieces of information related to recreate.

Note about archive limitations

The archive is currently stored as a single object in the repository and thus limited in size to MAX_OBJECT_SIZE (20MiB).

As one chunk list entry is ~40B, that means we can reference ~500.000 item metadata stream chunks per archive.

Each item metadata stream chunk is ~128kiB (see hardcoded ITEMS_CHUNKER_PARAMS).

So that means the whole item metadata stream is limited to ~64GiB chunks. If compression is used, the amount of storable metadata is bigger - by the compression factor.

If the medium size of an item entry is 100B (small size file, no ACLs/xattrs), that means a limit of ~640 million files/directories per archive.

If the medium size of an item entry is 2kB (~100MB size files or more ACLs/xattrs), the limit will be ~32 million files/directories per archive.

If one tries to create an archive object bigger than MAX_OBJECT_SIZE, a fatal IntegrityError will be raised.

A workaround is to create multiple archives with fewer items each, see also #1452.

Items

Each item represents a file, directory or other file system item and is stored as a dictionary created by the Item class that contains:

  • path

  • list of data chunks (size: count * ~40B)

  • user

  • group

  • uid

  • gid

  • mode (item type + permissions)

  • source (for symlinks, and for hardlinks within one archive)

  • rdev (for device files)

  • mtime, atime, ctime in nanoseconds

  • xattrs

  • acl (various OS-dependent fields)

  • flags

All items are serialized using msgpack and the resulting byte stream is fed into the same chunker algorithm as used for regular file data and turned into deduplicated chunks. The reference to these chunks is then added to the archive metadata. To achieve a finer granularity on this metadata stream, we use different chunker params for this chunker, which result in smaller chunks.

A chunk is stored as an object as well, of course.

Chunks

Borg has these chunkers:

  • "fixed": a simple, low cpu overhead, fixed blocksize chunker, optionally supporting a header block of different size.

  • "buzhash": variable, content-defined blocksize, uses a rolling hash computed by the Buzhash algorithm.

For some more general usage hints see also --chunker-params.

"fixed" chunker

The fixed chunker triggers (chunks) at even-spaced offsets, e.g. every 4MiB, producing chunks of same block size (the last chunk is not required to be full-size).

Optionally, it supports processing a differently sized "header" first, before it starts to cut chunks of the desired block size. The default is not to have a differently sized header.

borg create --chunker-params fixed,BLOCK_SIZE[,HEADER_SIZE]

  • BLOCK_SIZE: no default value, multiple of the system page size (usually 4096 bytes) recommended. E.g.: 4194304 would cut 4MiB sized chunks.

  • HEADER_SIZE: optional, defaults to 0 (no header).

The fixed chunker also supports processing sparse files (reading only the ranges with data and seeking over the empty hole ranges).

borg create --sparse --chunker-params fixed,BLOCK_SIZE[,HEADER_SIZE]

"buzhash" chunker

The buzhash chunker triggers (chunks) when the last HASH_MASK_BITS bits of the hash are zero, producing chunks with a target size of 2^HASH_MASK_BITS bytes.

Buzhash is only used for cutting the chunks at places defined by the content, the buzhash value is not used as the deduplication criteria (we use a cryptographically strong hash/MAC over the chunk contents for this, the id_hash).

The idea of content-defined chunking is assigning every byte where a cut could be placed a hash. The hash is based on some number of bytes (the window size) before the byte in question. Chunks are cut where the hash satisfies some condition (usually "n numbers of trailing/leading zeroes"). This causes chunks to be cut in the same location relative to the file's contents, even if bytes are inserted or removed before/after a cut, as long as the bytes within the window stay the same. This results in a high chance that a single cluster of changes to a file will only result in 1-2 new chunks, aiding deduplication.

Using normal hash functions this would be extremely slow, requiring hashing approximately window size * file size bytes. A rolling hash is used instead, which allows to add a new input byte and compute a new hash as well as remove a previously added input byte from the computed hash. This makes the cost of computing a hash for each input byte largely independent of the window size.

Borg defines minimum and maximum chunk sizes (CHUNK_MIN_EXP and CHUNK_MAX_EXP, respectively) which narrows down where cuts may be made, greatly reducing the amount of data that is actually hashed for content-defined chunking.

borg create --chunker-params buzhash,CHUNK_MIN_EXP,CHUNK_MAX_EXP,HASH_MASK_BITS,HASH_WINDOW_SIZE can be used to tune the chunker parameters, the default is:

  • CHUNK_MIN_EXP = 19 (minimum chunk size = 2^19 B = 512 kiB)

  • CHUNK_MAX_EXP = 23 (maximum chunk size = 2^23 B = 8 MiB)

  • HASH_MASK_BITS = 21 (target chunk size ~= 2^21 B = 2 MiB)

  • HASH_WINDOW_SIZE = 4095 [B] (0xFFF)

The buzhash table is altered by XORing it with a seed randomly generated once for the repository, and stored encrypted in the keyfile. This is to prevent chunk size based fingerprinting attacks on your encrypted repo contents (to guess what files you have based on a specific set of chunk sizes).

The cache

The files cache is stored in cache/files and is used at backup time to quickly determine whether a given file is unchanged and we have all its chunks.

In memory, the files cache is a key -> value mapping (a Python dict) and contains:

  • key: id_hash of the encoded, absolute file path

  • value:

    • file inode number

    • file size

    • file mtime_ns

    • age (0 [newest], 1, 2, 3, ..., BORG_FILES_CACHE_TTL - 1)

    • list of chunk ids representing the file's contents

To determine whether a file has not changed, cached values are looked up via the key in the mapping and compared to the current file attribute values.

If the file's size, mtime_ns and inode number is still the same, it is considered to not have changed. In that case, we check that all file content chunks are (still) present in the repository (we check that via the chunks cache).

If everything is matching and all chunks are present, the file is not read / chunked / hashed again (but still a file metadata item is written to the archive, made from fresh file metadata read from the filesystem). This is what makes borg so fast when processing unchanged files.

If there is a mismatch or a chunk is missing, the file is read / chunked / hashed. Chunks already present in repo won't be transferred to repo again.

The inode number is stored and compared to make sure we distinguish between different files, as a single path may not be unique across different archives in different setups.

Not all filesystems have stable inode numbers. If that is the case, borg can be told to ignore the inode number in the check via --ignore-inode.

The age value is used for cache management. If a file is "seen" in a backup run, its age is reset to 0, otherwise its age is incremented by one. If a file was not seen in BORG_FILES_CACHE_TTL backups, its cache entry is removed. See also: It always chunks all my files, even unchanged ones! and I am seeing 'A' (added) status for an unchanged file!?

The files cache is a python dictionary, storing python objects, which generates a lot of overhead.

Borg can also work without using the files cache (saves memory if you have a lot of files or not much RAM free), then all files are assumed to have changed. This is usually much slower than with files cache.

The on-disk format of the files cache is a stream of msgpacked tuples (key, value). Loading the files cache involves reading the file, one msgpack object at a time, unpacking it, and msgpacking the value (in an effort to save memory).

The chunks cache is stored in cache/chunks and is used to determine whether we already have a specific chunk, to count references to it and also for statistics.

The chunks cache is a key -> value mapping and contains:

  • key:

    • chunk id_hash

  • value:

    • reference count

    • size

    • encrypted/compressed size

The chunks cache is a HashIndex. Due to some restrictions of HashIndex, the reference count of each given chunk is limited to a constant, MAX_VALUE (introduced below in HashIndex), approximately 2**32. If a reference count hits MAX_VALUE, decrementing it yields MAX_VALUE again, i.e. the reference count is pinned to MAX_VALUE.

Indexes / Caches memory usage

Here is the estimated memory usage of Borg - it's complicated:

chunk_size ~= 2 ^ HASH_MASK_BITS  (for buzhash chunker, BLOCK_SIZE for fixed chunker)
chunk_count ~= total_file_size / chunk_size

repo_index_usage = chunk_count * 40

chunks_cache_usage = chunk_count * 44

files_cache_usage = total_file_count * 240 + chunk_count * 80

mem_usage ~= repo_index_usage + chunks_cache_usage + files_cache_usage
           = chunk_count * 164 + total_file_count * 240

Due to the hashtables, the best/usual/worst cases for memory allocation can be estimated like that:

mem_allocation = mem_usage / load_factor  # l_f = 0.25 .. 0.75

mem_allocation_peak = mem_allocation * (1 + growth_factor)  # g_f = 1.1 .. 2

All units are Bytes.

It is assuming every chunk is referenced exactly once (if you have a lot of duplicate chunks, you will have fewer chunks than estimated above).

It is also assuming that typical chunk size is 2^HASH_MASK_BITS (if you have a lot of files smaller than this statistical medium chunk size, you will have more chunks than estimated above, because 1 file is at least 1 chunk).

If a remote repository is used the repo index will be allocated on the remote side.

The chunks cache, files cache and the repo index are all implemented as hash tables. A hash table must have a significant amount of unused entries to be fast - the so-called load factor gives the used/unused elements ratio.

When a hash table gets full (load factor getting too high), it needs to be grown (allocate new, bigger hash table, copy all elements over to it, free old hash table) - this will lead to short-time peaks in memory usage each time this happens. Usually does not happen for all hashtables at the same time, though. For small hash tables, we start with a growth factor of 2, which comes down to ~1.1x for big hash tables.

E.g. backing up a total count of 1 Mi (IEC binary prefix i.e. 2^20) files with a total size of 1TiB.

  1. with create --chunker-params buzhash,10,23,16,4095 (custom, like borg < 1.0 or attic):

mem_usage = 2.8GiB

  1. with create --chunker-params buzhash,19,23,21,4095 (default):

mem_usage = 0.31GiB

Note

There is also the --files-cache=disabled option to disable the files cache. You'll save some memory, but it will need to read / chunk all the files as it can not skip unmodified files then.

HashIndex

The chunks cache and the repository index are stored as hash tables, with only one slot per bucket, spreading hash collisions to the following buckets. As a consequence the hash is just a start position for a linear search. If a key is looked up that is not in the table, then the hash table is searched from the start position (the hash) until the first empty bucket is reached.

This particular mode of operation is open addressing with linear probing.

When the hash table is filled to 75%, its size is grown. When it's emptied to 25%, its size is shrinked. Operations on it have a variable complexity between constant and linear with low factor, and memory overhead varies between 33% and 300%.

If an element is deleted, and the slot behind the deleted element is not empty, then the element will leave a tombstone, a bucket marked as deleted. Tombstones are only removed by insertions using the tombstone's bucket, or by resizing the table. They present the same load to the hash table as a real entry, but do not count towards the regular load factor.

Thus, if the number of empty slots becomes too low (recall that linear probing for an element not in the index stops at the first empty slot), the hash table is rebuilt. The maximum effective load factor, i.e. including tombstones, is 93%.

Data in a HashIndex is always stored in little-endian format, which increases efficiency for almost everyone, since basically no one uses big-endian processors any more.

HashIndex does not use a hashing function, because all keys (save manifest) are outputs of a cryptographic hash or MAC and thus already have excellent distribution. Thus, HashIndex simply uses the first 32 bits of the key as its "hash".

The format is easy to read and write, because the buckets array has the same layout in memory and on disk. Only the header formats differ. The on-disk header is struct HashHeader:

  • First, the HashIndex magic, the eight byte ASCII string "BORG_IDX".

  • Second, the signed 32-bit number of entries (i.e. buckets which are not deleted and not empty).

  • Third, the signed 32-bit number of buckets, i.e. the length of the buckets array contained in the file, and the modulus for index calculation.

  • Fourth, the signed 8-bit length of keys.

  • Fifth, the signed 8-bit length of values. This has to be at least four bytes.

All fields are packed.

The HashIndex is not a general purpose data structure. The value size must be at least 4 bytes, and these first bytes are used for in-band signalling in the data structure itself.

The constant MAX_VALUE (defined as 2**32-1025 = 4294966271) defines the valid range for these 4 bytes when interpreted as an uint32_t from 0 to MAX_VALUE (inclusive). The following reserved values beyond MAX_VALUE are currently in use (byte order is LE):

  • 0xffffffff marks empty buckets in the hash table

  • 0xfffffffe marks deleted buckets in the hash table

HashIndex is implemented in C and wrapped with Cython in a class-based interface. The Cython wrapper checks every passed value against these reserved values and raises an AssertionError if they are used.

Encryption

See also

The Cryptography in Borg section for an in-depth review.

AES-256 is used in CTR mode (so no need for padding). A 64 bit initialization vector is used, a MAC is computed on the encrypted chunk and both are stored in the chunk. Encryption and MAC use two different keys. Each chunk consists of TYPE(1) + MAC(32) + NONCE(8) + CIPHERTEXT:

_images/encryption.png

In AES-CTR mode you can think of the IV as the start value for the counter. The counter itself is incremented by one after each 16 byte block. The IV/counter is not required to be random but it must NEVER be reused. So to accomplish this Borg initializes the encryption counter to be higher than any previously used counter value before encrypting new data.

To reduce payload size, only 8 bytes of the 16 bytes nonce is saved in the payload, the first 8 bytes are always zeros. This does not affect security but limits the maximum repository capacity to only 295 exabytes (2**64 * 16 bytes).

Encryption keys (and other secrets) are kept either in a key file on the client ('keyfile' mode) or in the repository config on the server ('repokey' mode). In both cases, the secrets are generated from random and then encrypted by a key derived from your passphrase (this happens on the client before the key is stored into the keyfile or as repokey).

The passphrase is passed through the BORG_PASSPHRASE environment variable or prompted for interactive usage.

Key files

See also

The Offline key security section for an in-depth review of the key encryption.

When initialized with the init -e keyfile command, Borg needs an associated file in $HOME/.config/borg/keys to read and write the repository. The format is based on msgpack, base64 encoding and PBKDF2 SHA256 hashing, which is then encoded again in a msgpack.

The same data structure is also used in the "repokey" modes, which store it in the repository in the configuration file.

The internal data structure is as follows:

version

currently always an integer, 1

repository_id

the id field in the config INI file of the repository.

enc_key

the key used to encrypt data with AES (256 bits)

enc_hmac_key

the key used to HMAC the encrypted data (256 bits)

id_key

the key used to HMAC the plaintext chunk data to compute the chunk's id

chunk_seed

the seed for the buzhash chunking table (signed 32 bit integer)

These fields are packed using msgpack. The utf-8 encoded passphrase is processed with PBKDF2 (SHA256, 100000 iterations, random 256 bit salt) to derive a 256 bit key encryption key (KEK).

A HMAC-SHA256 checksum of the packed fields is generated with the KEK, then the KEK is also used to encrypt the same packed fields using AES-CTR.

The result is stored in a another msgpack formatted as follows:

version

currently always an integer, 1

salt

random 256 bits salt used to process the passphrase

iterations

number of iterations used to process the passphrase (currently 100000)

algorithm

the hashing algorithm used to process the passphrase and do the HMAC checksum (currently the string sha256)

hash

HMAC-SHA256 of the plaintext of the packed fields.

data

The encrypted, packed fields.

The resulting msgpack is then encoded using base64 and written to the key file, wrapped using the standard textwrap module with a header. The header is a single line with a MAGIC string, a space and a hexadecimal representation of the repository id.

Compression

Borg supports the following compression methods, each identified by two bytes:

  • none (no compression, pass through data 1:1), identified by \x00\x00

  • lz4 (low compression, but super fast), identified by \x01\x00

  • zstd (level 1-22 offering a wide range: level 1 is lower compression and high speed, level 22 is higher compression and lower speed) - since borg 1.1.4, identified by \x03\x00

  • zlib (level 0-9, level 0 is no compression [but still adding zlib overhead], level 1 is low, level 9 is high compression), identified by a zlib header (\x.8\x..)

  • lzma (level 0-9, level 0 is low, level 9 is high compression), identified by \x02\x00.

Speed: none > lz4 > zlib > lzma, lz4 > zstd Compression: lzma > zlib > lz4 > none, zstd > lz4

Be careful, higher compression levels might use a lot of resources (CPU/memory).

The overall speed of course also depends on the speed of your target storage. If that is slow, using a higher compression level might yield better overall performance. You need to experiment a bit. Maybe just watch your CPU load, if that is relatively low, increase compression until 1 core is 70-100% loaded.

Even if your target storage is rather fast, you might see interesting effects: while doing no compression at all (none) is a operation that takes no time, it likely will need to store more data to the storage compared to using lz4. The time needed to transfer and store the additional data might be much more than if you had used lz4 (which is super fast, but still might compress your data about 2:1). This is assuming your data is compressible (if you backup already compressed data, trying to compress them at backup time is usually pointless).

Compression is applied after deduplication, thus using different compression methods in one repo does not influence deduplication.

See borg create --help about how to specify the compression level and its default.

Lock files

Borg uses locks to get (exclusive or shared) access to the cache and the repository.

The locking system is based on renaming a temporary directory to lock.exclusive (for exclusive locks). Inside this directory, there is a file indicating hostname, process id and thread id of the lock holder.

There is also a json file lock.roster that keeps a directory of all shared and exclusive lockers.

If the process is able to rename a temporary directory (with the host/process/thread identifier prepared inside it) in the resource directory to lock.exclusive, it has the lock for it. If renaming fails (because this directory already exists and its host/process/thread identifier denotes a thread on the host which is still alive), lock acquisition fails.

The cache lock is usually in ~/.cache/borg/REPOID/lock.*. The repository lock is in repository/lock.*.

In case you run into troubles with the locks, you can use the borg break-lock command after you first have made sure that no Borg process is running on any machine that accesses this resource. Be very careful, the cache or repository might get damaged if multiple processes use it at the same time.

Checksumming data structures

As detailed in the previous sections, Borg generates and stores various files containing important meta data, such as the repository index, repository hints, chunks caches and files cache.

Data corruption in these files can damage the archive data in a repository, e.g. due to wrong reference counts in the chunks cache. Only some parts of Borg were designed to handle corrupted data structures, so a corrupted files cache may cause crashes or write incorrect archives.

Therefore, Borg calculates checksums when writing these files and tests checksums when reading them. Checksums are generally 64-bit XXH64 hashes. The canonical xxHash representation is used, i.e. big-endian. Checksums are stored as hexadecimal ASCII strings.

For compatibility, checksums are not required and absent checksums do not trigger errors. The mechanisms have been designed to avoid false-positives when various Borg versions are used alternately on the same repositories.

Checksums are a data safety mechanism. They are not a security mechanism.

Choice of algorithm

XXH64 has been chosen for its high speed on all platforms, which avoids performance degradation in CPU-limited parts (e.g. cache synchronization). Unlike CRC32, it neither requires hardware support (crc32c or CLMUL) nor vectorized code nor large, cache-unfriendly lookup tables to achieve good performance. This simplifies deployment of it considerably (cf. src/borg/algorithms/crc32...).

Further, XXH64 is a non-linear hash function and thus has a "more or less" good chance to detect larger burst errors, unlike linear CRCs where the probability of detection decreases with error size.

The 64-bit checksum length is considered sufficient for the file sizes typically checksummed (individual files up to a few GB, usually less). xxHash was expressly designed for data blocks of these sizes.

Lower layer — file_integrity

To accommodate the different transaction models used for the cache and repository, there is a lower layer (borg.crypto.file_integrity.IntegrityCheckedFile) wrapping a file-like object, performing streaming calculation and comparison of checksums. Checksum errors are signalled by raising an exception (borg.crypto.file_integrity.FileIntegrityError) at the earliest possible moment.

Calculating checksums

Before feeding the checksum algorithm any data, the file name (i.e. without any path) is mixed into the checksum, since the name encodes the context of the data for Borg.

The various indices used by Borg have separate header and main data parts. IntegrityCheckedFile allows borg to checksum them independently, which avoids even reading the data when the header is corrupted. When a part is signalled, the length of the part name is mixed into the checksum state first (encoded as an ASCII string via %10d printf format), then the name of the part is mixed in as an UTF-8 string. Lastly, the current position (length) in the file is mixed in as well.

The checksum state is not reset at part boundaries.

A final checksum is always calculated in the same way as the parts described above, after seeking to the end of the file. The final checksum cannot prevent code from processing corrupted data during reading, however, it prevents use of the corrupted data.

Serializing checksums

All checksums are compiled into a simple JSON structure called integrity data:

{
    "algorithm": "XXH64",
    "digests": {
        "HashHeader": "eab6802590ba39e3",
        "final": "e2a7f132fc2e8b24"
    }
}

The algorithm key notes the used algorithm. When reading, integrity data containing an unknown algorithm is not inspected further.

The digests key contains a mapping of part names to their digests.

Integrity data is generally stored by the upper layers, introduced below. An exception is the DetachedIntegrityCheckedFile, which automatically writes and reads it from a ".integrity" file next to the data file. It is used for archive chunks indexes in chunks.archive.d.

Upper layer

Storage of integrity data depends on the component using it, since they have different transaction mechanisms, and integrity data needs to be transacted with the data it is supposed to protect.

Main cache files: chunks and files cache

The integrity data of the chunks and files caches is stored in the cache config, since all three are transacted together.

The [integrity] section is used:

[cache]
version = 1
repository = 3c4...e59
manifest = 10e...21c
timestamp = 2017-06-01T21:31:39.699514
key_type = 2
previous_location = /path/to/repo

[integrity]
manifest = 10e...21c
chunks = {"algorithm": "XXH64", "digests": {"HashHeader": "eab...39e3", "final": "e2a...b24"}}

The manifest ID is duplicated in the integrity section due to the way all Borg versions handle the config file. Instead of creating a "new" config file from an internal representation containing only the data understood by Borg, the config file is read in entirety (using the Python ConfigParser) and modified. This preserves all sections and values not understood by the Borg version modifying it.

Thus, if an older versions uses a cache with integrity data, it would preserve the integrity section and its contents. If a integrity-aware Borg version would read this cache, it would incorrectly report checksum errors, since the older version did not update the checksums.

However, by duplicating the manifest ID in the integrity section, it is easy to tell whether the checksums concern the current state of the cache.

Integrity errors are fatal in these files, terminating the program, and are not automatically corrected at this time.

chunks.archive.d

Indices in chunks.archive.d are not transacted and use DetachedIntegrityCheckedFile, which writes the integrity data to a separate ".integrity" file.

Integrity errors result in deleting the affected index and rebuilding it. This logs a warning and increases the exit code to WARNING (1).

Repository index and hints

The repository associates index and hints files with a transaction by including the transaction ID in the file names. Integrity data is stored in a third file ("integrity.<TRANSACTION_ID>"). Like the hints file, it is msgpacked:

{
    b'version': 2,
    b'hints': b'{"algorithm": "XXH64", "digests": {"final": "411208db2aa13f1a"}}',
    b'index': b'{"algorithm": "XXH64", "digests": {"HashHeader": "846b7315f91b8e48", "final": "cb3e26cadc173e40"}}'
}

The version key started at 2, the same version used for the hints. Since Borg has many versioned file formats, this keeps the number of different versions in use a bit lower.

The other keys map an auxiliary file, like index or hints to their integrity data. Note that the JSON is stored as-is, and not as part of the msgpack structure.

Integrity errors result in deleting the affected file(s) (index/hints) and rebuilding the index, which is the same action taken when corruption is noticed in other ways (e.g. HashIndex can detect most corrupted headers, but not data corruption). A warning is logged as well. The exit code is not influenced, since remote repositories cannot perform that action. Raising the exit code would be possible for local repositories, but is not implemented.

Unlike the cache design this mechanism can have false positives whenever an older version rewrites the auxiliary files for a transaction created by a newer version, since that might result in a different index (due to hash-table resizing) or hints file (hash ordering, or the older version 1 format), while not invalidating the integrity file.

For example, using 1.1 on a repository, noticing corruption or similar issues and then running borg-1.0 check --repair, which rewrites the index and hints, results in this situation. Borg 1.1 would erroneously report checksum errors in the hints and/or index files and trigger an automatic rebuild of these files.

All about JSON: How to develop frontends

Borg does not have a public API on the Python level. That does not keep you from writing import borg, but does mean that there are no release-to-release guarantees on what you might find in that package, not even for point releases (1.1.x), and there is no documentation beyond the code and the internals documents.

Borg does on the other hand provide an API on a command-line level. In other words, a frontend should (for example) create a backup archive just invoke borg create, give commandline parameters/options as needed and parse JSON output from borg.

Important: JSON output is expected to be UTF-8, but currently borg depends on the locale being configured for that (must be a UTF-8 locale and not "C" or "ascii"), so that Python will choose to encode to UTF-8. The same applies to any inputs read by borg, they are expected to be UTF-8 encoded also.

We consider this a bug (see #2273) and might fix it later, so borg will use UTF-8 independent of the locale.

On POSIX systems, you can usually set environment vars to choose a UTF-8 locale:

export LANG=en_US.UTF-8
export LC_CTYPE=en_US.UTF-8
Logging

Especially for graphical frontends it is important to be able to convey and reformat progress information in meaningful ways. The --log-json option turns the stderr stream of Borg into a stream of JSON lines, where each line is a JSON object. The type key of the object determines its other contents.

Warning

JSON logging requires successful argument parsing. Even with --log-json specified, a parsing error will be printed in plain text, because logging set-up happens after all arguments are parsed.

Since JSON can only encode text, any string representing a file system path may miss non-text parts.

The following types are in use. Progress information is governed by the usual rules for progress information, it is not produced unless --progress is specified.

archive_progress

Output during operations creating archives (borg create and borg recreate). The following keys exist, each represents the current progress.

original_size

Original size of data processed so far (before compression and deduplication, may be empty/absent)

compressed_size

Compressed size (may be empty/absent)

deduplicated_size

Deduplicated size (may be empty/absent)

nfiles

Number of (regular) files processed so far (may be empty/absent)

path

Current path (may be empty/absent)

time

Unix timestamp (float)

finished

boolean indicating whether the operation has finished, only the last object for an operation can have this property set to true.

progress_message

A message-based progress information with no concrete progress information, just a message saying what is currently being worked on.

operation

unique, opaque integer ID of the operation

msgid

Message ID of the operation (may be null)

finished

boolean indicating whether the operation has finished, only the last object for an operation can have this property set to true.

message

current progress message (may be empty/absent)

time

Unix timestamp (float)

progress_percent

Absolute progress information with defined end/total and current value.

operation

unique, opaque integer ID of the operation

msgid

Message ID of the operation (may be null)

finished

boolean indicating whether the operation has finished, only the last object for an operation can have this property set to true.

message

A formatted progress message, this will include the percentage and perhaps other information (absent for finished == true)

current

Current value (always less-or-equal to total, absent for finished == true)

info

Array that describes the current item, may be null, contents depend on msgid (absent for finished == true)

total

Total value (absent for finished == true)

time

Unix timestamp (float)

file_status

This is only output by borg create and borg recreate if --list is specified. The usual rules for the file listing applies, including the --filter option.

status

Single-character status as for regular list output

path

Path of the file system object

log_message

Any regular log output invokes this type. Regular log options and filtering applies to these as well.

time

Unix timestamp (float)

levelname

Upper-case log level name (also called severity). Defined levels are: DEBUG, INFO, WARNING, ERROR, CRITICAL

name

Name of the emitting entity

message

Formatted log message

msgid

Message ID, may be null or absent

See Prompts for the types used by prompts.

Examples (reformatted, each object would be on exactly one line)

borg extract progress:

{"message": "100.0% Extracting: src/borgbackup.egg-info/entry_points.txt",
 "current": 13000228, "total": 13004993, "info": ["src/borgbackup.egg-info/entry_points.txt"],
 "operation": 1, "msgid": "extract", "type": "progress_percent", "finished": false}
{"message": "100.0% Extracting: src/borgbackup.egg-info/SOURCES.txt",
 "current": 13004993, "total": 13004993, "info": ["src/borgbackup.egg-info/SOURCES.txt"],
 "operation": 1, "msgid": "extract", "type": "progress_percent", "finished": false}
{"operation": 1, "msgid": "extract", "type": "progress_percent", "finished": true}

borg create file listing with progress:

{"original_size": 0, "compressed_size": 0, "deduplicated_size": 0, "nfiles": 0, "type": "archive_progress", "path": "src"}
{"type": "file_status", "status": "U", "path": "src/borgbackup.egg-info/entry_points.txt"}
{"type": "file_status", "status": "U", "path": "src/borgbackup.egg-info/SOURCES.txt"}
{"type": "file_status", "status": "d", "path": "src/borgbackup.egg-info"}
{"type": "file_status", "status": "d", "path": "src"}
{"original_size": 13176040, "compressed_size": 11386863, "deduplicated_size": 503, "nfiles": 277, "type": "archive_progress", "path": ""}

Internal transaction progress:

{"message": "Saving files cache", "operation": 2, "msgid": "cache.commit", "type": "progress_message", "finished": false}
{"message": "Saving cache config", "operation": 2, "msgid": "cache.commit", "type": "progress_message", "finished": false}
{"message": "Saving chunks cache", "operation": 2, "msgid": "cache.commit", "type": "progress_message", "finished": false}
{"operation": 2, "msgid": "cache.commit", "type": "progress_message", "finished": true}

A debug log message:

{"message": "35 self tests completed in 0.08 seconds",
 "type": "log_message", "created": 1488278449.5575905, "levelname": "DEBUG", "name": "borg.archiver"}
Prompts

Prompts assume a JSON form as well when the --log-json option is specified. Responses are still read verbatim from stdin, while prompts are JSON messages printed to stderr, just like log messages.

Prompts use the question_prompt and question_prompt_retry types for the prompt itself, and question_invalid_answer, question_accepted_default, question_accepted_true, question_accepted_false and question_env_answer types for information about prompt processing.

The message property contains the same string displayed regularly in the same situation, while the msgid property may contain a msgid, typically the name of the environment variable that can be used to override the prompt. It is the same for all JSON messages pertaining to the same prompt.

Examples (reformatted, each object would be on exactly one line)

Providing an invalid answer:

{"type": "question_prompt", "msgid": "BORG_CHECK_I_KNOW_WHAT_I_AM_DOING",
 "message": "... Type 'YES' if you understand this and want to continue: "}
incorrect answer  # input on stdin
{"type": "question_invalid_answer", "msgid": "BORG_CHECK_I_KNOW_WHAT_I_AM_DOING", "is_prompt": false,
 "message": "Invalid answer, aborting."}

Providing a false (negative) answer:

{"type": "question_prompt", "msgid": "BORG_CHECK_I_KNOW_WHAT_I_AM_DOING",
 "message": "... Type 'YES' if you understand this and want to continue: "}
NO  # input on stdin
{"type": "question_accepted_false", "msgid": "BORG_CHECK_I_KNOW_WHAT_I_AM_DOING",
 "message": "Aborting.", "is_prompt": false}

Providing a true (affirmative) answer:

{"type": "question_prompt", "msgid": "BORG_CHECK_I_KNOW_WHAT_I_AM_DOING",
 "message": "... Type 'YES' if you understand this and want to continue: "}
YES  # input on stdin
# no further output, just like the prompt without --log-json
Passphrase prompts

Passphrase prompts should be handled differently. Use the environment variables BORG_PASSPHRASE and BORG_NEW_PASSPHRASE (see Environment Variables for reference) to pass passphrases to Borg, don't use the interactive passphrase prompts.

When setting a new passphrase (borg init, borg key change-passphrase) normally Borg prompts whether it should display the passphrase. This can be suppressed by setting the environment variable BORG_DISPLAY_PASSPHRASE to no.

When "confronted" with an unknown repository, where the application does not know whether the repository is encrypted, the following algorithm can be followed to detect encryption:

  1. Set BORG_PASSPHRASE to gibberish (for example a freshly generated UUID4, which cannot possibly be the passphrase)

  2. Invoke borg list repository ...

  3. If this fails, due the repository being encrypted and the passphrase obviously being wrong, you'll get an error with the PassphraseWrong msgid.

    The repository is encrypted, for further access the application will need the passphrase.

  4. If this does not fail, then the repository is not encrypted.

Standard output

stdout is different and more command-dependent than logging. Commands like borg info, borg create and borg list implement a --json option which turns their regular output into a single JSON object.

Some commands, like borg list and borg diff, can produce a lot of JSON. Since many JSON implementations don't support a streaming mode of operation, which is pretty much required to deal with this amount of JSON, these commands implement a --json-lines option which generates output in the JSON lines format, which is simply a number of JSON objects separated by new lines.

Dates are formatted according to ISO 8601 in local time. No explicit time zone is specified at this time (subject to change). The equivalent strftime format string is '%Y-%m-%dT%H:%M:%S.%f', e.g. 2017-08-07T12:27:20.123456.

The root object of '--json' output will contain at least a repository key with an object containing:

id

The ID of the repository, normally 64 hex characters

location

Canonicalized repository path, thus this may be different from what is specified on the command line

last_modified

Date when the repository was last modified by the Borg client

The encryption key, if present, contains:

mode

Textual encryption mode name (same as borg init --encryption names)

keyfile

Path to the local key file used for access. Depending on mode this key may be absent.

The cache key, if present, contains:

path

Path to the local repository cache

stats

Object containing cache stats:

total_chunks

Number of chunks

total_unique_chunks

Number of unique chunks

total_size

Total uncompressed size of all chunks multiplied with their reference counts

total_csize

Total compressed and encrypted size of all chunks multiplied with their reference counts

unique_size

Uncompressed size of all chunks

unique_csize

Compressed and encrypted size of all chunks

Example borg info output:

{
    "cache": {
        "path": "/home/user/.cache/borg/0cbe6166b46627fd26b97f8831e2ca97584280a46714ef84d2b668daf8271a23",
        "stats": {
            "total_chunks": 511533,
            "total_csize": 17948017540,
            "total_size": 22635749792,
            "total_unique_chunks": 54892,
            "unique_csize": 1920405405,
            "unique_size": 2449675468
        }
    },
    "encryption": {
        "mode": "repokey"
    },
    "repository": {
        "id": "0cbe6166b46627fd26b97f8831e2ca97584280a46714ef84d2b668daf8271a23",
        "last_modified": "2017-08-07T12:27:20.789123",
        "location": "/home/user/testrepo"
    },
    "security_dir": "/home/user/.config/borg/security/0cbe6166b46627fd26b97f8831e2ca97584280a46714ef84d2b668daf8271a23",
    "archives": []
}
Archive formats

borg info uses an extended format for archives, which is more expensive to retrieve, while borg list uses a simpler format that is faster to retrieve. Either return archives in an array under the archives key, while borg create returns a single archive object under the archive key.

Both formats contain a name key with the archive name, the id key with the hexadecimal archive ID, and the start key with the start timestamp.

borg info and borg create further have:

end

End timestamp

duration

Duration in seconds between start and end in seconds (float)

stats

Archive statistics (freshly calculated, this is what makes "info" more expensive)

original_size

Size of files and metadata before compression

compressed_size

Size after compression

deduplicated_size

Deduplicated size (against the current repository, not when the archive was created)

nfiles

Number of regular files in the archive

limits

Object describing the utilization of Borg limits

max_archive_size

Float between 0 and 1 describing how large this archive is relative to the maximum size allowed by Borg

command_line

Array of strings of the command line that created the archive

The note about paths from above applies here as well.

chunker_params

The chunker parameters the archive has been created with.

borg info further has:

hostname

Hostname of the creating host

username

Name of the creating user

comment

Archive comment, if any

Some keys/values are more expensive to compute than others (e.g. because it requires opening the archive, not just the manifest). To optimize for speed, borg list repo does not determine these values except when they are requested. The --format option is used for that (for normal mode as well as for --json mode), so, to have the comment included in the json output, you will need:

borg list repo --format "{name}{comment}" --json`

Example of a simple archive listing (borg list --last 1 --json):

{
    "archives": [
        {
            "id": "80cd07219ad725b3c5f665c1dcf119435c4dee1647a560ecac30f8d40221a46a",
            "name": "host-system-backup-2017-02-27",
            "start": "2017-08-07T12:27:20.789123"
        }
    ],
    "encryption": {
        "mode": "repokey"
    },
    "repository": {
        "id": "0cbe6166b46627fd26b97f8831e2ca97584280a46714ef84d2b668daf8271a23",
        "last_modified": "2017-08-07T12:27:20.789123",
        "location": "/home/user/repository"
    }
}

The same archive with more information (borg info --last 1 --json):

{
    "archives": [
        {
            "chunker_params": [
                "buzhash",
                13,
                23,
                16,
                4095
            ],
            "command_line": [
                "/home/user/.local/bin/borg",
                "create",
                "/home/user/repository",
                "..."
            ],
            "comment": "",
            "duration": 5.641542,
            "end": "2017-02-27T12:27:20.789123",
            "hostname": "host",
            "id": "80cd07219ad725b3c5f665c1dcf119435c4dee1647a560ecac30f8d40221a46a",
            "limits": {
                "max_archive_size": 0.0001330855110409714
            },
            "name": "host-system-backup-2017-02-27",
            "start": "2017-02-27T12:27:20.789123",
            "stats": {
                "compressed_size": 1880961894,
                "deduplicated_size": 2791,
                "nfiles": 53669,
                "original_size": 2400471280
            },
            "username": "user"
        }
    ],
    "cache": {
        "path": "/home/user/.cache/borg/0cbe6166b46627fd26b97f8831e2ca97584280a46714ef84d2b668daf8271a23",
        "stats": {
            "total_chunks": 511533,
            "total_csize": 17948017540,
            "total_size": 22635749792,
            "total_unique_chunks": 54892,
            "unique_csize": 1920405405,
            "unique_size": 2449675468
        }
    },
    "encryption": {
        "mode": "repokey"
    },
    "repository": {
        "id": "0cbe6166b46627fd26b97f8831e2ca97584280a46714ef84d2b668daf8271a23",
        "last_modified": "2017-08-07T12:27:20.789123",
        "location": "/home/user/repository"
    }
}
File listings

Each archive item (file, directory, ...) is described by one object in the borg list output. Refer to the borg list documentation for the available keys and their meaning.

Example (excerpt) of borg list --json-lines:

{"type": "d", "mode": "drwxr-xr-x", "user": "user", "group": "user", "uid": 1000, "gid": 1000, "path": "linux", "healthy": true, "source": "", "linktarget": "", "flags": null, "mtime": "2017-02-27T12:27:20.023407", "size": 0}
{"type": "d", "mode": "drwxr-xr-x", "user": "user", "group": "user", "uid": 1000, "gid": 1000, "path": "linux/baz", "healthy": true, "source": "", "linktarget": "", "flags": null, "mtime": "2017-02-27T12:27:20.585407", "size": 0}
Archive Differencing

Each archive difference item (file contents, user/group/mode) output by borg diff is represented by an ItemDiff object. The propertiese of an ItemDiff object are:

path:

The filename/path of the Item (file, directory, symlink).

changes:

A list of Change objects describing the changes made to the item in the two archives. For example, there will be two changes if the contents of a file are changed, and its ownership are changed.

The Change object can contain a number of properties depending on the type of change that occurred. If a 'property' is not required for the type of change, it is not output. The possible properties of a Change object are:

type:

The type property is always present. It identifies the type of change and will be one of these values:

  • modified - file contents changed.

  • added - the file was added.

  • removed - the file was removed.

  • added directory - the directory was added.

  • removed directory - the directory was removed.

  • added link - the symlink was added.

  • removed link - the symlink was removed.

  • changed link - the symlink target was changed.

  • mode - the file/directory/link mode was changed. Note - this could indicate a change from a file/directory/link type to a different type (file/directory/link), such as -- a file is deleted and replaced with a directory of the same name.

  • owner - user and/or group ownership changed.

size:

If type == 'added' or 'removed', then size provides the size of the added or removed file.

added:

If type == 'modified' and chunk ids can be compared, then added and removed indicate the amount of data 'added' and 'removed'. If chunk ids can not be compared, then added and removed properties are not provided and the only information available is that the file contents were modified.

removed:

See added property.

old_mode:

If type == 'mode', then old_mode and new_mode provide the mode and permissions changes.

new_mode:

See old_mode property.

old_user:

If type == 'owner', then old_user, new_user, old_group and new_group provide the user and group ownership changes.

old_group:

See old_user property.

new_user:

See old_user property.

new_group:

See old_user property.

Example (excerpt) of borg diff --json-lines:

{"path": "file1", "changes": [{"path": "file1", "changes": [{"type": "modified", "added": 17, "removed": 5}, {"type": "mode", "old_mode": "-rw-r--r--", "new_mode": "-rwxr-xr-x"}]}]}
{"path": "file2", "changes": [{"type": "modified", "added": 135, "removed": 252}]}
{"path": "file4", "changes": [{"type": "added", "size": 0}]}
{"path": "file3", "changes": [{"type": "removed", "size": 0}]}
Message IDs

Message IDs are strings that essentially give a log message or operation a name, without actually using the full text, since texts change more frequently. Message IDs are unambiguous and reduce the need to parse log messages.

Assigned message IDs are:

Errors
Archive.AlreadyExists

Archive {} already exists

Archive.DoesNotExist

Archive {} does not exist

Archive.IncompatibleFilesystemEncodingError

Failed to encode filename "{}" into file system encoding "{}". Consider configuring the LANG environment variable.

Cache.CacheInitAbortedError

Cache initialization aborted

Cache.EncryptionMethodMismatch

Repository encryption method changed since last access, refusing to continue

Cache.RepositoryAccessAborted

Repository access aborted

Cache.RepositoryIDNotUnique

Cache is newer than repository - do you have multiple, independently updated repos with same ID?

Cache.RepositoryReplay

Cache is newer than repository - this is either an attack or unsafe (multiple repos with same ID)

Buffer.MemoryLimitExceeded

Requested buffer size {} is above the limit of {}.

ExtensionModuleError

The Borg binary extension modules do not seem to be properly installed

IntegrityError

Data integrity error: {}

NoManifestError

Repository has no manifest.

PlaceholderError

Formatting Error: "{}".format({}): {}({})

KeyfileInvalidError

Invalid key file for repository {} found in {}.

KeyfileMismatchError

Mismatch between repository {} and key file {}.

KeyfileNotFoundError

No key file for repository {} found in {}.

PassphraseWrong

passphrase supplied in BORG_PASSPHRASE is incorrect

PasswordRetriesExceeded

exceeded the maximum password retries

RepoKeyNotFoundError

No key entry found in the config of repository {}.

UnsupportedManifestError

Unsupported manifest envelope. A newer version is required to access this repository.

UnsupportedPayloadError

Unsupported payload type {}. A newer version is required to access this repository.

NotABorgKeyFile

This file is not a borg key backup, aborting.

RepoIdMismatch

This key backup seems to be for a different backup repository, aborting.

UnencryptedRepo

Keymanagement not available for unencrypted repositories.

UnknownKeyType

Keytype {0} is unknown.

LockError

Failed to acquire the lock {}.

LockErrorT

Failed to acquire the lock {}.

ConnectionClosed

Connection closed by remote host

InvalidRPCMethod

RPC method {} is not valid

PathNotAllowed

Repository path not allowed

RemoteRepository.RPCServerOutdated

Borg server is too old for {}. Required version {}

UnexpectedRPCDataFormatFromClient

Borg {}: Got unexpected RPC data format from client.

UnexpectedRPCDataFormatFromServer

Got unexpected RPC data format from server: {}

Repository.AlreadyExists

Repository {} already exists.

Repository.CheckNeeded

Inconsistency detected. Please run "borg check {}".

Repository.DoesNotExist

Repository {} does not exist.

Repository.InsufficientFreeSpaceError

Insufficient free space to complete transaction (required: {}, available: {}).

Repository.InvalidRepository

{} is not a valid repository. Check repo config.

Repository.AtticRepository

Attic repository detected. Please run "borg upgrade {}".

Repository.ObjectNotFound

Object with key {} not found in repository {}.

Operations
  • cache.begin_transaction

  • cache.download_chunks, appears with borg create --no-cache-sync

  • cache.commit

  • cache.sync

    info is one string element, the name of the archive currently synced.

  • repository.compact_segments

  • repository.replay_segments

  • repository.check

  • check.verify_data

  • check.rebuild_manifest

  • check.rebuild_refcounts

  • extract

    info is one string element, the name of the path currently extracted.

  • extract.permissions

  • archive.delete

  • archive.calc_stats

  • prune

  • upgrade.convert_segments

Prompts
BORG_UNKNOWN_UNENCRYPTED_REPO_ACCESS_IS_OK

For "Warning: Attempting to access a previously unknown unencrypted repository"

BORG_RELOCATED_REPO_ACCESS_IS_OK

For "Warning: The repository at location ... was previously located at ..."

BORG_CHECK_I_KNOW_WHAT_I_AM_DOING

For "This is a potentially dangerous function..." (check --repair)

BORG_DELETE_I_KNOW_WHAT_I_AM_DOING

For "You requested to completely DELETE the repository including all archives it contains:"

Development

This chapter will get you started with Borg development.

Borg is written in Python (with a little bit of Cython and C for the performance critical parts).

Contributions

... are welcome!

Some guidance for contributors:

  • Discuss changes on the GitHub issue tracker, on IRC or on the mailing list.

  • Make your PRs on the master branch (see Branching Model for details).

  • Do clean changesets:

    • Focus on some topic, resist changing anything else.

    • Do not do style changes mixed with functional changes.

    • Try to avoid refactorings mixed with functional changes.

    • If you need to fix something after commit/push:

      • If there are ongoing reviews: do a fixup commit you can squash into the bad commit later.

      • If there are no ongoing reviews or you did not push the bad commit yet: amend the commit to include your fix or merge the fixup commit before pushing.

    • Have a nice, clear, typo-free commit comment.

    • If you fixed an issue, refer to it in your commit comment.

    • Follow the style guide (see below).

  • If you write new code, please add tests and docs for it.

  • Run the tests, fix any issues that come up.

  • Make a pull request on GitHub.

  • Wait for review by other developers.

Branching model

Borg development happens on the master branch and uses GitHub pull requests (if you don't have GitHub or don't want to use it you can send smaller patches via the borgbackup mailing list to the maintainers).

Stable releases are maintained on maintenance branches named x.y-maint, eg. the maintenance branch of the 1.0.x series is 1.0-maint.

Most PRs should be filed against the master branch. Only if an issue affects only a particular maintenance branch a PR should be filed against it directly.

While discussing / reviewing a PR it will be decided whether the change should be applied to maintenance branches. Each maintenance branch has a corresponding backport/x.y-maint label, which will then be applied.

Changes that are typically considered for backporting:

  • Data loss, corruption and inaccessibility fixes.

  • Security fixes.

  • Forward-compatibility improvements.

  • Documentation corrections.

Maintainer part

From time to time a maintainer will backport the changes for a maintenance branch, typically before a release or if enough changes were collected:

  1. Notify others that you're doing this to avoid duplicate work.

  2. Branch a backporting branch off the maintenance branch.

  3. Cherry pick and backport the changes from each labelled PR, remove the label for each PR you've backported.

    To preserve authorship metadata, do not follow the git cherry-pick instructions to use git commit after resolving conflicts. Instead, stage conflict resolutions and run git cherry-pick --continue, much like using git rebase.

    To avoid merge issues (a cherry pick is a form of merge), use these options (similar to the git merge options used previously, the -x option adds a reference to the original commit):

    git cherry-pick --strategy recursive -X rename-threshold=5% -x
    
  4. Make a PR of the backporting branch against the maintenance branch for backport review. Mention the backported PRs in this PR, e.g.:

    Includes changes from #2055 #2057 #2381

    This way GitHub will automatically show in these PRs where they were backported.

Historic model

Previously (until release 1.0.10) Borg used a "merge upwards" model where most minor changes and fixes where committed to a maintenance branch (eg. 1.0-maint), and the maintenance branch(es) were regularly merged back into the main development branch. This became more and more troublesome due to merges growing more conflict-heavy and error-prone.

Code and issues

Code is stored on GitHub, in the Borgbackup organization. Issues and pull requests should be sent there as well. See also the Support section for more details.

Style guide

We generally follow pep8, with 120 columns instead of 79. We do not use form-feed (^L) characters to separate sections either. Compliance is tested automatically when you run the tests.

Continuous Integration

All pull requests go through GitHub Actions, which runs the tests on misc. Python versions and on misc. platforms as well as some additional checks.

Output and Logging

When writing logger calls, always use correct log level (debug only for debugging, info for informative messages, warning for warnings, error for errors, critical for critical errors/states).

When directly talking to the user (e.g. Y/N questions), do not use logging, but directly output to stderr (not: stdout, it could be connected to a pipe).

To control the amount and kinds of messages output emitted at info level, use flags like --stats or --list, then create a topic logger for messages controlled by that flag. See _setup_implied_logging() in borg/archiver.py for the entry point to topic logging.

Building a development environment

First, just install borg into a virtual env as described before.

To install some additional packages needed for running the tests, activate your virtual env and run:

pip install -r requirements.d/development.txt

This project utilizes pre-commit to lint code before it is committed. Although pre-commit is installed when running the command above, the pre-commit hooks will have to be installed separately. Run this command to install the pre-commit hooks:

pre-commit install

Running the tests

The tests are in the borg/testsuite package.

To run all the tests, you need to have fakeroot installed. If you do not have fakeroot, you still will be able to run most tests, just leave away the fakeroot -u from the given command lines.

To run the test suite use the following command:

fakeroot -u tox  # run all tests

Some more advanced examples:

# verify a changed tox.ini (run this after any change to tox.ini):
fakeroot -u tox --recreate

fakeroot -u tox -e py38  # run all tests, but only on python 3.8

fakeroot -u tox borg.testsuite.locking  # only run 1 test module

fakeroot -u tox borg.testsuite.locking -- -k '"not Timer"'  # exclude some tests

fakeroot -u tox borg.testsuite -- -v  # verbose py.test

Important notes:

  • When using -- to give options to py.test, you MUST also give borg.testsuite[.module].

Adding a compression algorithm

If you want to add a new compression algorithm, please refer to #1633 and leave a post there in order to discuss about the proposal.

Documentation

Generated files

Usage documentation (found in docs/usage/) and man pages (docs/man/) are generated automatically from the command line parsers declared in the program and their documentation, which is embedded in the program (see archiver.py). These are committed to git for easier use by packagers downstream.

When a command is added, a command line flag changed, added or removed, the usage docs need to be rebuilt as well:

python setup.py build_usage
python setup.py build_man

However, we prefer to do this as part of our Creating a new release preparations, so it is generally not necessary to update these when submitting patches that change something about the command line.

Building the docs with Sphinx

The documentation (in reStructuredText format, .rst) is in docs/.

To build the html version of it, you need to have Sphinx installed (in your Borg virtualenv with Python 3):

pip install -r requirements.d/docs.txt

Now run:

cd docs/
make html

Then point a web browser at docs/_build/html/index.html.

The website is updated automatically by ReadTheDocs through GitHub web hooks on the main repository.

Using Vagrant

We use Vagrant for the automated creation of testing environments and borgbackup standalone binaries for various platforms.

For better security, there is no automatic sync in the VM to host direction. The plugin vagrant-scp is useful to copy stuff from the VMs to the host.

The "windows10" box requires the reload plugin (vagrant plugin install vagrant-reload).

Usage:

# To create and provision the VM:
vagrant up OS
# same, but use 6 VM cpus and 12 workers for pytest:
VMCPUS=6 XDISTN=12 vagrant up OS
# To create an ssh session to the VM:
vagrant ssh OS
# To execute a command via ssh in the VM:
vagrant ssh OS -c "command args"
# To shut down the VM:
vagrant halt OS
# To shut down and destroy the VM:
vagrant destroy OS
# To copy files from the VM (in this case, the generated binary):
vagrant scp OS:/vagrant/borg/borg.exe .

Creating standalone binaries

Make sure you have everything built and installed (including fuse stuff). When using the Vagrant VMs, pyinstaller will already be installed.

With virtual env activated:

pip install pyinstaller  # or git checkout master
pyinstaller -F -n borg-PLATFORM borg/__main__.py
for file in dist/borg-*; do gpg --armor --detach-sign $file; done

If you encounter issues, see also our Vagrantfile for details.

Note

Standalone binaries built with pyinstaller are supposed to work on same OS, same architecture (x86 32bit, amd64 64bit) without external dependencies.

Creating a new release

Checklist:

  • Make sure all issues for this milestone are closed or moved to the next milestone.

  • Check if there are any pending fixes for security issues.

  • Find and fix any low hanging fruit left on the issue tracker.

  • Check that GitHub Actions CI is happy.

  • Update CHANGES.rst, based on git log $PREVIOUS_RELEASE...

  • Check version number of upcoming release in CHANGES.rst.

  • Render CHANGES.rst via make html and check for markup errors.

  • Verify that MANIFEST.in and setup.py are complete.

  • python setup.py build_usage ; python setup.py build_man and commit.

  • Tag the release:

    git tag -s -m "tagged/signed release X.Y.Z" X.Y.Z
    
  • Create a clean repo and use it for the following steps:

    git clone borg borg-clean
    

    This makes sure no uncommitted files get into the release archive. It will also reveal uncommitted required files. Moreover, it makes sure the vagrant machines only get committed files and do a fresh start based on that.

  • Run tox and/or binary builds on all supported platforms via vagrant, check for test failures.

  • Create sdist, sign it, upload release to (test) PyPi:

    scripts/sdist-sign X.Y.Z
    scripts/upload-pypi X.Y.Z test
    scripts/upload-pypi X.Y.Z
    

    Note: the signature is not uploaded to PyPi any more, but we upload it to github releases.

  • Put binaries into dist/borg-OSNAME and sign them:

    scripts/sign-binaries 201912312359
    
  • Close the release milestone on GitHub.

  • Update borgbackup.org with the new version number and release date.

  • Announce on:

  • Mailing list.

  • Twitter.

  • IRC channel (change /topic).

  • Create a GitHub release, include:

    • pypi dist package and signature

    • Standalone binaries (see above for how to create them).

      • For macOS, document the macFUSE version in the README of the binaries. macFUSE uses a kernel extension that needs to be compatible with the code contained in the binary.

    • A link to CHANGES.rst.

Authors

E-mail addresses listed here are not intended for support, please see the support section instead.

Borg authors ("The Borg Collective")

Retired

Borg is a fork of Attic.

Attic authors

Attic is written and maintained by Jonas Borgström and various contributors:

Attic Development Lead
Attic Patches and Suggestions
  • Brian Johnson

  • Cyril Roussillon

  • Dan Christensen

  • Jeremy Maitin-Shepard

  • Johann Klähn

  • Petros Moisiadis

  • Thomas Waldmann

BLAKE2

Borg includes BLAKE2: Copyright 2012, Samuel Neves <sneves@dei.uc.pt>, licensed under the terms of the CC0, the OpenSSL Licence, or the Apache Public License 2.0.

Slicing CRC32

Borg includes a fast slice-by-8 implementation of CRC32, Copyright 2011-2015 Stephan Brumme, licensed under the terms of a zlib license. See http://create.stephan-brumme.com/crc32/

Folding CRC32

Borg includes an extremely fast folding implementation of CRC32, Copyright 2013 Intel Corporation, licensed under the terms of the zlib license.

xxHash

XXH64, a fast non-cryptographic hash algorithm. Copyright 2012-2016 Yann Collet, licensed under a BSD 2-clause license.

License

Copyright (C) 2015-2023 The Borg Collective (see AUTHORS file)
Copyright (C) 2010-2014 Jonas Borgström <jonas@borgstrom.se>
All rights reserved.

Redistribution and use in source and binary forms, with or without
modification, are permitted provided that the following conditions
are met:

 1. Redistributions of source code must retain the above copyright
    notice, this list of conditions and the following disclaimer.
 2. Redistributions in binary form must reproduce the above copyright
    notice, this list of conditions and the following disclaimer in
    the documentation and/or other materials provided with the
    distribution.
 3. The name of the author may not be used to endorse or promote
    products derived from this software without specific prior
    written permission.

THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.