12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049105010511052105310541055105610571058105910601061106210631064106510661067106810691070107110721073107410751076107710781079108010811082108310841085108610871088 |
- .. include:: global.rst.inc
- .. highlight:: none
- .. _faq:
- Frequently asked questions
- ==========================
- Usage & Limitations
- ###################
- What is the difference between a repo on an external hard drive vs. repo on a server?
- -------------------------------------------------------------------------------------
- If Borg is running in client/server mode, the client uses SSH as a transport to
- talk to the remote agent, which is another Borg process (Borg is installed on
- the server, too) started automatically by the client. The Borg server is doing
- storage-related low-level repo operations (list, load and store objects),
- while the Borg client does the high-level stuff: deduplication, encryption,
- compression, dealing with archives, backups, restores, etc., which reduces the
- amount of data that goes over the network.
- When Borg is writing to a repo on a locally mounted remote file system, e.g.
- SSHFS, the Borg client only can do file system operations and has no agent
- running on the remote side, so *every* operation needs to go over the network,
- which is slower.
- Can I back up from multiple servers into a single repository?
- -------------------------------------------------------------
- Yes, you can! Even simultaneously.
- Can I back up to multiple, swapped backup targets?
- --------------------------------------------------
- It is possible to swap your backup disks if each backup medium is assigned its
- own repository by creating a new one with :ref:`borg_rcreate`.
- Can I copy or synchronize my repo to another location?
- ------------------------------------------------------
- If you want to have redundant backup repositories (preferably at separate
- locations), the recommended way to do that is like this:
- - ``borg repo-create repo1 --encryption=X``
- - ``borg repo-create repo2 --encryption=X --other-repo=repo1``
- - maybe do a snapshot to have stable and same input data for both borg create.
- - client machine ---borg create---> repo1
- - client machine ---borg create---> repo2
- This will create distinct (different repo ID), but related repositories.
- Related means using the same chunker secret and the same id_key, thus producing
- the same chunks / the same chunk ids if the input data is the same.
- The 2 independent borg create invocations mean that there is no error propagation
- from repo1 to repo2 when done like that.
- An alternative way would be to use ``borg transfer`` to copy backup archives
- from repo1 to repo2. Likely a bit more efficient and the archives would be identical,
- but suffering from potential error propagation.
- Warning: using borg with multiple repositories with identical repository ID (like when
- creating 1:1 repository copies) is not supported and can lead to all sorts of issues,
- like e.g. cache coherency issues, malfunction, data corruption.
- "this is either an attack or unsafe" warning
- --------------------------------------------
- About the warning:
- Cache, or information obtained from the security directory is newer than
- repository - this is either an attack or unsafe (multiple repos with same ID)
- "unsafe": If not following the advice from the previous section, you can easily
- run into this by yourself by restoring an older copy of your repository.
- "attack": maybe an attacker has replaced your repo by an older copy, trying to
- trick you into AES counter reuse, trying to break your repo encryption.
- Borg users have also reported that fs issues (like hw issues / I/O errors causing
- the fs to become read-only) can cause this warning, see :issue:`7853`.
- If you decide to ignore this and accept unsafe operation for this repository,
- you could delete the manifest-timestamp and the local cache:
- ::
- borg config id # shows the REPO_ID
- rm ~/.config/borg/security/REPO_ID/manifest-timestamp
- borg repo-delete --cache-only
- This is an unsafe and unsupported way to use borg, you have been warned.
- Which file types, attributes, etc. are *not* preserved?
- -------------------------------------------------------
- * UNIX domain sockets (because it does not make sense - they are
- meaningless without the running process that created them and the process
- needs to recreate them in any case). So, don't panic if your backup
- misses a UDS!
- * The precise on-disk (or rather: not-on-disk) representation of the holes
- in a sparse file.
- Archive creation has no special support for sparse files, holes are
- backed up as (deduplicated and compressed) runs of zero bytes.
- Archive extraction has optional support to extract all-zero chunks as
- holes in a sparse file.
- * Some filesystem specific attributes, like btrfs NOCOW, see :ref:`platforms`.
- Are there other known limitations?
- ----------------------------------
- - borg extract supports restoring only into an empty destination. After extraction,
- the destination will have exactly the contents of the extracted archive.
- If you extract into a non-empty destination, borg will (for example) not
- remove files which are in the destination, but not in the archive.
- See :issue:`4598` for a workaround and more details.
- .. _interrupted_backup:
- If a backup stops mid-way, does the already-backed-up data stay there?
- ----------------------------------------------------------------------
- Yes, the data transferred into the repo stays there - just avoid running
- ``borg compact`` before you completed the backup, because that would remove
- chunks that were already transferred to the repo, but not (yet) referenced
- by an archive.
- If a backup was interrupted, you normally do not need to do anything special,
- just invoke ``borg create`` as you always do. You may use the same archive name
- as in previous attempt or a different one (e.g. if you always include the
- current datetime), it does not matter.
- Borg always does full single-pass backups, so it will start again
- from the beginning - but it will be much faster, because some of the data was
- already stored into the repo, so it does not need to get transmitted and stored
- again.
- How can I back up huge file(s) over a unstable connection?
- ----------------------------------------------------------
- Yes. For more details, see :ref:`interrupted_backup`.
- How can I restore huge file(s) over an unstable connection?
- -----------------------------------------------------------
- Try using ``borg mount`` and ``rsync`` (or a similar tool that supports
- resuming a partial file copy from what's already copied).
- How can I switch append-only mode on and off?
- -----------------------------------------------------------------------------------------------------------------------------------
- You could do that (via borg config REPO append_only 0/1), but using different
- ssh keys and different entries in ``authorized_keys`` is much easier and also
- maybe has less potential of things going wrong somehow.
- My machine goes to sleep causing `Broken pipe`
- ----------------------------------------------
- While backing up your data over the network, your machine should not go to sleep.
- On macOS you can use `caffeinate` to avoid that.
- How can I compare contents of an archive to my local filesystem?
- -----------------------------------------------------------------
- You can instruct ``export-tar`` to send a tar stream to the stdout, and
- then use ``tar`` to perform the comparison:
- ::
- borg export-tar archive-name - | tar --compare -f - -C /path/to/compare/to
- Can Borg add redundancy to the backup data to deal with hardware malfunction?
- -----------------------------------------------------------------------------
- No, it can't. While that at first sounds like a good idea to defend against
- some defect HDD sectors or SSD flash blocks, dealing with this in a
- reliable way needs a lot of low-level storage layout information and
- control which we do not have (and also can't get, even if we wanted).
- So, if you need that, consider RAID or a filesystem that offers redundant
- storage or just make backups to different locations / different hardware.
- See also :issue:`225`.
- Can Borg verify data integrity of a backup archive?
- ---------------------------------------------------
- Yes, if you want to detect accidental data damage (like bit rot), use the
- ``check`` operation. It will notice corruption using CRCs and hashes.
- If you want to be able to detect malicious tampering also, use an encrypted
- repo. It will then be able to check using CRCs and HMACs.
- .. _faq-integrityerror:
- I get an IntegrityError or similar - what now?
- ----------------------------------------------
- A single error does not necessarily indicate bad hardware or a Borg
- bug. All hardware exhibits a bit error rate (BER). Hard drives are typically
- specified as exhibiting fewer than one error every 12 to 120 TB
- (one bit error in 10e14 to 10e15 bits). The specification is often called
- *unrecoverable read error rate* (URE rate).
- Apart from these very rare errors there are two main causes of errors:
- (i) Defective hardware: described below.
- (ii) Bugs in software (Borg, operating system, libraries):
- Ensure software is up to date.
- Check whether the issue is caused by any fixed bugs described in
- :ref:`important_notes`.
- .. rubric:: Finding defective hardware
- .. note::
- Hardware diagnostics are operating system dependent and do not
- apply universally. The commands shown apply for popular Unix-like
- systems. Refer to your operating system's manual.
- Checking hard drives
- Find the drive containing the repository and use *findmnt*, *mount* or *lsblk*
- to learn the device path (typically */dev/...*) of the drive.
- Then, smartmontools can retrieve self-diagnostics of the drive in question::
- # smartctl -a /dev/sdSomething
- The *Offline_Uncorrectable*, *Current_Pending_Sector* and *Reported_Uncorrect*
- attributes indicate data corruption. A high *UDMA_CRC_Error_Count* usually
- indicates a bad cable.
- I/O errors logged by the system (refer to the system journal or
- dmesg) can point to issues as well. I/O errors only affecting the
- file system easily go unnoticed, since they are not reported to
- applications (e.g. Borg), while these errors can still corrupt data.
- Drives can corrupt some sectors in one event, while remaining
- reliable otherwise. Conversely, drives can fail completely with no
- advance warning. If in doubt, copy all data from the drive in
- question to another drive -- just in case it fails completely.
- If any of these are suspicious, a self-test is recommended::
- # smartctl -t long /dev/sdSomething
- Running ``fsck`` if not done already might yield further insights.
- Checking memory
- Intermittent issues, such as ``borg check`` finding errors
- inconsistently between runs, are frequently caused by bad memory.
- Run memtest86+ (or an equivalent memory tester) to verify that
- the memory subsystem is operating correctly.
- Checking processors
- Processors rarely cause errors. If they do, they are usually overclocked
- or otherwise operated outside their specifications. We do not recommend to
- operate hardware outside its specifications for productive use.
- Tools to verify correct processor operation include Prime95 (mprime), linpack,
- and the `Intel Processor Diagnostic Tool
- <https://downloadcenter.intel.com/download/19792/Intel-Processor-Diagnostic-Tool>`_
- (applies only to Intel processors).
- .. rubric:: Repairing a damaged repository
- With any defective hardware found and replaced, the damage done to the repository
- needs to be ascertained and fixed.
- :ref:`borg_check` provides diagnostics and ``--repair`` options for repositories with
- issues. We recommend to first run without ``--repair`` to assess the situation.
- If the found issues and proposed repairs seem right, re-run "check" with ``--repair`` enabled.
- How probable is it to get a hash collision problem?
- ---------------------------------------------------
- If you noticed, there are some issues (:issue:`170` (**warning: hell**) and :issue:`4884`)
- about the probability of a chunk having the same hash as another chunk, making the file
- corrupted because it grabbed the wrong chunk. This is called the `Birthday Problem
- <https://en.wikipedia.org/wiki/Birthday_problem>`_.
- There is a lot of probability in here so, I can give you my interpretation of
- such math but it's honestly better that you read it yourself and grab your own
- resolution from that.
- Assuming that all your chunks have a size of :math:`2^{21}` bytes (approximately 2.1 MB)
- and we have a "perfect" hash algorithm, we can think that the probability of collision
- would be of :math:`p^2/2^{n+1}` then, using SHA-256 (:math:`n=256`) and for example
- we have 1000 million chunks (:math:`p=10^9`) (1000 million chunks would be about 2100TB).
- The probability would be around 0.0000000000000000000000000000000000000000000000000000000000043.
- A mass-murderer space rock happens about once every 30 million years on average.
- This leads to a probability of such an event occurring in the next second to about :math:`10^{-15}`.
- That's **45** orders of magnitude more probable than the SHA-256 collision. Briefly stated,
- if you find SHA-256 collisions scary then your priorities are wrong. This example was grabbed from
- `this SO answer <https://stackoverflow.com/a/4014407/13359375>`_, it's great honestly.
- Still, the real question is whether Borg tries not to make this happen?
- Well... previously it did not check anything until there was a feature added which saves the size
- of the chunks too, so the size of the chunks is compared to the size that you got with the
- hash and if the check says there is a mismatch it will raise an exception instead of corrupting
- the file. This doesn't save us from everything but reduces the chances of corruption.
- There are other ways of trying to escape this but it would affect performance so much that
- it wouldn't be worth it and it would contradict Borg's design, so if you don't want this to
- happen, simply don't use Borg.
- Why is the time elapsed in the archive stats different from wall clock time?
- ----------------------------------------------------------------------------
- Borg needs to write the time elapsed into the archive metadata before finalizing
- the archive and saving the files cache.
- This means when Borg is run with e.g. the ``time`` command, the duration shown
- in the archive stats may be shorter than the full time the command runs for.
- How do I configure different prune policies for different directories?
- ----------------------------------------------------------------------
- Say you want to prune ``/var/log`` faster than the rest of
- ``/``. How do we implement that? The answer is to back up to different
- archive *names* and then implement different prune policies for
- different prefixes. For example, you could have a script that does::
- borg create --exclude var/log main-$(date +%Y-%m-%d) /
- borg create logs-$(date +%Y-%m-%d) /var/log
- Then you would have two different prune calls with different policies::
- borg prune --verbose --list -d 30 -a 'sh:main-*'
- borg prune --verbose --list -d 7 -a 'sh:logs-*'
- This will keep 7 days of logs and 30 days of everything else.
- How do I remove files from an existing backup?
- ----------------------------------------------
- A file is only removed from a BorgBackup repository if all archives that contain
- the file are deleted and the corresponding data chunks are removed from the
- repository. There are two ways how to remove files from a repository.
- 1. Use :ref:`borg_delete` to remove all archives that contain the files. This
- will of course delete everything in the archive, not only some files.
- 2. If you really want to remove only some specific files, you can run the
- :ref:`borg_recreate` command to rewrite all archives with a different
- ``--exclude`` pattern. See the examples in the manpage for more information.
- Finally, run :ref:`borg_compact` to delete the data chunks from the repository.
- Can I safely change the compression level or algorithm?
- --------------------------------------------------------
- The compression level and algorithm don't affect deduplication. Chunk ID hashes
- are calculated *before* compression. New compression settings
- will only be applied to new chunks, not existing chunks. So it's safe
- to change them.
- Use ``borg repo-compress`` to efficiently recompress a complete repository.
- Security
- ########
- .. _home_config_borg:
- How important is the $HOME/.config/borg directory?
- --------------------------------------------------
- The Borg config directory has content that you should take care of:
- ``keys`` subdirectory
- All your borg keyfile keys are stored in this directory. Please note that
- borg repokey keys are stored inside the repository. You MUST make sure to have an
- independent backup of these keyfiles, otherwise you cannot access your backups anymore if you lose
- them. You also MUST keep these files secret; everyone who gains access to your repository and has
- the corresponding keyfile (and the key passphrase) can extract it.
- Make sure that only you have access to the Borg config directory.
- .. _home_data_borg:
- How important is the $HOME/.local/share/borg directory?
- -------------------------------------------------------
- The Borg data directory has content that you should take care of:
- ``security`` subdirectory
- Each directory here represents one Borg repository by its ID and contains the last known status.
- If a repository's status is different from this information at the beginning of BorgBackup
- operation, Borg outputs warning messages and asks for confirmation, so make sure you do not lose
- or manipulate these files. However, apart from those warnings, a loss of these files can be
- recovered.
- Make sure that only you have access to the Borg data directory.
- .. _cache_security:
- Do I need to take security precautions regarding the cache?
- -----------------------------------------------------------
- The cache contains a lot of metadata information about the files in
- your repositories and it is not encrypted.
- However, the assumption is that the cache is being stored on the very
- same system which also contains the original files which are being
- backed up. So someone with access to the cache files would also have
- access the original files anyway.
- The Internals section contains more details about :ref:`cache`. If you ever need to move the cache
- to a different location, this can be achieved by using the appropriate :ref:`env_vars`.
- How can I specify the encryption passphrase programmatically?
- -------------------------------------------------------------
- There are several ways to specify a passphrase without human intervention:
- Setting ``BORG_PASSPHRASE``
- The passphrase can be specified using the ``BORG_PASSPHRASE`` environment variable.
- This is often the simplest option, but can be insecure if the script that sets it
- is world-readable.
- .. _password_env:
- .. note:: Be careful how you set the environment; using the ``env``
- command, a ``system()`` call or using inline shell scripts
- (e.g. ``BORG_PASSPHRASE=hunter2 borg ...``)
- might expose the credentials in the process list directly
- and they will be readable to all users on a system. Using
- ``export`` in a shell script file should be safe, however, as
- the environment of a process is `accessible only to that
- user
- <https://security.stackexchange.com/questions/14000/environment-variable-accessibility-in-linux/14009#14009>`_.
- Using ``BORG_PASSCOMMAND`` with a file of proper permissions
- Another option is to create a file with a password in it in your home
- directory and use permissions to keep anyone else from reading it. For
- example, first create a key::
- (umask 0077; head -c 32 /dev/urandom | base64 -w 0 > ~/.borg-passphrase)
- Then in an automated script one can put::
- export BORG_PASSCOMMAND="cat $HOME/.borg-passphrase"
- and Borg will automatically use that passphrase.
- Using keyfile-based encryption with a blank passphrase
- It is possible to encrypt your repository in ``keyfile`` mode instead of the default
- ``repokey`` mode and use a blank passphrase for the key file (simply press Enter twice
- when ``borg repo-create`` asks for the password). See :ref:`encrypted_repos`
- for more details.
- Using ``BORG_PASSCOMMAND`` with macOS Keychain
- macOS has a native manager for secrets (such as passphrases) which is safer
- than just using a file as it is encrypted at rest and unlocked manually
- (fortunately, the login keyring automatically unlocks when you log in). With
- the built-in ``security`` command, you can access it from the command line,
- making it useful for ``BORG_PASSCOMMAND``.
- First generate a passphrase and use ``security`` to save it to your login
- (default) keychain::
- security add-generic-password -D secret -U -a $USER -s borg-passphrase -w $(head -c 32 /dev/urandom | base64 -w 0)
- In your backup script retrieve it in the ``BORG_PASSCOMMAND``::
- export BORG_PASSCOMMAND="security find-generic-password -a $USER -s borg-passphrase -w"
- Using ``BORG_PASSCOMMAND`` with GNOME Keyring
- GNOME also has a keyring daemon that can be used to store a Borg passphrase.
- First ensure ``libsecret-tools``, ``gnome-keyring`` and ``libpam-gnome-keyring``
- are installed. If ``libpam-gnome-keyring`` wasn't already installed, ensure it
- runs on login::
- sudo sh -c "echo session optional pam_gnome_keyring.so auto_start >> /etc/pam.d/login"
- sudo sh -c "echo password optional pam_gnome_keyring.so >> /etc/pam.d/passwd"
- # you may need to relogin afterwards to activate the login keyring
- Then add a secret to the login keyring::
- head -c 32 /dev/urandom | base64 -w 0 | secret-tool store borg-repository repo-name --label="Borg Passphrase"
- If a dialog box pops up prompting you to pick a password for a new keychain, use your
- login password. If there is a checkbox for automatically unlocking on login, check it
- to allow backups without any user intervention whatsoever.
- Once the secret is saved, retrieve it in a backup script using ``BORG_PASSCOMMAND``::
- export BORG_PASSCOMMAND="secret-tool lookup borg-repository repo-name"
- .. note:: For this to unlock the keychain automatically it must be run
- in the ``dbus`` session of an unlocked terminal; for example, running a backup
- script as a ``cron`` job might not work unless you also ``export DISPLAY=:0``
- so ``secret-tool`` can pick up your open session. `It gets even more complicated`__
- when you are running the tool as a different user (e.g. running a backup as root
- with the password stored in the user keyring).
- __ https://github.com/borgbackup/borg/pull/2837#discussion_r127641330
- Using ``BORG_PASSCOMMAND`` with KWallet
- KDE also has a keychain feature in the form of KWallet. The command-line tool
- ``kwalletcli`` can be used to store and retrieve secrets. Ensure ``kwalletcli``
- is installed, generate a passphrase, and store it in your "wallet"::
- head -c 32 /dev/urandom | base64 -w 0 | kwalletcli -Pe borg-passphrase -f Passwords
- Once the secret is saved, retrieve it in a backup script using ``BORG_PASSCOMMAND``::
- export BORG_PASSCOMMAND="kwalletcli -e borg-passphrase -f Passwords"
- When backing up to remote encrypted repos, is encryption done locally?
- ----------------------------------------------------------------------
- Yes, file and directory metadata and data is locally encrypted, before
- leaving the local machine. We do not mean the transport layer encryption
- by that, but the data/metadata itself. Transport layer encryption (e.g.
- when ssh is used as a transport) applies additionally.
- When backing up to remote servers, do I have to trust the remote server?
- ------------------------------------------------------------------------
- Yes and No.
- No, as far as data confidentiality is concerned - if you use encryption,
- all your files/dirs data and metadata are stored in their encrypted form
- into the repository.
- Yes, as an attacker with access to the remote server could delete (or
- otherwise make unavailable) all your backups.
- How can I protect against a hacked backup client?
- -------------------------------------------------
- Assume you back up your backup client machine C to the backup server S and
- C gets hacked. In a simple push setup, the attacker could then use borg on
- C to delete all backups residing on S.
- These are your options to protect against that:
- - Do not allow to delete data permanently from the repo, see :ref:`append_only_mode`.
- - Use a pull-mode setup using ``ssh -R``, see :ref:`pull_backup` for more information.
- - Mount C's filesystem on another machine and then create a backup of it.
- - Do not give C filesystem-level access to S.
- See :ref:`hosting_repositories` for a detailed protection guide.
- How can I protect against a hacked backup server?
- -------------------------------------------------
- Just in case you got the impression that pull-mode backups are way more safe
- than push-mode, you also need to consider the case that your backup server S
- gets hacked. In case S has access to a lot of clients C, that might bring you
- into even bigger trouble than a hacked backup client in the previous FAQ entry.
- These are your options to protect against that:
- - Use the standard push-mode setup (see also previous FAQ entry).
- - Mount (the repo part of) S's filesystem on C.
- - Do not give S file-system level access to C.
- - Have your backup server at a well protected place (maybe not reachable from
- the internet), configure it safely, apply security updates, monitor it, ...
- How can I protect against theft, sabotage, lightning, fire, ...?
- ----------------------------------------------------------------
- In general: if your only backup medium is nearby the backupped machine and
- always connected, you can easily get into trouble: they likely share the same
- fate if something goes really wrong.
- Thus:
- - have multiple backup media
- - have media disconnected from network, power, computer
- - have media at another place
- - have a relatively recent backup on your media
- How do I report a security issue with Borg?
- -------------------------------------------
- Send a private email to the :ref:`security contact <security-contact>`
- if you think you have discovered a security issue.
- Please disclose security issues responsibly.
- Common issues
- #############
- /path/to/repo is not a valid repository. Check repo config.
- -----------------------------------------------------------
- There can be many causes of this error. E.g. you have incorrectly specified the repository path.
- You will also get this error if you try to access a repository with a key that uses the argon2 key algorithm using an old version of borg.
- We recommend upgrading to the latest stable version and trying again. We are sorry. We should have thought about forward
- compatibility and implemented a more helpful error message.
- Why does Borg extract hang after some time?
- -------------------------------------------
- When I do a ``borg extract``, after a while all activity stops, no cpu usage,
- no downloads.
- This may happen when the SSH connection is stuck on server side. You can
- configure SSH on client side to prevent this by sending keep-alive requests,
- for example in ~/.ssh/config:
- ::
- Host borg.example.com
- # Client kills connection after 3*30 seconds without server response:
- ServerAliveInterval 30
- ServerAliveCountMax 3
- You can also do the opposite and configure SSH on server side in
- /etc/ssh/sshd_config, to make the server send keep-alive requests to the client:
- ::
- # Server kills connection after 3*30 seconds without client response:
- ClientAliveInterval 30
- ClientAliveCountMax 3
- How can I deal with my very unstable SSH connection?
- ----------------------------------------------------
- If you have issues with lost connections during long-running borg commands, you
- could try to work around:
- - Make partial extracts like ``borg extract PATTERN`` to do multiple
- smaller extraction runs that complete before your connection has issues.
- - Try using ``borg mount MOUNTPOINT`` and ``rsync -avH`` from
- ``MOUNTPOINT`` to your desired extraction directory. If the connection breaks
- down, just repeat that over and over again until rsync does not find anything
- to do any more. Due to the way borg mount works, this might be less efficient
- than borg extract for bigger volumes of data.
- Why do I get "connection closed by remote" after a while?
- ---------------------------------------------------------
- When doing a backup to a remote server (using a ssh: repo URL), it sometimes
- stops after a while (some minutes, hours, ... - not immediately) with
- "connection closed by remote" error message. Why?
- That's a good question and we are trying to find a good answer in :issue:`636`.
- Why am I seeing idle borg serve processes on the repo server?
- -------------------------------------------------------------
- Maybe the ssh connection between client and server broke down and that was not
- yet noticed on the server. Try these settings:
- ::
- # /etc/ssh/sshd_config on borg repo server - kill connection to client
- # after ClientAliveCountMax * ClientAliveInterval seconds with no response
- ClientAliveInterval 20
- ClientAliveCountMax 3
- If you have multiple borg create ... ; borg create ... commands in a already
- serialized way in a single script, you need to give them ``--lock-wait N`` (with N
- being a bit more than the time the server needs to terminate broken down
- connections and release the lock).
- Can I back up my root partition (/) with Borg?
- ----------------------------------------------
- Backing up your entire root partition works just fine, but remember to
- exclude directories that make no sense to back up, such as /dev, /proc,
- /sys, /tmp and /run, and to use ``--one-file-system`` if you only want to
- back up the root partition (and not any mounted devices e.g.).
- If it crashes with a UnicodeError, what can I do?
- -------------------------------------------------
- Check if your encoding is set correctly. For most POSIX-like systems, try::
- export LANG=en_US.UTF-8 # or similar, important is correct charset
- If that does not help:
- - check for typos, check if you really used ``export``.
- - check if you have set ``LC_ALL`` - if so, try not setting it.
- - check if you generated the respective locale via ``locale-gen``.
- I can't extract non-ascii filenames by giving them on the commandline!?
- -----------------------------------------------------------------------
- This might be due to different ways to represent some characters in unicode
- or due to other non-ascii encoding issues.
- If you run into that, try this:
- - avoid the non-ascii characters on the commandline by e.g. extracting
- the parent directory (or even everything)
- - mount the repo using FUSE and use some file manager
- .. _expected_performance:
- What's the expected backup performance?
- ---------------------------------------
- Compared to simply copying files (e.g. with ``rsync``), Borg has more work to do.
- This can make creation of the first archive slower, but saves time
- and disk space on subsequent runs. Here what Borg does when you run ``borg create``:
- - Borg chunks the file (using the relatively expensive buzhash algorithm)
- - It then computes the "id" of the chunk (hmac-sha256 (slow, except
- if your CPU has sha256 acceleration) or blake2b (fast, in software))
- - Then it checks whether this chunk is already in the repo (local hashtable lookup,
- fast). If so, the processing of the chunk is completed here. Otherwise it needs to
- process the chunk:
- - Compresses (the default lz4 is super fast)
- - Encrypts and authenticates (AES-OCB, usually fast if your CPU has AES acceleration as usual
- since about 10y, or chacha20-poly1305, fast pure-software crypto)
- - Transmits to repo. If the repo is remote, this usually involves an SSH connection
- (does its own encryption / authentication).
- - Stores the chunk into a key/value store (the key is the chunk id, the value
- is the data). While doing that, it computes XXH64 of the data (repo low-level
- checksum, used by borg check --repository).
- Subsequent backups are usually very fast if most files are unchanged and only
- a few are new or modified. The high performance on unchanged files primarily depends
- only on a few factors (like FS recursion + metadata reading performance and the
- files cache working as expected) and much less on other factors.
- E.g., for this setup:
- - server grade machine (4C/8T 2013 Xeon, 64GB RAM, 2x good 7200RPM disks)
- - local zfs filesystem (mirrored) containing the backup source data
- - repository is remote (does not matter much for unchanged files)
- - backup job runs while machine is otherwise idle
- The observed performance is that Borg can process about
- **1 million unchanged files (and a few small changed ones) in 4 minutes!**
- If you are seeing much less than that in similar circumstances, read the next
- few FAQ entries below.
- .. _slow_backup:
- Why is my backup so slow?
- --------------------------
- If you feel your Borg backup is too slow somehow, here is what you can do:
- - Make sure Borg has enough RAM (depends on how big your repo is / how many
- files you have)
- - Use one of the blake2 modes for --encryption except if you positively know
- your CPU (and openssl) accelerates sha256 (then stay with hmac-sha256).
- - Don't use any expensive compression. The default is lz4 and super fast.
- Uncompressed is often slower than lz4.
- - Just wait. You can also interrupt it and start it again as often as you like,
- it will converge against a valid "completed" state. It is starting
- from the beginning each time, but it is still faster then as it does not store
- data into the repo which it already has there.
- - If you don’t need additional file attributes, you can disable them with ``--noflags``,
- ``--noacls``, ``--noxattrs``. This can lead to noticeable performance improvements
- when your backup consists of many small files.
- To see what files have changed and take more time processing, you can also add
- ``--list --filter=AME --stats`` to your ``borg create`` call to produce more log output,
- including a file list (with file status characters) and also some statistics at
- the end of the backup.
- Then you do the backup and look at the log output:
- - stats: Do you really have little changes or are there more changes than you thought?
- In the stats you can see the overall volume of changed data, which needed to be
- added to the repo. If that is a lot, that can be the reason why it is slow.
- - ``A`` status ("added") in the file list:
- If you see that often, you have a lot of new files (files that Borg did not find
- in the files cache). If you think there is something wrong with that (the file was there
- already in the previous backup), please read the FAQ entries below.
- - ``M`` status ("modified") in the file list:
- If you see that often, Borg thinks that a lot of your files might be modified
- (Borg found them in the files cache, but the metadata read from the filesystem did
- not match the metadata stored in the files cache).
- In such a case, Borg will need to process the files' contents completely, which is
- much slower than processing unmodified files (Borg does not read their contents!).
- The metadata values used in this comparison are determined by the ``--files-cache`` option
- and could be e.g. size, ctime and inode number (see the ``borg create`` docs for more
- details and potential issues).
- You can use the ``stat`` command on files to look at fs metadata manually to debug if
- there is any unexpected change triggering the ``M`` status.
- Also, the ``--debug-topic=files_cache`` option of ``borg create`` provides a lot of debug
- output helping to analyse why the files cache does not give its expected high performance.
- When borg runs inside a virtual machine, there are some more things to look at:
- Some hypervisors (e.g. kvm on proxmox) give some broadly compatible CPU type to the
- VM (usually to ease migration between VM hosts of potentially different hardware CPUs).
- It is broadly compatible because they leave away modern CPU features that could be
- not present in older or other CPUs, e.g. hardware acceleration for AES crypto, for
- sha2 hashes, for (P)CLMUL(QDQ) computations useful for crc32.
- So, basically you pay for compatibility with bad performance. If you prefer better
- performance, you should try to expose the host CPU's misc. hw acceleration features
- to the VM which runs borg.
- On Linux, check ``/proc/cpuinfo`` for the CPU flags inside the VM.
- For kvm check the docs about "Host model" and "Host passthrough".
- See also the next few FAQ entries for more details.
- .. _a_status_oddity:
- I am seeing 'A' (added) status for an unchanged file!?
- ------------------------------------------------------
- The files cache is used to determine whether Borg already
- "knows" / has backed up a file and if so, to skip the file from
- chunking. It intentionally *excludes* files that have a timestamp
- which is the same as the newest timestamp in the created archive.
- So, if you see an 'A' status for unchanged file(s), they are likely the files
- with the most recent timestamp in that archive.
- This is expected: it is to avoid data loss with files that are backed up from
- a snapshot and that are immediately changed after the snapshot (but within
- timestamp granularity time, so the timestamp would not change). Without the code that
- removes these files from the files cache, the change that happened right after
- the snapshot would not be contained in the next backup as Borg would
- think the file is unchanged.
- This does not affect deduplication, the file will be chunked, but as the chunks
- will often be the same and already stored in the repo (except in the above
- mentioned rare condition), it will just re-use them as usual and not store new
- data chunks.
- If you want to avoid unnecessary chunking, just create or touch a small or
- empty file in your backup source file set (so that one has the latest timestamp,
- not your 50GB VM disk image) and, if you do snapshots, do the snapshot after
- that.
- Since only the files cache is used in the display of files status,
- those files are reported as being added when, really, chunks are
- already used.
- By default, ctime (change time) is used for the timestamps to have a rather
- safe change detection (see also the --files-cache option).
- Furthermore, pathnames recorded in files cache are always absolute, even if you
- specify source directories with relative pathname. If relative pathnames are
- stable, but absolute are not (for example if you mount a filesystem without
- stable mount points for each backup or if you are running the backup from a
- filesystem snapshot whose name is not stable), borg will assume that files are
- different and will report them as 'added', even though no new chunks will be
- actually recorded for them. To avoid this, you could bind mount your source
- directory in a directory with the stable path.
- .. _always_chunking:
- It always chunks all my files, even unchanged ones!
- ---------------------------------------------------
- Borg maintains a files cache where it remembers the timestamp, size and
- inode of files. When Borg does a new backup and starts processing a
- file, it first looks whether the file has changed (compared to the values
- stored in the files cache). If the values are the same, the file is assumed
- unchanged and thus its contents won't get chunked (again).
- Borg can't keep an infinite history of files of course, thus entries
- in the files cache have a "maximum time to live" which is set via the
- environment variable BORG_FILES_CACHE_TTL (and defaults to 20).
- Every time you do a backup (on the same machine, using the same user), the
- cache entries' ttl values of files that were not "seen" are incremented by 1
- and if they reach BORG_FILES_CACHE_TTL, the entry is removed from the cache.
- So, for example, if you do daily backups of 26 different data sets A, B,
- C, ..., Z on one machine (using the default TTL), the files from A will be
- already forgotten when you repeat the same backups on the next day and it
- will be slow because it would chunk all the files each time. If you set
- BORG_FILES_CACHE_TTL to at least 26 (or maybe even a small multiple of that),
- it would be much faster.
- Besides using a higher BORG_FILES_CACHE_TTL (which also increases memory usage),
- there is also BORG_FILES_CACHE_SUFFIX which can be used to have separate (smaller)
- files caches for each backup set instead of the default one (big) unified files cache.
- Another possible reason is that files don't always have the same path, for
- example if you mount a filesystem without stable mount points for each backup
- or if you are running the backup from a filesystem snapshot whose name is not
- stable. If the directory where you mount a filesystem is different every time,
- Borg assumes they are different files. This is true even if you back up these
- files with relative pathnames - borg uses full pathnames in files cache regardless.
- It is possible for some filesystems, such as ``mergerfs`` or network filesystems,
- to return inconsistent inode numbers across runs, causing borg to consider them changed.
- A workaround is to set the option ``--files-cache=ctime,size`` to exclude the inode
- number comparison from the files cache check so that files with different inode
- numbers won't be treated as modified.
- Is there a way to limit bandwidth with Borg?
- --------------------------------------------
- To limit upload (i.e. :ref:`borg_create`) bandwidth, use the
- ``--remote-ratelimit`` option.
- There is no built-in way to limit *download*
- (i.e. :ref:`borg_extract`) bandwidth, but limiting download bandwidth
- can be accomplished with pipeviewer_:
- Create a wrapper script: /usr/local/bin/pv-wrapper
- ::
- #!/bin/sh
- ## -q, --quiet do not output any transfer information at all
- ## -L, --rate-limit RATE limit transfer to RATE bytes per second
- RATE=307200
- pv -q -L $RATE | "$@"
- Add BORG_RSH environment variable to use pipeviewer wrapper script with ssh.
- ::
- export BORG_RSH='/usr/local/bin/pv-wrapper ssh'
- Now Borg will be bandwidth limited. The nice thing about ``pv`` is that you can
- change rate-limit on the fly:
- ::
- pv -R $(pidof pv) -L 102400
- .. _pipeviewer: http://www.ivarch.com/programs/pv.shtml
- How can I avoid unwanted base directories getting stored into archives?
- -----------------------------------------------------------------------
- Possible use cases:
- - Another file system is mounted and you want to back it up with original paths.
- - You have created a BTRFS snapshot in a ``/.snapshots`` directory for backup.
- To achieve this, run ``borg create`` within the mountpoint/snapshot directory:
- ::
- # Example: Some file system mounted in /mnt/rootfs.
- cd /mnt/rootfs
- borg create rootfs_backup .
- Another way (without changing the directory) is to use the slashdot hack:
- ::
- borg create rootfs_backup /mnt/rootfs/./
- I am having troubles with some network/FUSE/special filesystem, why?
- --------------------------------------------------------------------
- Borg is doing nothing special in the filesystem, it only uses very
- common and compatible operations (even the locking is just "rename").
- So, if you are encountering issues like slowness, corruption or malfunction
- when using a specific filesystem, please try if you can reproduce the issues
- with a local (non-network) and proven filesystem (like ext4 on Linux).
- If you can't reproduce the issue then, you maybe have found an issue within
- the filesystem code you used (not with Borg). For this case, it is
- recommended that you talk to the developers / support of the network fs and
- maybe open an issue in their issue tracker. Do not file an issue in the
- Borg issue tracker.
- If you can reproduce the issue with the proven filesystem, please file an
- issue in the Borg issue tracker about that.
- Why does running 'borg check --repair' warn about data loss?
- ------------------------------------------------------------
- Repair usually works for recovering data in a corrupted archive. However,
- it's impossible to predict all modes of corruption. In some very rare
- instances, such as malfunctioning storage hardware, additional repo
- corruption may occur. If you can't afford to lose the repo, it's strongly
- recommended that you perform repair on a copy of the repo.
- In other words, the warning is there to emphasize that Borg:
- - Will perform automated routines that modify your backup repository
- - Might not actually fix the problem you are experiencing
- - Might, in very rare cases, further corrupt your repository
- In the case of malfunctioning hardware, such as a drive or USB hub
- corrupting data when read or written, it's best to diagnose and fix the
- cause of the initial corruption before attempting to repair the repo. If
- the corruption is caused by a one time event such as a power outage,
- running `borg check --repair` will fix most problems.
- Why isn't there more progress / ETA information displayed?
- ----------------------------------------------------------
- Some borg runs take quite a bit, so it would be nice to see a progress display,
- maybe even including a ETA (expected time of "arrival" [here rather "completion"]).
- For some functionality, this can be done: if the total amount of work is more or
- less known, we can display progress. So check if there is a ``--progress`` option.
- But sometimes, the total amount is unknown (e.g. for ``borg create`` we just do
- a single pass over the filesystem, so we do not know the total file count or data
- volume before reaching the end). Adding another pass just to determine that would
- take additional time and could be incorrect, if the filesystem is changing.
- Even if the fs does not change and we knew count and size of all files, we still
- could not compute the ``borg create`` ETA as we do not know the amount of changed
- chunks, how the bandwidth of source and destination or system performance might
- fluctuate.
- You see, trying to display ETA would be futile. The borg developers prefer to
- rather not implement progress / ETA display than doing futile attempts.
- See also: https://xkcd.com/612/
- Why am I getting 'Operation not permitted' errors when backing up on sshfs?
- ---------------------------------------------------------------------------
- By default, ``sshfs`` is not entirely POSIX-compliant when renaming files due to
- a technicality in the SFTP protocol. Fortunately, it also provides a workaround_
- to make it behave correctly::
- sshfs -o workaround=rename user@host:dir /mnt/dir
- .. _workaround: https://unix.stackexchange.com/a/123236
- How do I rename a repository?
- -----------------------------
- There is nothing special that needs to be done, you can simply rename the
- directory that corresponds to the repository. However, the next time borg
- interacts with the repository (i.e, via ``borg list``), depending on the value
- of ``BORG_RELOCATED_REPO_ACCESS_IS_OK``, borg may warn you that the repository
- has been moved. You will be given a prompt to confirm you are OK with this.
- If ``BORG_RELOCATED_REPO_ACCESS_IS_OK`` is unset, borg will interactively ask for
- each repository whether it's OK.
- It may be useful to set ``BORG_RELOCATED_REPO_ACCESS_IS_OK=yes`` to avoid the
- prompts when renaming multiple repositories or in a non-interactive context
- such as a script. See :doc:`deployment` for an example.
- My backup disk is full, what can I do?
- --------------------------------------
- Borg cannot work if you really have zero free space on the backup disk, so the
- first thing you must do is deleting some files to regain free disk space. See
- :ref:`about_free_space` for further details.
- Some Borg commands that do not change the repository might work under disk-full
- conditions, but generally this should be avoided. If your backup disk is already
- full when Borg starts a write command like `borg create`, it will abort
- immediately and the repository will stay as-is.
- Miscellaneous
- #############
- macOS: borg mounts not shown in Finder's side bar
- -------------------------------------------------
- https://github.com/osxfuse/osxfuse/wiki/Mount-options#local
- Read the above first and use this on your own risk::
- borg mount -olocal REPO MOUNTPOINT
- Requirements for the borg single-file binary, esp. (g)libc?
- -----------------------------------------------------------
- We try to build the binary on old, but still supported systems - to keep the
- minimum requirement for the (g)libc low. The (g)libc can't be bundled into
- the binary as it needs to fit your kernel and OS, but Python and all other
- required libraries will be bundled into the binary.
- If your system fulfills the minimum (g)libc requirement (see the README that
- is released with the binary), there should be no problem. If you are slightly
- below the required version, maybe just try. Due to the dynamic loading (or not
- loading) of some shared libraries, it might still work depending on what
- libraries are actually loaded and used.
- In the borg git repository, there is scripts/glibc_check.py that can determine
- (based on the symbols' versions they want to link to) whether a set of given
- (Linux) binaries works with a given glibc version.
|