123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192 |
- from binascii import hexlify
- import os
- import shutil
- import time
- from .helpers import get_keys_dir, get_cache_dir
- from .locking import UpgradableLock
- from .repository import Repository, MAGIC
- from .key import KeyfileKey, KeyfileNotFoundError
- ATTIC_MAGIC = b'ATTICSEG'
- class AtticRepositoryConverter(Repository):
- def convert(self, dryrun=True):
- """convert an attic repository to a borg repository
- those are the files that need to be converted here, from most
- important to least important: segments, key files, and various
- caches, the latter being optional, as they will be rebuilt if
- missing."""
- print("reading segments from attic repository using borg")
- # we need to open it to load the configuration and other fields
- self.open(self.path, exclusive=False)
- segments = [ filename for i, filename in self.io.segment_iterator() ]
- try:
- keyfile = self.find_attic_keyfile()
- except KeyfileNotFoundError:
- print("no key file found for repository")
- else:
- self.convert_keyfiles(keyfile, dryrun)
- self.close()
- # partial open: just hold on to the lock
- self.lock = UpgradableLock(os.path.join(self.path, 'lock'),
- exclusive=True).acquire()
- try:
- self.convert_segments(segments, dryrun)
- self.convert_cache(dryrun)
- finally:
- self.lock.release()
- self.lock = None
- @staticmethod
- def convert_segments(segments, dryrun):
- """convert repository segments from attic to borg
- replacement pattern is `s/ATTICSEG/BORG_SEG/` in files in
- `$ATTIC_REPO/data/**`.
- luckily the magic string length didn't change so we can just
- replace the 8 first bytes of all regular files in there."""
- print("converting %d segments..." % len(segments))
- i = 0
- for filename in segments:
- i += 1
- print("\rconverting segment %d/%d in place, %.2f%% done (%s)"
- % (i, len(segments), float(i)/len(segments), filename), end='')
- if dryrun:
- time.sleep(0.001)
- else:
- AtticRepositoryConverter.header_replace(filename, ATTIC_MAGIC, MAGIC)
- print()
- @staticmethod
- def header_replace(filename, old_magic, new_magic):
- print("changing header on %s" % filename)
- with open(filename, 'r+b') as segment:
- segment.seek(0)
- # only write if necessary
- if (segment.read(len(old_magic)) == old_magic):
- segment.seek(0)
- segment.write(new_magic)
- def find_attic_keyfile(self):
- """find the attic keyfiles
- the keyfiles are loaded by `KeyfileKey.find_key_file()`. that
- finds the keys with the right identifier for the repo.
- this is expected to look into $HOME/.attic/keys or
- $ATTIC_KEYS_DIR for key files matching the given Borg
- repository.
- it is expected to raise an exception (KeyfileNotFoundError) if
- no key is found. whether that exception is from Borg or Attic
- is unclear.
- this is split in a separate function in case we want to use
- the attic code here directly, instead of our local
- implementation."""
- return AtticKeyfileKey.find_key_file(self)
- @staticmethod
- def convert_keyfiles(keyfile, dryrun):
- """convert key files from attic to borg
- replacement pattern is `s/ATTIC KEY/BORG_KEY/` in
- `get_keys_dir()`, that is `$ATTIC_KEYS_DIR` or
- `$HOME/.attic/keys`, and moved to `$BORG_KEYS_DIR` or
- `$HOME/.borg/keys`.
- no need to decrypt to convert. we need to rewrite the whole
- key file because magic string length changed, but that's not a
- problem because the keyfiles are small (compared to, say,
- all the segments)."""
- print("converting keyfile %s" % keyfile)
- with open(keyfile, 'r') as f:
- data = f.read()
- data = data.replace(AtticKeyfileKey.FILE_ID, KeyfileKey.FILE_ID, 1)
- keyfile = os.path.join(get_keys_dir(), os.path.basename(keyfile))
- print("writing borg keyfile to %s" % keyfile)
- if not dryrun:
- with open(keyfile, 'w') as f:
- f.write(data)
- def convert_cache(self, dryrun):
- """convert caches from attic to borg
- those are all hash indexes, so we need to
- `s/ATTICIDX/BORG_IDX/` in a few locations:
-
- * the repository index (in `$ATTIC_REPO/index.%d`, where `%d`
- is the `Repository.get_index_transaction_id()`), which we
- should probably update, with a lock, see
- `Repository.open()`, which i'm not sure we should use
- because it may write data on `Repository.close()`...
- * the `files` and `chunks` cache (in `$ATTIC_CACHE_DIR` or
- `$HOME/.cache/attic/<repoid>/`), which we could just drop,
- but if we'd want to convert, we could open it with the
- `Cache.open()`, edit in place and then `Cache.close()` to
- make sure we have locking right
- """
- caches = []
- transaction_id = self.get_index_transaction_id()
- if transaction_id is None:
- print('no index file found for repository %s' % self.path)
- else:
- caches += [os.path.join(self.path, 'index.%d' % transaction_id).encode('utf-8')]
- # copy of attic's get_cache_dir()
- attic_cache_dir = os.environ.get('ATTIC_CACHE_DIR',
- os.path.join(os.path.expanduser('~'), '.cache', 'attic'))
- # XXX: untested, because generating cache files is a PITA, see
- # Archiver.do_create() for proof
- for cache in [ 'files', 'chunks' ]:
- attic_cache = os.path.join(attic_cache_dir, hexlify(self.id).decode('ascii'), cache)
- if os.path.exists(attic_cache):
- borg_cache = os.path.join(get_cache_dir(), hexlify(self.id).decode('ascii'), cache)
- shutil.copy(attic_cache, borg_cache)
- caches += [borg_cache]
- for cache in caches:
- print("converting cache %s" % cache)
- AtticRepositoryConverter.header_replace(cache, b'ATTICIDX', b'BORG_IDX')
- class AtticKeyfileKey(KeyfileKey):
- """backwards compatible Attic key file parser"""
- FILE_ID = 'ATTIC KEY'
- # verbatim copy from attic
- @staticmethod
- def get_keys_dir():
- """Determine where to repository keys and cache"""
- return os.environ.get('ATTIC_KEYS_DIR',
- os.path.join(os.path.expanduser('~'), '.attic', 'keys'))
- @classmethod
- def find_key_file(cls, repository):
- """copy of attic's `find_key_file`_
- this has two small modifications:
- 1. it uses the above `get_keys_dir`_ instead of the global one,
- assumed to be borg's
- 2. it uses `repository.path`_ instead of
- `repository._location.canonical_path`_ because we can't
- assume the repository has been opened by the archiver yet
- """
- get_keys_dir = cls.get_keys_dir
- id = hexlify(repository.id).decode('ascii')
- keys_dir = get_keys_dir()
- for name in os.listdir(keys_dir):
- filename = os.path.join(keys_dir, name)
- with open(filename, 'r') as fd:
- line = fd.readline().strip()
- if line and line.startswith(cls.FILE_ID) and line[10:] == id:
- return filename
- raise KeyfileNotFoundError(repository.path, get_keys_dir())
|