| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447 | 
							- from __future__ import with_statement
 
- from ConfigParser import RawConfigParser
 
- import fcntl
 
- import os
 
- import re
 
- import shutil
 
- import struct
 
- import tempfile
 
- import unittest
 
- from zlib import crc32
 
- from .hashindex import NSIndex
 
- from .helpers import IntegrityError, deferrable, read_msgpack, write_msgpack
 
- from .lrucache import LRUCache
 
- MAX_OBJECT_SIZE = 20 * 1024 * 1024
 
- TAG_PUT = 0
 
- TAG_DELETE = 1
 
- TAG_COMMIT = 2
 
- class Store(object):
 
-     """Filesystem based transactional key value store
 
-     On disk layout:
 
-     dir/README
 
-     dir/config
 
-     dir/data/<X / SEGMENTS_PER_DIR>/<X>
 
-     dir/index.X
 
-     dir/hints.X
 
-     """
 
-     DEFAULT_MAX_SEGMENT_SIZE = 5 * 1024 * 1024
 
-     DEFAULT_SEGMENTS_PER_DIR = 10000
 
-     class DoesNotExist(KeyError):
 
-         """Requested key does not exist"""
 
-     def __init__(self, path, create=False):
 
-         if create:
 
-             self.create(path)
 
-         self.open(path)
 
-     def create(self, path):
 
-         """Create a new empty store at `path`
 
-         """
 
-         if os.path.exists(path) and (not os.path.isdir(path) or os.listdir(path)):
 
-             raise Exception('Path "%s" already exists' % path)
 
-         if not os.path.exists(path):
 
-             os.mkdir(path)
 
-         with open(os.path.join(path, 'README'), 'wb') as fd:
 
-             fd.write('This is a DARC store')
 
-         os.mkdir(os.path.join(path, 'data'))
 
-         config = RawConfigParser()
 
-         config.add_section('store')
 
-         config.set('store', 'version', '1')
 
-         config.set('store', 'segments_per_dir', self.DEFAULT_SEGMENTS_PER_DIR)
 
-         config.set('store', 'max_segment_size', self.DEFAULT_MAX_SEGMENT_SIZE)
 
-         config.set('store', 'id', os.urandom(32).encode('hex'))
 
-         with open(os.path.join(path, 'config'), 'w') as fd:
 
-             config.write(fd)
 
-     def open(self, path):
 
-         self.head = None
 
-         self.path = path
 
-         if not os.path.isdir(path):
 
-             raise Exception('%s Does not look like a darc store' % path)
 
-         self.lock_fd = open(os.path.join(path, 'README'), 'r+')
 
-         fcntl.flock(self.lock_fd, fcntl.LOCK_EX)
 
-         self.config = RawConfigParser()
 
-         self.config.read(os.path.join(self.path, 'config'))
 
-         if self.config.getint('store', 'version') != 1:
 
-             raise Exception('%s Does not look like a darc store')
 
-         self.max_segment_size = self.config.getint('store', 'max_segment_size')
 
-         self.segments_per_dir = self.config.getint('store', 'segments_per_dir')
 
-         self.id = self.config.get('store', 'id').decode('hex')
 
-         self.rollback()
 
-     def close(self):
 
-         self.lock_fd.close()
 
-     def commit(self, rollback=True):
 
-         """Commit transaction
 
-         """
 
-         self.io.write_commit()
 
-         self.compact_segments()
 
-         self.write_index()
 
-         self.rollback()
 
-     def _available_indices(self, reverse=False):
 
-         names = [int(name[6:]) for name in os.listdir(self.path) if re.match('index\.\d+', name)]
 
-         names.sort(reverse=reverse)
 
-         return names
 
-     def open_index(self, head, read_only=False):
 
-         if head is None:
 
-             self.index = NSIndex.create(os.path.join(self.path, 'index.tmp'))
 
-             self.segments = {}
 
-             self.compact = set()
 
-         else:
 
-             if read_only:
 
-                 self.index = NSIndex(os.path.join(self.path, 'index.%d') % head)
 
-             else:
 
-                 shutil.copy(os.path.join(self.path, 'index.%d' % head),
 
-                             os.path.join(self.path, 'index.tmp'))
 
-                 self.index = NSIndex(os.path.join(self.path, 'index.tmp'))
 
-             hints = read_msgpack(os.path.join(self.path, 'hints.%d' % head))
 
-             if hints['version'] != 1:
 
-                 raise ValueError('Unknown hints file version: %d' % hints['version'])
 
-             self.segments = hints['segments']
 
-             self.compact = set(hints['compact'])
 
-     def write_index(self):
 
-         hints = {'version': 1,
 
-                  'segments': self.segments,
 
-                  'compact': list(self.compact)}
 
-         write_msgpack(os.path.join(self.path, 'hints.%d' % self.io.head), hints)
 
-         self.index.flush()
 
-         os.rename(os.path.join(self.path, 'index.tmp'),
 
-                   os.path.join(self.path, 'index.%d' % self.io.head))
 
-         # Remove old indices
 
-         current = '.%d' % self.io.head
 
-         for name in os.listdir(self.path):
 
-             if not name.startswith('index.') and not name.startswith('hints.'):
 
-                 continue
 
-             if name.endswith(current):
 
-                 continue
 
-             os.unlink(os.path.join(self.path, name))
 
-     def compact_segments(self):
 
-         """Compact sparse segments by copying data into new segments
 
-         """
 
-         if not self.compact:
 
-             return
 
-         def lookup(tag, key):
 
-             return tag == TAG_PUT and self.index.get(key, (-1, -1))[0] == segment
 
-         segments = self.segments
 
-         for segment in self.compact:
 
-             if segments[segment] > 0:
 
-                 for tag, key, data in self.io.iter_objects(segment, lookup, include_data=True):
 
-                     new_segment, offset = self.io.write_put(key, data)
 
-                     self.index[key] = new_segment, offset
 
-                     segments.setdefault(new_segment, 0)
 
-                     segments[new_segment] += 1
 
-                     segments[segment] -= 1
 
-                 assert segments[segment] == 0
 
-         self.io.write_commit()
 
-         for segment in self.compact:
 
-             assert self.segments.pop(segment) == 0
 
-             self.io.delete_segment(segment)
 
-         self.compact = set()
 
-     def recover(self, path):
 
-         """Recover missing index by replaying logs"""
 
-         start = None
 
-         available = self._available_indices()
 
-         if available:
 
-             start = available[-1]
 
-         self.open_index(start)
 
-         for segment, filename in self.io._segment_names():
 
-             if start is not None and segment <= start:
 
-                 continue
 
-             self.segments[segment] = 0
 
-             for tag, key, offset in self.io.iter_objects(segment):
 
-                 if tag == TAG_PUT:
 
-                     try:
 
-                         s, _ = self.index[key]
 
-                         self.compact.add(s)
 
-                         self.segments[s] -= 1
 
-                     except KeyError:
 
-                         pass
 
-                     self.index[key] = segment, offset
 
-                     self.segments[segment] += 1
 
-                 elif tag == TAG_DELETE:
 
-                     try:
 
-                         s, _ = self.index.pop(key)
 
-                         self.segments[s] -= 1
 
-                         self.compact.add(s)
 
-                         self.compact.add(segment)
 
-                     except KeyError:
 
-                         pass
 
-             if self.segments[segment] == 0:
 
-                 self.compact.add(segment)
 
-         if self.io.head is not None:
 
-             self.write_index()
 
-     def rollback(self):
 
-         """
 
-         """
 
-         self._active_txn = False
 
-         self.io = LoggedIO(self.path, self.max_segment_size, self.segments_per_dir)
 
-         if self.io.head is not None and not os.path.exists(os.path.join(self.path, 'index.%d' % self.io.head)):
 
-             self.recover(self.path)
 
-         self.open_index(self.io.head, read_only=True)
 
-     @deferrable
 
-     def get(self, id):
 
-         try:
 
-             segment, offset = self.index[id]
 
-             return self.io.read(segment, offset, id)
 
-         except KeyError:
 
-             raise self.DoesNotExist
 
-     @deferrable
 
-     def put(self, id, data):
 
-         if not self._active_txn:
 
-             self._active_txn = True
 
-             self.open_index(self.io.head)
 
-         try:
 
-             segment, _ = self.index[id]
 
-             self.segments[segment] -= 1
 
-             self.compact.add(segment)
 
-             self.compact.add(self.io.write_delete(id))
 
-         except KeyError:
 
-             pass
 
-         segment, offset = self.io.write_put(id, data)
 
-         self.segments.setdefault(segment, 0)
 
-         self.segments[segment] += 1
 
-         self.index[id] = segment, offset
 
-     @deferrable
 
-     def delete(self, id):
 
-         if not self._active_txn:
 
-             self._active_txn = True
 
-             self.open_index(self.io.head)
 
-         try:
 
-             segment, offset = self.index.pop(id)
 
-             self.segments[segment] -= 1
 
-             self.compact.add(segment)
 
-             self.compact.add(self.io.write_delete(id))
 
-         except KeyError:
 
-             raise self.DoesNotExist
 
-     def flush_rpc(self, *args):
 
-         pass
 
-     def add_callback(self, cb, data):
 
-         cb(None, None, data)
 
- class LoggedIO(object):
 
-     header_fmt = struct.Struct('<IIB')
 
-     assert header_fmt.size == 9
 
-     put_header_fmt = struct.Struct('<IIB32s')
 
-     assert put_header_fmt.size == 41
 
-     header_no_crc_fmt = struct.Struct('<IB')
 
-     assert header_no_crc_fmt.size == 5
 
-     crc_fmt = struct.Struct('<I')
 
-     assert crc_fmt.size == 4
 
-     _commit = header_no_crc_fmt.pack(9, TAG_COMMIT)
 
-     COMMIT = crc_fmt.pack(crc32(_commit)) + _commit
 
-     def __init__(self, path, limit, segments_per_dir, capacity=100):
 
-         self.path = path
 
-         self.fds = LRUCache(capacity)
 
-         self.segment = None
 
-         self.limit = limit
 
-         self.segments_per_dir = segments_per_dir
 
-         self.offset = 0
 
-         self._write_fd = None
 
-         self.head = None
 
-         self.cleanup()
 
-     def close(self):
 
-         for segment in self.fds.keys():
 
-             self.fds.pop(segment).close()
 
-         self.close_segment()
 
-         self.fds = None  # Just to make sure we're disabled
 
-     def _segment_names(self, reverse=False):
 
-         for dirpath, dirs, filenames in os.walk(os.path.join(self.path, 'data')):
 
-             dirs.sort(lambda a, b: cmp(int(a), int(b)), reverse=reverse)
 
-             filenames.sort(lambda a, b: cmp(int(a), int(b)), reverse=reverse)
 
-             for filename in filenames:
 
-                 yield int(filename), os.path.join(dirpath, filename)
 
-     def cleanup(self):
 
-         """Delete segment files left by aborted transactions
 
-         """
 
-         self.head = None
 
-         self.segment = 0
 
-         for segment, filename in self._segment_names(reverse=True):
 
-             if self.is_complete_segment(filename):
 
-                 self.head = segment
 
-                 self.segment = self.head + 1
 
-                 return
 
-             else:
 
-                 os.unlink(filename)
 
-     def is_complete_segment(self, filename):
 
-         with open(filename, 'rb') as fd:
 
-             fd.seek(-self.header_fmt.size, 2)
 
-             return fd.read(self.header_fmt.size) == self.COMMIT
 
-     def segment_filename(self, segment):
 
-         return os.path.join(self.path, 'data', str(segment / self.segments_per_dir), str(segment))
 
-     def get_write_fd(self):
 
-         if self.offset and self.offset > self.limit:
 
-             self.close_segment()
 
-         if not self._write_fd:
 
-             if self.segment % self.segments_per_dir == 0:
 
-                 dirname = os.path.join(self.path, 'data', str(self.segment / self.segments_per_dir))
 
-                 if not os.path.exists(dirname):
 
-                     os.mkdir(dirname)
 
-             self._write_fd = open(self.segment_filename(self.segment), 'ab')
 
-             self._write_fd.write('DSEGMENT')
 
-             self.offset = 8
 
-         return self._write_fd
 
-     def get_fd(self, segment):
 
-         try:
 
-             return self.fds[segment]
 
-         except KeyError:
 
-             fd = open(self.segment_filename(segment), 'rb')
 
-             self.fds[segment] = fd
 
-             return fd
 
-     def delete_segment(self, segment):
 
-         try:
 
-             os.unlink(self.segment_filename(segment))
 
-         except OSError:
 
-             pass
 
-     def iter_objects(self, segment, lookup=None, include_data=False):
 
-         fd = self.get_fd(segment)
 
-         fd.seek(0)
 
-         if fd.read(8) != 'DSEGMENT':
 
-             raise IntegrityError('Invalid segment header')
 
-         offset = 8
 
-         header = fd.read(self.header_fmt.size)
 
-         while header:
 
-             crc, size, tag = self.header_fmt.unpack(header)
 
-             if size > MAX_OBJECT_SIZE:
 
-                 raise IntegrityError('Invalid segment object size')
 
-             rest = fd.read(size - self.header_fmt.size)
 
-             if crc32(rest, crc32(buffer(header, 4))) & 0xffffffff != crc:
 
-                 raise IntegrityError('Segment checksum mismatch')
 
-             if tag not in (TAG_PUT, TAG_DELETE, TAG_COMMIT):
 
-                 raise IntegrityError('Invalid segment entry header')
 
-             key = None
 
-             if tag in (TAG_PUT, TAG_DELETE):
 
-                 key = rest[:32]
 
-             if not lookup or lookup(tag, key):
 
-                 if include_data:
 
-                     yield tag, key, rest[32:]
 
-                 else:
 
-                     yield tag, key, offset
 
-             offset += size
 
-             header = fd.read(self.header_fmt.size)
 
-     def read(self, segment, offset, id):
 
-         if segment == self.segment:
 
-             self._write_fd.flush()
 
-         fd = self.get_fd(segment)
 
-         fd.seek(offset)
 
-         header = fd.read(self.put_header_fmt.size)
 
-         crc, size, tag, key = self.put_header_fmt.unpack(header)
 
-         if size > MAX_OBJECT_SIZE:
 
-             raise IntegrityError('Invalid segment object size')
 
-         data = fd.read(size - self.put_header_fmt.size)
 
-         if crc32(data, crc32(buffer(header, 4))) & 0xffffffff != crc:
 
-             raise IntegrityError('Segment checksum mismatch')
 
-         if tag != TAG_PUT or id != key:
 
-             raise IntegrityError('Invalid segment entry header')
 
-         return data
 
-     def write_put(self, id, data):
 
-         size = len(data) + self.put_header_fmt.size
 
-         fd = self.get_write_fd()
 
-         offset = self.offset
 
-         header = self.header_no_crc_fmt.pack(size, TAG_PUT)
 
-         crc = self.crc_fmt.pack(crc32(data, crc32(id, crc32(header))) & 0xffffffff)
 
-         fd.write(''.join((crc, header, id, data)))
 
-         self.offset += size
 
-         return self.segment, offset
 
-     def write_delete(self, id):
 
-         fd = self.get_write_fd()
 
-         header = self.header_no_crc_fmt.pack(self.put_header_fmt.size, TAG_DELETE)
 
-         crc = self.crc_fmt.pack(crc32(id, crc32(header)) & 0xffffffff)
 
-         fd.write(''.join((crc, header, id)))
 
-         self.offset += self.put_header_fmt.size
 
-         return self.segment
 
-     def write_commit(self):
 
-         fd = self.get_write_fd()
 
-         header = self.header_no_crc_fmt.pack(self.header_fmt.size, TAG_COMMIT)
 
-         crc = self.crc_fmt.pack(crc32(header) & 0xffffffff)
 
-         fd.write(''.join((crc, header)))
 
-         self.head = self.segment
 
-         self.close_segment()
 
-     def close_segment(self):
 
-         if self._write_fd:
 
-             self.segment += 1
 
-             self.offset = 0
 
-             os.fsync(self._write_fd)
 
-             self._write_fd.close()
 
-             self._write_fd = None
 
- class StoreTestCase(unittest.TestCase):
 
-     def setUp(self):
 
-         self.tmppath = tempfile.mkdtemp()
 
-         self.store = Store(os.path.join(self.tmppath, 'store'), create=True)
 
-     def tearDown(self):
 
-         shutil.rmtree(self.tmppath)
 
-     def test1(self):
 
-         for x in range(100):
 
-             self.store.put('%-32d' % x, 'SOMEDATA')
 
-         key50 = '%-32d' % 50
 
-         self.assertEqual(self.store.get(key50), 'SOMEDATA')
 
-         self.store.delete(key50)
 
-         self.assertRaises(self.store.DoesNotExist, lambda: self.store.get(key50))
 
-         self.store.commit()
 
-         self.store.close()
 
-         store2 = Store(os.path.join(self.tmppath, 'store'))
 
-         self.assertRaises(store2.DoesNotExist, lambda: store2.get(key50))
 
-         for x in range(100):
 
-             if x == 50:
 
-                 continue
 
-             self.assertEqual(self.store.get('%-32d' % x), 'SOMEDATA')
 
-     def test2(self):
 
-         """Test multiple sequential transactions
 
-         """
 
-         self.store.put('00000000000000000000000000000000', 'foo')
 
-         self.store.put('00000000000000000000000000000001', 'foo')
 
-         self.store.commit()
 
-         self.store.delete('00000000000000000000000000000000')
 
-         self.store.put('00000000000000000000000000000001', 'bar')
 
-         self.store.commit()
 
-         self.assertEqual(self.store.get('00000000000000000000000000000001'), 'bar')
 
- def suite():
 
-     return unittest.TestLoader().loadTestsFromTestCase(StoreTestCase)
 
- if __name__ == '__main__':
 
-     unittest.main()
 
 
  |