borg/borg/fuse.py

300 lines
11 KiB
Python
Raw Normal View History

2013-07-21 22:41:06 +00:00
from collections import defaultdict
import errno
import io
2013-07-21 22:41:06 +00:00
import llfuse
import os
2013-07-23 11:07:48 +00:00
import stat
import tempfile
2013-07-21 22:41:06 +00:00
import time
from distutils.version import LooseVersion
import msgpack
2015-10-08 21:03:35 +00:00
from .archive import Archive
from .helpers import daemonize, bigint_to_int, remove_surrogates
from .logger import create_logger
logger = create_logger()
# Does this version of llfuse support ns precision?
have_fuse_xtime_ns = hasattr(llfuse.EntryAttributes, 'st_mtime_ns')
fuse_version = LooseVersion(getattr(llfuse, '__version__', '0.1'))
if fuse_version >= '0.42':
def fuse_main():
return llfuse.main(workers=1)
else:
def fuse_main():
llfuse.main(single=True)
return None
2013-07-21 22:41:06 +00:00
2016-02-17 00:05:04 +00:00
class ItemCache:
def __init__(self):
self.fd = tempfile.TemporaryFile(prefix='borg-tmp')
self.offset = 1000000
def add(self, item):
pos = self.fd.seek(0, io.SEEK_END)
self.fd.write(msgpack.packb(item))
return pos + self.offset
def get(self, inode):
self.fd.seek(inode - self.offset, io.SEEK_SET)
return next(msgpack.Unpacker(self.fd, read_size=1024))
class FuseOperations(llfuse.Operations):
"""Export archive as a fuse filesystem
2013-07-21 22:41:06 +00:00
"""
allow_damaged_files = True
def __init__(self, key, repository, manifest, archive, cached_repo):
super().__init__()
2013-07-23 11:40:54 +00:00
self._inode_count = 0
2013-07-21 22:41:06 +00:00
self.key = key
self.repository = cached_repo
2013-07-21 22:41:06 +00:00
self.items = {}
2013-07-23 11:40:54 +00:00
self.parent = {}
self.contents = defaultdict(dict)
self.default_dir = {b'mode': 0o40755, b'mtime': int(time.time() * 1e9), b'uid': os.getuid(), b'gid': os.getgid()}
2014-03-27 21:43:06 +00:00
self.pending_archives = {}
self.accounted_chunks = {}
self.cache = ItemCache()
if archive:
self.process_archive(archive)
else:
# Create root inode
2014-03-27 21:43:06 +00:00
self.parent[1] = self.allocate_inode()
self.items[1] = self.default_dir
for archive_name in manifest.archives:
# Create archive placeholder inode
2014-03-27 21:43:06 +00:00
archive_inode = self.allocate_inode()
self.items[archive_inode] = self.default_dir
self.parent[archive_inode] = 1
self.contents[1][os.fsencode(archive_name)] = archive_inode
self.pending_archives[archive_inode] = Archive(repository, key, manifest, archive_name)
def process_archive(self, archive, prefix=[]):
2014-09-07 12:54:18 +00:00
"""Build fuse inode hierarchy from archive metadata
"""
unpacker = msgpack.Unpacker()
for key, chunk in zip(archive.metadata[b'items'], self.repository.get_many(archive.metadata[b'items'])):
data = self.key.decrypt(key, chunk)
unpacker.feed(data)
for item in unpacker:
try:
# This can happen if an archive was created with a command line like
# $ borg create ... dir1/file dir1
# In this case the code below will have created a default_dir inode for dir1 already.
inode = self._find_inode(item[b'path'], prefix)
except KeyError:
pass
else:
self.items[inode] = item
continue
segments = prefix + os.fsencode(os.path.normpath(item[b'path'])).split(b'/')
del item[b'path']
num_segments = len(segments)
parent = 1
for i, segment in enumerate(segments, 1):
# Insert a default root inode if needed
if self._inode_count == 0 and segment:
archive_inode = self.allocate_inode()
self.items[archive_inode] = self.default_dir
self.parent[archive_inode] = parent
# Leaf segment?
if i == num_segments:
if b'source' in item and stat.S_ISREG(item[b'mode']):
inode = self._find_inode(item[b'source'], prefix)
item = self.cache.get(inode)
item[b'nlink'] = item.get(b'nlink', 1) + 1
self.items[inode] = item
else:
inode = self.cache.add(item)
self.parent[inode] = parent
if segment:
self.contents[parent][segment] = inode
elif segment in self.contents[parent]:
parent = self.contents[parent][segment]
2013-07-24 11:05:47 +00:00
else:
inode = self.allocate_inode()
self.items[inode] = self.default_dir
self.parent[inode] = parent
if segment:
self.contents[parent][segment] = inode
parent = inode
2013-07-23 11:40:54 +00:00
def allocate_inode(self):
self._inode_count += 1
return self._inode_count
2013-07-21 22:41:06 +00:00
def statfs(self, ctx=None):
2013-07-27 12:31:28 +00:00
stat_ = llfuse.StatvfsData()
stat_.f_bsize = 512
stat_.f_frsize = 512
stat_.f_blocks = 0
stat_.f_bfree = 0
stat_.f_bavail = 0
stat_.f_files = 0
stat_.f_ffree = 0
stat_.f_favail = 0
return stat_
def get_item(self, inode):
try:
return self.items[inode]
except KeyError:
return self.cache.get(inode)
def _find_inode(self, path, prefix=[]):
segments = prefix + os.fsencode(os.path.normpath(path)).split(b'/')
2013-07-23 11:07:48 +00:00
inode = 1
for segment in segments:
2013-07-23 11:40:54 +00:00
inode = self.contents[inode][segment]
return inode
2013-07-21 22:41:06 +00:00
def getattr(self, inode, ctx=None):
item = self.get_item(inode)
2013-07-21 22:41:06 +00:00
size = 0
dsize = 0
2013-07-21 22:41:06 +00:00
try:
for key, chunksize, _ in item[b'chunks']:
size += chunksize
if self.accounted_chunks.get(key, inode) == inode:
self.accounted_chunks[key] = inode
dsize += chunksize
2013-07-21 22:41:06 +00:00
except KeyError:
pass
entry = llfuse.EntryAttributes()
2013-07-23 11:40:54 +00:00
entry.st_ino = inode
2013-07-21 22:41:06 +00:00
entry.generation = 0
entry.entry_timeout = 300
entry.attr_timeout = 300
entry.st_mode = item[b'mode']
2013-07-23 11:40:54 +00:00
entry.st_nlink = item.get(b'nlink', 1)
2013-07-21 22:41:06 +00:00
entry.st_uid = item[b'uid']
2013-07-25 20:54:19 +00:00
entry.st_gid = item[b'gid']
2013-07-21 22:41:06 +00:00
entry.st_rdev = item.get(b'rdev', 0)
entry.st_size = size
entry.st_blksize = 512
entry.st_blocks = dsize / 512
# note: older archives only have mtime (not atime nor ctime)
if have_fuse_xtime_ns:
entry.st_mtime_ns = bigint_to_int(item[b'mtime'])
if b'atime' in item:
entry.st_atime_ns = bigint_to_int(item[b'atime'])
else:
entry.st_atime_ns = bigint_to_int(item[b'mtime'])
if b'ctime' in item:
entry.st_ctime_ns = bigint_to_int(item[b'ctime'])
else:
entry.st_ctime_ns = bigint_to_int(item[b'mtime'])
else:
2016-03-31 21:31:53 +00:00
entry.st_mtime = bigint_to_int(item[b'mtime']) / 1e9
if b'atime' in item:
2016-03-31 21:31:53 +00:00
entry.st_atime = bigint_to_int(item[b'atime']) / 1e9
else:
2016-03-31 21:31:53 +00:00
entry.st_atime = bigint_to_int(item[b'mtime']) / 1e9
if b'ctime' in item:
2016-03-31 21:31:53 +00:00
entry.st_ctime = bigint_to_int(item[b'ctime']) / 1e9
else:
2016-03-31 21:31:53 +00:00
entry.st_ctime = bigint_to_int(item[b'mtime']) / 1e9
2013-07-21 22:41:06 +00:00
return entry
def listxattr(self, inode, ctx=None):
item = self.get_item(inode)
2013-07-27 20:33:43 +00:00
return item.get(b'xattrs', {}).keys()
2013-07-23 08:44:29 +00:00
def getxattr(self, inode, name, ctx=None):
item = self.get_item(inode)
2013-07-23 08:44:29 +00:00
try:
return item.get(b'xattrs', {})[name]
except KeyError:
raise llfuse.FUSEError(llfuse.ENOATTR) from None
2013-07-23 08:44:29 +00:00
2014-03-27 21:43:06 +00:00
def _load_pending_archive(self, inode):
# Check if this is an archive we need to load
archive = self.pending_archives.pop(inode, None)
if archive:
self.process_archive(archive, [os.fsencode(archive.name)])
def lookup(self, parent_inode, name, ctx=None):
2014-03-27 21:43:06 +00:00
self._load_pending_archive(parent_inode)
2013-07-21 22:41:06 +00:00
if name == b'.':
inode = parent_inode
elif name == b'..':
2013-07-23 11:40:54 +00:00
inode = self.parent[parent_inode]
2013-07-21 22:41:06 +00:00
else:
2013-07-23 11:40:54 +00:00
inode = self.contents[parent_inode].get(name)
2013-07-21 22:41:06 +00:00
if not inode:
raise llfuse.FUSEError(errno.ENOENT)
return self.getattr(inode)
def open(self, inode, flags, ctx=None):
if not self.allow_damaged_files:
item = self.get_item(inode)
if b'chunks_healthy' in item:
# Processed archive items don't carry the path anymore; for converting the inode
# to the path we'd either have to store the inverse of the current structure,
# or search the entire archive. So we just don't print it. It's easy to correlate anyway.
logger.warning('File has damaged (all-zero) chunks. Try running borg check --repair. '
'Mount with allow_damaged_files to read damaged files.')
raise llfuse.FUSEError(errno.EIO)
2013-07-21 22:41:06 +00:00
return inode
def opendir(self, inode, ctx=None):
2014-03-27 21:43:06 +00:00
self._load_pending_archive(inode)
2013-07-21 22:41:06 +00:00
return inode
def read(self, fh, offset, size):
parts = []
item = self.get_item(fh)
2013-07-21 22:41:06 +00:00
for id, s, csize in item[b'chunks']:
if s < offset:
offset -= s
continue
n = min(size, s - offset)
chunk = self.key.decrypt(id, self.repository.get(id))
2016-01-30 20:32:45 +00:00
parts.append(chunk[offset:offset + n])
2013-07-21 22:41:06 +00:00
offset = 0
size -= n
if not size:
break
return b''.join(parts)
def readdir(self, fh, off):
2013-07-23 11:40:54 +00:00
entries = [(b'.', fh), (b'..', self.parent[fh])]
entries.extend(self.contents[fh].items())
2013-07-21 22:41:06 +00:00
for i, (name, inode) in enumerate(entries[off:], off):
yield name, self.getattr(inode), i + 1
def readlink(self, inode, ctx=None):
item = self.get_item(inode)
return os.fsencode(item[b'source'])
2013-07-23 11:40:54 +00:00
def mount(self, mountpoint, extra_options, foreground=False):
options = ['fsname=borgfs', 'ro']
if extra_options:
options.extend(extra_options.split(','))
try:
options.remove('allow_damaged_files')
self.allow_damaged_files = True
except ValueError:
self.allow_damaged_files = False
llfuse.init(self, mountpoint, options)
if not foreground:
daemonize()
# If the file system crashes, we do not want to umount because in that
# case the mountpoint suddenly appears to become empty. This can have
# nasty consequences, imagine the user has e.g. an active rsync mirror
# job - seeing the mountpoint empty, rsync would delete everything in the
# mirror.
umount = False
2013-07-23 11:40:54 +00:00
try:
signal = fuse_main()
2016-02-17 00:05:04 +00:00
umount = (signal is None) # no crash and no signal -> umount request
2015-03-17 22:16:12 +00:00
finally:
llfuse.close(umount)