1
0
Fork 0
mirror of https://github.com/borgbackup/borg.git synced 2024-12-26 01:37:20 +00:00

minor cleanup: fix typos and punctuation in comments

This commit is contained in:
Thomas Waldmann 2022-08-21 11:36:27 +02:00
parent 709d162ee7
commit 15d56b2ff7

View file

@ -429,7 +429,7 @@ def check_transaction(self):
msg = '%s" - although likely this is "beyond repair' % self.path # dirty hack msg = '%s" - although likely this is "beyond repair' % self.path # dirty hack
raise self.CheckNeeded(msg) raise self.CheckNeeded(msg)
# Attempt to automatically rebuild index if we crashed between commit # Attempt to automatically rebuild index if we crashed between commit
# tag write and index save # tag write and index save.
if index_transaction_id != segments_transaction_id: if index_transaction_id != segments_transaction_id:
if index_transaction_id is not None and index_transaction_id > segments_transaction_id: if index_transaction_id is not None and index_transaction_id > segments_transaction_id:
replay_from = None replay_from = None
@ -563,7 +563,7 @@ def prepare_txn(self, transaction_id, do_cleanup=True):
# self.exclusive is either True or False, thus a new client is active here. # self.exclusive is either True or False, thus a new client is active here.
# if it is False and we get here, the caller did not use exclusive=True although # if it is False and we get here, the caller did not use exclusive=True although
# it is needed for a write operation. if it is True and we get here, something else # it is needed for a write operation. if it is True and we get here, something else
# went very wrong, because we should have a exclusive lock, but we don't. # went very wrong, because we should have an exclusive lock, but we don't.
raise AssertionError("bug in code, exclusive lock should exist here") raise AssertionError("bug in code, exclusive lock should exist here")
# if we are here, this is an old client talking to a new server (expecting lock upgrade). # if we are here, this is an old client talking to a new server (expecting lock upgrade).
# or we are replaying segments and might need a lock upgrade for that. # or we are replaying segments and might need a lock upgrade for that.
@ -876,10 +876,10 @@ def complete_xfer(intermediate=True):
# #
# However, this only happens if the crash also affects the FS to the effect that file deletions # However, this only happens if the crash also affects the FS to the effect that file deletions
# did not materialize consistently after journal recovery. If they always materialize in-order # did not materialize consistently after journal recovery. If they always materialize in-order
# then this is not a problem, because the old segment containing a deleted object would be deleted # then this is not a problem, because the old segment containing a deleted object would be
# before the segment containing the delete. # deleted before the segment containing the delete.
# #
# Consider the following series of operations if we would not do this, ie. this entire if: # Consider the following series of operations if we would not do this, i.e. this entire if:
# would be removed. # would be removed.
# Columns are segments, lines are different keys (line 1 = some key, line 2 = some other key) # Columns are segments, lines are different keys (line 1 = some key, line 2 = some other key)
# Legend: P=TAG_PUT/TAG_PUT2, D=TAG_DELETE, c=commit, i=index is written for latest commit # Legend: P=TAG_PUT/TAG_PUT2, D=TAG_DELETE, c=commit, i=index is written for latest commit
@ -902,7 +902,7 @@ def complete_xfer(intermediate=True):
# fixed by borg-check --repair. # fixed by borg-check --repair.
# #
# Note that in this check the index state is the proxy for a "most definitely settled" repository state, # Note that in this check the index state is the proxy for a "most definitely settled" repository state,
# ie. the assumption is that *all* operations on segments <= index state are completed and stable. # i.e. the assumption is that *all* operations on segments <= index state are completed and stable.
try: try:
new_segment, size = self.io.write_delete(key, raise_full=True) new_segment, size = self.io.write_delete(key, raise_full=True)
except LoggedIO.SegmentFull: except LoggedIO.SegmentFull:
@ -1001,7 +1001,7 @@ def _rebuild_sparse(self, segment):
try: try:
segment_size = self.io.segment_size(segment) segment_size = self.io.segment_size(segment)
except FileNotFoundError: except FileNotFoundError:
# segment does not exist any more, remove it from the mappings # segment does not exist any more, remove it from the mappings.
# note: no need to self.compact.pop(segment), as we start from empty mapping. # note: no need to self.compact.pop(segment), as we start from empty mapping.
self.segments.pop(segment) self.segments.pop(segment)
return return
@ -1015,10 +1015,10 @@ def _rebuild_sparse(self, segment):
if tag in (TAG_PUT2, TAG_PUT): if tag in (TAG_PUT2, TAG_PUT):
in_index = self.index.get(key) in_index = self.index.get(key)
if not in_index or (in_index.segment, in_index.offset) != (segment, offset): if not in_index or (in_index.segment, in_index.offset) != (segment, offset):
# This PUT is superseded later # This PUT is superseded later.
self.compact[segment] += header_size(tag) + size self.compact[segment] += header_size(tag) + size
elif tag == TAG_DELETE: elif tag == TAG_DELETE:
# The outcome of the DELETE has been recorded in the PUT branch already # The outcome of the DELETE has been recorded in the PUT branch already.
self.compact[segment] += header_size(tag) + size self.compact[segment] += header_size(tag) + size
def check(self, repair=False, save_space=False, max_duration=0): def check(self, repair=False, save_space=False, max_duration=0):
@ -1108,8 +1108,8 @@ def report_error(msg):
self.save_config(self.path, self.config) self.save_config(self.path, self.config)
pi.finish() pi.finish()
# self.index, self.segments, self.compact now reflect the state of the segment files up to <transaction_id> # self.index, self.segments, self.compact now reflect the state of the segment files up to <transaction_id>.
# We might need to add a commit tag if no committed segment is found # We might need to add a commit tag if no committed segment is found.
if repair and segments_transaction_id is None: if repair and segments_transaction_id is None:
report_error(f"Adding commit tag to segment {transaction_id}") report_error(f"Adding commit tag to segment {transaction_id}")
self.io.segment = transaction_id + 1 self.io.segment = transaction_id + 1
@ -1177,7 +1177,6 @@ def scan_low_level(self, segment=None, offset=None):
) )
def _rollback(self, *, cleanup): def _rollback(self, *, cleanup):
""" """
if cleanup: if cleanup:
self.io.cleanup(self.io.get_segments_transaction_id()) self.io.cleanup(self.io.get_segments_transaction_id())
self.index = None self.index = None
@ -1234,7 +1233,7 @@ def scan(self, limit=None, marker=None):
tag, id, offset, size = next(obj_iterator) tag, id, offset, size = next(obj_iterator)
except (StopIteration, IntegrityError): except (StopIteration, IntegrityError):
# either end-of-segment or an error - we can not seek to objects at # either end-of-segment or an error - we can not seek to objects at
# higher offsets than one that has an error in the header fields # higher offsets than one that has an error in the header fields.
break break
if start_offset > 0: if start_offset > 0:
# we are using a marker and the marker points to the last object we have already # we are using a marker and the marker points to the last object we have already
@ -1255,7 +1254,7 @@ def get(self, id):
if not self.index: if not self.index:
self.index = self.open_index(self.get_transaction_id()) self.index = self.open_index(self.get_transaction_id())
try: try:
in_index = NSIndexEntry(*((self.index[id] + (None,))[:3])) # legacy: index entriess have no size element in_index = NSIndexEntry(*((self.index[id] + (None,))[:3])) # legacy: index entries have no size element
return self.io.read(in_index.segment, in_index.offset, id, expected_size=in_index.size) return self.io.read(in_index.segment, in_index.offset, id, expected_size=in_index.size)
except KeyError: except KeyError:
raise self.ObjectNotFound(id, self.path) from None raise self.ObjectNotFound(id, self.path) from None
@ -1473,7 +1472,7 @@ def get_write_fd(self, no_new=False, want_new=False, raise_full=False):
if self.segment in self.fds: if self.segment in self.fds:
# we may have a cached fd for a segment file we already deleted and # we may have a cached fd for a segment file we already deleted and
# we are writing now a new segment file to same file name. get rid of # we are writing now a new segment file to same file name. get rid of
# of the cached fd that still refers to the old file, so it will later # the cached fd that still refers to the old file, so it will later
# get repopulated (on demand) with a fd that refers to the new file. # get repopulated (on demand) with a fd that refers to the new file.
del self.fds[self.segment] del self.fds[self.segment]
return self._write_fd return self._write_fd