mirror of
https://github.com/morpheus65535/bazarr
synced 2024-12-30 19:46:25 +00:00
184 lines
6.6 KiB
Python
184 lines
6.6 KiB
Python
import datetime
|
|
import os
|
|
import threading
|
|
import time
|
|
import pickle
|
|
|
|
try:
|
|
import pymongo
|
|
import pymongo.errors
|
|
import bson
|
|
except ImportError:
|
|
pymongo = None
|
|
bson = None
|
|
|
|
from beaker.container import NamespaceManager
|
|
from beaker.synchronization import SynchronizerImpl
|
|
from beaker.util import SyncDict, machine_identifier
|
|
from beaker.crypto.util import sha1
|
|
from beaker._compat import string_type, PY2
|
|
|
|
|
|
class MongoNamespaceManager(NamespaceManager):
|
|
"""Provides the :class:`.NamespaceManager` API over MongoDB.
|
|
|
|
Provided ``url`` can be both a mongodb connection string or
|
|
an already existing MongoClient instance.
|
|
|
|
The data will be stored into ``beaker_cache`` collection of the
|
|
*default database*, so make sure your connection string or
|
|
MongoClient point to a default database.
|
|
"""
|
|
MAX_KEY_LENGTH = 1024
|
|
|
|
clients = SyncDict()
|
|
|
|
def __init__(self, namespace, url, **kw):
|
|
super(MongoNamespaceManager, self).__init__(namespace)
|
|
self.lock_dir = None # MongoDB uses mongo itself for locking.
|
|
|
|
if pymongo is None:
|
|
raise RuntimeError('pymongo3 is not available')
|
|
|
|
if isinstance(url, string_type):
|
|
self.client = MongoNamespaceManager.clients.get(url, pymongo.MongoClient, url)
|
|
else:
|
|
self.client = url
|
|
self.db = self.client.get_default_database()
|
|
|
|
def _format_key(self, key):
|
|
if not isinstance(key, str):
|
|
key = key.decode('ascii')
|
|
if len(key) > (self.MAX_KEY_LENGTH - len(self.namespace) - 1):
|
|
if not PY2:
|
|
key = key.encode('utf-8')
|
|
key = sha1(key).hexdigest()
|
|
return '%s:%s' % (self.namespace, key)
|
|
|
|
def get_creation_lock(self, key):
|
|
return MongoSynchronizer(self._format_key(key), self.client)
|
|
|
|
def __getitem__(self, key):
|
|
self._clear_expired()
|
|
entry = self.db.backer_cache.find_one({'_id': self._format_key(key)})
|
|
if entry is None:
|
|
raise KeyError(key)
|
|
return pickle.loads(entry['value'])
|
|
|
|
def __contains__(self, key):
|
|
self._clear_expired()
|
|
entry = self.db.backer_cache.find_one({'_id': self._format_key(key)})
|
|
return entry is not None
|
|
|
|
def has_key(self, key):
|
|
return key in self
|
|
|
|
def set_value(self, key, value, expiretime=None):
|
|
self._clear_expired()
|
|
|
|
expiration = None
|
|
if expiretime is not None:
|
|
expiration = time.time() + expiretime
|
|
|
|
value = pickle.dumps(value)
|
|
self.db.backer_cache.update_one({'_id': self._format_key(key)},
|
|
{'$set': {'value': bson.Binary(value),
|
|
'expiration': expiration}},
|
|
upsert=True)
|
|
|
|
def __setitem__(self, key, value):
|
|
self.set_value(key, value)
|
|
|
|
def __delitem__(self, key):
|
|
self._clear_expired()
|
|
self.db.backer_cache.delete_many({'_id': self._format_key(key)})
|
|
|
|
def do_remove(self):
|
|
self.db.backer_cache.delete_many({'_id': {'$regex': '^%s' % self.namespace}})
|
|
|
|
def keys(self):
|
|
return [e['key'].split(':', 1)[-1] for e in self.db.backer_cache.find_all(
|
|
{'_id': {'$regex': '^%s' % self.namespace}}
|
|
)]
|
|
|
|
def _clear_expired(self):
|
|
now = time.time()
|
|
self.db.backer_cache.delete_many({'_id': {'$regex': '^%s' % self.namespace},
|
|
'expiration': {'$ne': None, '$lte': now}})
|
|
|
|
|
|
class MongoSynchronizer(SynchronizerImpl):
|
|
"""Provides a Writer/Reader lock based on MongoDB.
|
|
|
|
Provided ``url`` can be both a mongodb connection string or
|
|
an already existing MongoClient instance.
|
|
|
|
The data will be stored into ``beaker_locks`` collection of the
|
|
*default database*, so make sure your connection string or
|
|
MongoClient point to a default database.
|
|
|
|
Locks are identified by local machine, PID and threadid, so
|
|
are suitable for use in both local and distributed environments.
|
|
"""
|
|
# If a cache entry generation function can take a lot,
|
|
# but 15 minutes is more than a reasonable time.
|
|
LOCK_EXPIRATION = 900
|
|
MACHINE_ID = machine_identifier()
|
|
|
|
def __init__(self, identifier, url):
|
|
super(MongoSynchronizer, self).__init__()
|
|
self.identifier = identifier
|
|
if isinstance(url, string_type):
|
|
self.client = MongoNamespaceManager.clients.get(url, pymongo.MongoClient, url)
|
|
else:
|
|
self.client = url
|
|
self.db = self.client.get_default_database()
|
|
|
|
def _clear_expired_locks(self):
|
|
now = datetime.datetime.utcnow()
|
|
expired = now - datetime.timedelta(seconds=self.LOCK_EXPIRATION)
|
|
self.db.beaker_locks.delete_many({'_id': self.identifier, 'timestamp': {'$lte': expired}})
|
|
return now
|
|
|
|
def _get_owner_id(self):
|
|
return '%s-%s-%s' % (self.MACHINE_ID, os.getpid(), threading.current_thread().ident)
|
|
|
|
def do_release_read_lock(self):
|
|
owner_id = self._get_owner_id()
|
|
self.db.beaker_locks.update_one({'_id': self.identifier, 'readers': owner_id},
|
|
{'$pull': {'readers': owner_id}})
|
|
|
|
def do_acquire_read_lock(self, wait):
|
|
now = self._clear_expired_locks()
|
|
owner_id = self._get_owner_id()
|
|
while True:
|
|
try:
|
|
self.db.beaker_locks.update_one({'_id': self.identifier, 'owner': None},
|
|
{'$set': {'timestamp': now},
|
|
'$push': {'readers': owner_id}},
|
|
upsert=True)
|
|
return True
|
|
except pymongo.errors.DuplicateKeyError:
|
|
if not wait:
|
|
return False
|
|
time.sleep(0.2)
|
|
|
|
def do_release_write_lock(self):
|
|
self.db.beaker_locks.delete_one({'_id': self.identifier, 'owner': self._get_owner_id()})
|
|
|
|
def do_acquire_write_lock(self, wait):
|
|
now = self._clear_expired_locks()
|
|
owner_id = self._get_owner_id()
|
|
while True:
|
|
try:
|
|
self.db.beaker_locks.update_one({'_id': self.identifier, 'owner': None,
|
|
'readers': []},
|
|
{'$set': {'owner': owner_id,
|
|
'timestamp': now}},
|
|
upsert=True)
|
|
return True
|
|
except pymongo.errors.DuplicateKeyError:
|
|
if not wait:
|
|
return False
|
|
time.sleep(0.2)
|
|
|