mirror of https://github.com/morpheus65535/bazarr
Initial commit
This commit is contained in:
parent
4da97428bc
commit
963947d636
|
@ -1,53 +1,45 @@
|
|||
# coding=utf-8
|
||||
import os
|
||||
import platform
|
||||
import re
|
||||
import subprocess
|
||||
import tarfile
|
||||
import logging
|
||||
import requests
|
||||
import sqlite3
|
||||
import json
|
||||
import requests
|
||||
|
||||
from get_args import args
|
||||
from config import settings
|
||||
|
||||
if not args.no_update:
|
||||
import git
|
||||
bazarr_version = '7.2.0'
|
||||
|
||||
current_working_directory = os.path.dirname(os.path.dirname(__file__))
|
||||
LATEST_VERSION = None
|
||||
INSTALL_TYPE = None
|
||||
UPDATE_AVAILABLE = None
|
||||
COMMITS_BEHIND = None
|
||||
LATEST_RELEASE = None
|
||||
PREV_RELEASE = bazarr_version
|
||||
|
||||
|
||||
def gitconfig():
|
||||
g = git.Repo.init(current_working_directory)
|
||||
config_read = g.config_reader()
|
||||
config_write = g.config_writer()
|
||||
class FakeLock(object):
|
||||
"""
|
||||
If no locking or request throttling is needed, use this
|
||||
"""
|
||||
|
||||
try:
|
||||
username = config_read.get_value("user", "name")
|
||||
except:
|
||||
logging.debug('BAZARR Settings git username')
|
||||
config_write.set_value("user", "name", "Bazarr")
|
||||
def __enter__(self):
|
||||
"""
|
||||
Do nothing on enter
|
||||
"""
|
||||
pass
|
||||
|
||||
try:
|
||||
email = config_read.get_value("user", "email")
|
||||
except:
|
||||
logging.debug('BAZARR Settings git email')
|
||||
config_write.set_value("user", "email", "bazarr@fake.email")
|
||||
def __exit__(self, type, value, traceback):
|
||||
"""
|
||||
Do nothing on exit
|
||||
"""
|
||||
pass
|
||||
|
||||
|
||||
def check_and_apply_update():
|
||||
gitconfig()
|
||||
check_releases()
|
||||
branch = settings.general.branch
|
||||
g = git.cmd.Git(current_working_directory)
|
||||
g.fetch('origin')
|
||||
result = g.diff('--shortstat', 'origin/' + branch)
|
||||
if len(result) == 0:
|
||||
logging.info('BAZARR No new version of Bazarr available.')
|
||||
else:
|
||||
g.reset('--hard', 'HEAD')
|
||||
g.checkout(branch)
|
||||
g.reset('--hard', 'origin/' + branch)
|
||||
g.pull()
|
||||
logging.info('BAZARR Updated to latest version. Restart required. ' + result)
|
||||
updated()
|
||||
fake_lock = FakeLock()
|
||||
|
||||
|
||||
def check_releases():
|
||||
|
@ -71,6 +63,374 @@ def check_releases():
|
|||
json.dump(releases, f)
|
||||
|
||||
|
||||
def runGit(args):
|
||||
git_locations = ['git']
|
||||
|
||||
if platform.system().lower() == 'darwin':
|
||||
git_locations.append('/usr/local/git/bin/git')
|
||||
|
||||
output = err = None
|
||||
|
||||
for cur_git in git_locations:
|
||||
cmd = cur_git + ' ' + args
|
||||
|
||||
try:
|
||||
logging.debug('Trying to execute: "' + cmd + '"')
|
||||
p = subprocess.Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT, shell=True)
|
||||
output, err = p.communicate()
|
||||
output = output.strip()
|
||||
|
||||
logging.debug('Git output: ' + output)
|
||||
except OSError:
|
||||
logging.debug('Command failed: %s', cmd)
|
||||
continue
|
||||
|
||||
if 'not found' in output or "not recognized as an internal or external command" in output:
|
||||
logging.debug('Unable to find git with command ' + cmd)
|
||||
output = None
|
||||
elif 'fatal:' in output or err:
|
||||
logging.error('Git returned bad info. Are you sure this is a git installation?')
|
||||
output = None
|
||||
elif output:
|
||||
break
|
||||
|
||||
return (output, err)
|
||||
|
||||
|
||||
def getVersion():
|
||||
if os.path.isdir(os.path.join(os.path.dirname(__file__), '..', '.git')):
|
||||
|
||||
INSTALL_TYPE = 'git'
|
||||
output, err = runGit('rev-parse HEAD')
|
||||
|
||||
if not output:
|
||||
logging.error('Could not find latest installed version.')
|
||||
cur_commit_hash = None
|
||||
|
||||
cur_commit_hash = str(output)
|
||||
|
||||
if not re.match('^[a-z0-9]+$', cur_commit_hash):
|
||||
logging.error('Output does not look like a hash, not using it.')
|
||||
cur_commit_hash = None
|
||||
|
||||
if settings.general.branch:
|
||||
branch_name = settings.general.branch
|
||||
|
||||
else:
|
||||
remote_branch, err = runGit('rev-parse --abbrev-ref --symbolic-full-name @{u}')
|
||||
remote_branch = remote_branch.rsplit('/', 1) if remote_branch else []
|
||||
if len(remote_branch) == 2:
|
||||
remote_name, branch_name = remote_branch
|
||||
else:
|
||||
remote_name = branch_name = None
|
||||
|
||||
if not remote_name and settings.general.branch:
|
||||
logging.error('Could not retrieve remote name from git. Defaulting to origin.')
|
||||
branch_name = settings.general.branch
|
||||
|
||||
if not branch_name:
|
||||
logging.error('Could not retrieve branch name from git. Defaulting to master.')
|
||||
branch_name = 'master'
|
||||
|
||||
return cur_commit_hash, 'origin', branch_name
|
||||
|
||||
else:
|
||||
INSTALL_TYPE = 'source'
|
||||
|
||||
if CURRENT_VERSION:
|
||||
return CURRENT_VERSION, 'origin', settings.general.branch
|
||||
else:
|
||||
return None, 'origin', settings.general.branch
|
||||
|
||||
|
||||
CURRENT_VERSION, GIT_REMOTE, GIT_BRANCH = getVersion()
|
||||
|
||||
|
||||
def check_updates():
|
||||
check_github()
|
||||
if not CURRENT_VERSION:
|
||||
UPDATE_AVAILABLE = None
|
||||
elif COMMITS_BEHIND > 0 and settings.general.branch in ('master') and \
|
||||
('V' + bazarr_version) != LATEST_RELEASE:
|
||||
UPDATE_AVAILABLE = 'release'
|
||||
elif COMMITS_BEHIND > 0 and CURRENT_VERSION != LATEST_VERSION:
|
||||
UPDATE_AVAILABLE = 'commit'
|
||||
else:
|
||||
UPDATE_AVAILABLE = False
|
||||
|
||||
|
||||
def check_github():
|
||||
COMMITS_BEHIND = 0
|
||||
|
||||
# Get the latest version available from github
|
||||
logging.info('Retrieving latest version information from GitHub')
|
||||
url = 'https://api.github.com/repos/morpheus65535/bazarr/commits/%s' % settings.general.branch
|
||||
version = request_json(url, timeout=20, validator=lambda x: type(x) == dict)
|
||||
|
||||
if version is None:
|
||||
logging.warn('Could not get the latest version from GitHub. Are you running a local development version?')
|
||||
return CURRENT_VERSION
|
||||
|
||||
LATEST_VERSION = version['sha']
|
||||
logging.debug("Latest version is %s", LATEST_VERSION)
|
||||
|
||||
# See how many commits behind we are
|
||||
if not CURRENT_VERSION:
|
||||
logging.info('You are running an unknown version of Bazarr. Run the updater to identify your version')
|
||||
return LATEST_VERSION
|
||||
|
||||
if LATEST_VERSION == CURRENT_VERSION:
|
||||
logging.info('Bazarr is up to date')
|
||||
return LATEST_VERSION
|
||||
|
||||
logging.info('Comparing currently installed version with latest GitHub version')
|
||||
url = 'https://api.github.com/repos/morpheus65535/bazarr/compare/%s...%s' % (LATEST_VERSION,
|
||||
CURRENT_VERSION)
|
||||
commits = request_json(url, timeout=20, whitelist_status_code=404, validator=lambda x: type(x) == dict)
|
||||
|
||||
if commits is None:
|
||||
logging.warn('Could not get commits behind from GitHub.')
|
||||
return LATEST_VERSION
|
||||
|
||||
try:
|
||||
COMMITS_BEHIND = int(commits['behind_by'])
|
||||
logging.debug("In total, %d commits behind", COMMITS_BEHIND)
|
||||
except KeyError:
|
||||
logging.info('Cannot compare versions. Are you running a local development version?')
|
||||
COMMITS_BEHIND = 0
|
||||
|
||||
if COMMITS_BEHIND > 0:
|
||||
logging.info('New version is available. You are %s commits behind' % COMMITS_BEHIND)
|
||||
|
||||
url = 'https://api.github.com/repos/morpheus65535/bazarr/releases'
|
||||
releases = request_json(url, timeout=20, whitelist_status_code=404, validator=lambda x: type(x) == list)
|
||||
|
||||
if releases is None:
|
||||
logging.warn('Could not get releases from GitHub.')
|
||||
return LATEST_VERSION
|
||||
|
||||
if settings.general.branch == 'master':
|
||||
release = next((r for r in releases if not r['prerelease']), releases[0])
|
||||
else:
|
||||
release = releases[0]
|
||||
LATEST_RELEASE = release['tag_name']
|
||||
|
||||
elif COMMITS_BEHIND == 0:
|
||||
logging.info('Bazarr is up to date')
|
||||
|
||||
return LATEST_VERSION
|
||||
|
||||
|
||||
def update():
|
||||
if INSTALL_TYPE == 'git':
|
||||
output, err = runGit('pull ' + 'origin' + ' ' + settings.general.branch)
|
||||
|
||||
if not output:
|
||||
logging.error('Unable to download latest version')
|
||||
return
|
||||
|
||||
for line in output.split('\n'):
|
||||
|
||||
if 'Already up-to-date.' in line:
|
||||
logging.info('No update available, not updating')
|
||||
logging.info('Output: ' + str(output))
|
||||
elif line.endswith(('Aborting', 'Aborting.')):
|
||||
logging.error('Unable to update from git: ' + line)
|
||||
logging.info('Output: ' + str(output))
|
||||
|
||||
else:
|
||||
tar_download_url = 'https://github.com/morpheus65535/bazarr/tarball/{}'.format(settings.general.branch)
|
||||
update_dir = os.path.join(os.path.dirname(__file__), '..', 'update')
|
||||
|
||||
logging.info('Downloading update from: ' + tar_download_url)
|
||||
data = request_content(tar_download_url)
|
||||
|
||||
if not data:
|
||||
logging.error("Unable to retrieve new version from '%s', can't update", tar_download_url)
|
||||
return
|
||||
|
||||
download_name = settings.general.branch + '-github'
|
||||
tar_download_path = os.path.join(os.path.dirname(__file__), '..', download_name)
|
||||
|
||||
# Save tar to disk
|
||||
with open(tar_download_path, 'wb') as f:
|
||||
f.write(data)
|
||||
|
||||
# Extract the tar to update folder
|
||||
logging.info('Extracting file: ' + tar_download_path)
|
||||
tar = tarfile.open(tar_download_path)
|
||||
tar.extractall(update_dir)
|
||||
tar.close()
|
||||
|
||||
# Delete the tar.gz
|
||||
logging.info('Deleting file: ' + tar_download_path)
|
||||
os.remove(tar_download_path)
|
||||
|
||||
# Find update dir name
|
||||
update_dir_contents = [x for x in os.listdir(update_dir) if os.path.isdir(os.path.join(update_dir, x))]
|
||||
if len(update_dir_contents) != 1:
|
||||
logging.error("Invalid update data, update failed: " + str(update_dir_contents))
|
||||
return
|
||||
content_dir = os.path.join(update_dir, update_dir_contents[0])
|
||||
|
||||
# walk temp folder and move files to main folder
|
||||
for dirname, dirnames, filenames in os.walk(content_dir):
|
||||
dirname = dirname[len(content_dir) + 1:]
|
||||
for curfile in filenames:
|
||||
old_path = os.path.join(content_dir, dirname, curfile)
|
||||
new_path = os.path.join(os.path.dirname(__file__), '..', dirname, curfile)
|
||||
|
||||
if os.path.isfile(new_path):
|
||||
os.remove(new_path)
|
||||
os.renames(old_path, new_path)
|
||||
|
||||
|
||||
def checkout_git_branch():
|
||||
if INSTALL_TYPE == 'git':
|
||||
output, err = runGit('fetch origin')
|
||||
output, err = runGit('checkout %s' % settings.general.branch)
|
||||
|
||||
if not output:
|
||||
logging.error('Unable to change git branch.')
|
||||
return
|
||||
|
||||
for line in output.split('\n'):
|
||||
if line.endswith(('Aborting', 'Aborting.')):
|
||||
logging.error('Unable to checkout from git: ' + line)
|
||||
logging.info('Output: ' + str(output))
|
||||
|
||||
output, err = runGit('pull origin %s' % settings.general.branch)
|
||||
|
||||
|
||||
def request_content(url, **kwargs):
|
||||
"""
|
||||
Wrapper for `request_response', which will return the raw content.
|
||||
"""
|
||||
|
||||
response = request_response(url, **kwargs)
|
||||
|
||||
if response is not None:
|
||||
return response.content
|
||||
|
||||
|
||||
def request_response(url, method="get", auto_raise=True,
|
||||
whitelist_status_code=None, lock=fake_lock, **kwargs):
|
||||
"""
|
||||
Convenient wrapper for `requests.get', which will capture the exceptions
|
||||
and log them. On success, the Response object is returned. In case of a
|
||||
exception, None is returned.
|
||||
|
||||
Additionally, there is support for rate limiting. To use this feature,
|
||||
supply a tuple of (lock, request_limit). The lock is used to make sure no
|
||||
other request with the same lock is executed. The request limit is the
|
||||
minimal time between two requests (and so 1/request_limit is the number of
|
||||
requests per seconds).
|
||||
"""
|
||||
|
||||
# Convert whitelist_status_code to a list if needed
|
||||
if whitelist_status_code and type(whitelist_status_code) != list:
|
||||
whitelist_status_code = [whitelist_status_code]
|
||||
|
||||
# Disable verification of SSL certificates if requested. Note: this could
|
||||
# pose a security issue!
|
||||
kwargs["verify"] = True
|
||||
|
||||
# Map method to the request.XXX method. This is a simple hack, but it
|
||||
# allows requests to apply more magic per method. See lib/requests/api.py.
|
||||
request_method = getattr(requests, method.lower())
|
||||
|
||||
try:
|
||||
# Request URL and wait for response
|
||||
with lock:
|
||||
logging.debug(
|
||||
"Requesting URL via %s method: %s", method.upper(), url)
|
||||
response = request_method(url, **kwargs)
|
||||
|
||||
# If status code != OK, then raise exception, except if the status code
|
||||
# is white listed.
|
||||
if whitelist_status_code and auto_raise:
|
||||
if response.status_code not in whitelist_status_code:
|
||||
try:
|
||||
response.raise_for_status()
|
||||
except:
|
||||
logging.debug(
|
||||
"Response status code %d is not white "
|
||||
"listed, raised exception", response.status_code)
|
||||
raise
|
||||
elif auto_raise:
|
||||
response.raise_for_status()
|
||||
|
||||
return response
|
||||
except requests.exceptions.SSLError as e:
|
||||
if kwargs["verify"]:
|
||||
logging.error(
|
||||
"Unable to connect to remote host because of a SSL error. "
|
||||
"It is likely that your system cannot verify the validity"
|
||||
"of the certificate. The remote certificate is either "
|
||||
"self-signed, or the remote server uses SNI. See the wiki for "
|
||||
"more information on this topic.")
|
||||
else:
|
||||
logging.error(
|
||||
"SSL error raised during connection, with certificate "
|
||||
"verification turned off: %s", e)
|
||||
except requests.ConnectionError:
|
||||
logging.error(
|
||||
"Unable to connect to remote host. Check if the remote "
|
||||
"host is up and running.")
|
||||
except requests.Timeout:
|
||||
logging.error(
|
||||
"Request timed out. The remote host did not respond timely.")
|
||||
except requests.HTTPError as e:
|
||||
if e.response is not None:
|
||||
if e.response.status_code >= 500:
|
||||
cause = "remote server error"
|
||||
elif e.response.status_code >= 400:
|
||||
cause = "local client error"
|
||||
else:
|
||||
# I don't think we will end up here, but for completeness
|
||||
cause = "unknown"
|
||||
|
||||
logging.error(
|
||||
"Request raise HTTP error with status code %d (%s).",
|
||||
e.response.status_code, cause)
|
||||
|
||||
# Debug response
|
||||
# if bazarr.DEBUG:
|
||||
# server_message(e.response)
|
||||
else:
|
||||
logging.error("Request raised HTTP error.")
|
||||
except requests.RequestException as e:
|
||||
logging.error("Request raised exception: %s", e)
|
||||
|
||||
|
||||
def request_json(url, **kwargs):
|
||||
"""
|
||||
Wrapper for `request_response', which will decode the response as JSON
|
||||
object and return the result, if no exceptions are raised.
|
||||
|
||||
As an option, a validator callback can be given, which should return True
|
||||
if the result is valid.
|
||||
"""
|
||||
|
||||
validator = kwargs.pop("validator", None)
|
||||
response = request_response(url, **kwargs)
|
||||
|
||||
if response is not None:
|
||||
try:
|
||||
result = response.json()
|
||||
|
||||
if validator and not validator(result):
|
||||
logging.error("JSON validation result failed")
|
||||
else:
|
||||
return result
|
||||
except ValueError:
|
||||
logging.error("Response returned invalid JSON data")
|
||||
|
||||
# Debug response
|
||||
# if bazarr.DEBUG:
|
||||
# server_message(response)
|
||||
|
||||
|
||||
def updated():
|
||||
conn = sqlite3.connect(os.path.join(args.config_dir, 'db', 'bazarr.db'), timeout=30)
|
||||
c = conn.cursor()
|
||||
|
|
|
@ -58,7 +58,7 @@ from get_series import *
|
|||
from get_episodes import *
|
||||
|
||||
if not args.no_update:
|
||||
from check_update import check_and_apply_update
|
||||
from check_update import check_updates
|
||||
from list_subtitles import store_subtitles, store_subtitles_movie, series_scan_subtitles, movies_scan_subtitles, \
|
||||
list_missing_subtitles, list_missing_subtitles_movies
|
||||
from get_subtitle import download_subtitle, series_download_subtitles, movies_download_subtitles, \
|
||||
|
|
|
@ -8,7 +8,7 @@ from get_subtitle import wanted_search_missing_subtitles, upgrade_subtitles
|
|||
from get_args import args
|
||||
|
||||
if not args.no_update:
|
||||
from check_update import check_and_apply_update, check_releases
|
||||
from check_update import check_updates, check_releases
|
||||
else:
|
||||
from check_update import check_releases
|
||||
from apscheduler.schedulers.background import BackgroundScheduler
|
||||
|
@ -84,10 +84,10 @@ scheduler.add_listener(task_listener, EVENT_JOB_SUBMITTED | EVENT_JOB_EXECUTED)
|
|||
|
||||
if not args.no_update:
|
||||
if settings.general.getboolean('auto_update'):
|
||||
scheduler.add_job(check_and_apply_update, IntervalTrigger(hours=6), max_instances=1, coalesce=True,
|
||||
scheduler.add_job(check_updates, IntervalTrigger(hours=6), max_instances=1, coalesce=True,
|
||||
misfire_grace_time=15, id='update_bazarr', name='Update bazarr from source on Github')
|
||||
else:
|
||||
scheduler.add_job(check_and_apply_update, CronTrigger(year='2100'), hour=4, id='update_bazarr',
|
||||
scheduler.add_job(check_updates, CronTrigger(year='2100'), hour=4, id='update_bazarr',
|
||||
name='Update bazarr from source on Github')
|
||||
scheduler.add_job(check_releases, IntervalTrigger(hours=6), max_instances=1, coalesce=True,
|
||||
misfire_grace_time=15, id='update_release', name='Update release info')
|
||||
|
|
Loading…
Reference in New Issue