2019-04-07 23:31:39 +00:00
|
|
|
# -*- coding: utf-8 -*-
|
2019-09-17 02:04:27 +00:00
|
|
|
from __future__ import absolute_import
|
2019-04-07 23:31:39 +00:00
|
|
|
import logging
|
|
|
|
import re
|
|
|
|
import io
|
|
|
|
import os
|
2020-05-24 10:02:34 +00:00
|
|
|
import codecs
|
|
|
|
from hashlib import sha1
|
2019-04-07 23:31:39 +00:00
|
|
|
from random import randint
|
|
|
|
from bs4 import BeautifulSoup
|
|
|
|
from zipfile import ZipFile, is_zipfile
|
|
|
|
from rarfile import RarFile, is_rarfile
|
|
|
|
from requests import Session
|
|
|
|
from guessit import guessit
|
2020-05-24 10:02:34 +00:00
|
|
|
from dogpile.cache.api import NO_VALUE
|
2019-04-07 23:31:39 +00:00
|
|
|
from subliminal_patch.providers import Provider
|
|
|
|
from subliminal_patch.subtitle import Subtitle
|
2020-03-20 07:56:18 +00:00
|
|
|
from subliminal_patch.utils import sanitize, fix_inconsistent_naming
|
2019-04-07 23:31:39 +00:00
|
|
|
from subliminal.subtitle import guess_matches
|
|
|
|
from subliminal.video import Episode, Movie
|
|
|
|
from subliminal.subtitle import fix_line_ending
|
2020-05-24 10:02:34 +00:00
|
|
|
from subliminal.cache import region
|
2019-04-07 23:31:39 +00:00
|
|
|
from subzero.language import Language
|
|
|
|
from .utils import FIRST_THOUSAND_OR_SO_USER_AGENTS as AGENT_LIST
|
|
|
|
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
2020-05-31 15:07:59 +00:00
|
|
|
|
2020-03-20 07:56:18 +00:00
|
|
|
def fix_tv_naming(title):
|
|
|
|
"""Fix TV show titles with inconsistent naming using dictionary, but do not sanitize them.
|
|
|
|
|
|
|
|
:param str title: original title.
|
|
|
|
:return: new title.
|
|
|
|
:rtype: str
|
|
|
|
|
|
|
|
"""
|
|
|
|
return fix_inconsistent_naming(title, {"Marvel's Daredevil": "Daredevil",
|
|
|
|
"Marvel's Luke Cage": "Luke Cage",
|
|
|
|
"Marvel's Iron Fist": "Iron Fist",
|
|
|
|
"Marvel's Jessica Jones": "Jessica Jones",
|
2020-04-21 20:16:15 +00:00
|
|
|
"DC's Legends of Tomorrow": "Legends of Tomorrow",
|
|
|
|
"Doctor Who (2005)": "Doctor Who",
|
2020-03-20 07:56:18 +00:00
|
|
|
}, True)
|
|
|
|
|
2020-05-31 15:07:59 +00:00
|
|
|
|
2020-09-13 11:24:49 +00:00
|
|
|
def fix_movie_naming(title):
|
|
|
|
return fix_inconsistent_naming(title, {"Back to the Future Part": "Back to the Future",
|
|
|
|
}, True)
|
|
|
|
|
|
|
|
|
2019-04-07 23:31:39 +00:00
|
|
|
class SubsSabBzSubtitle(Subtitle):
|
|
|
|
"""SubsSabBz Subtitle."""
|
|
|
|
provider_name = 'subssabbz'
|
|
|
|
|
2020-05-31 15:07:59 +00:00
|
|
|
def __init__(self, language, filename, type, video, link, fps, num_cds):
|
|
|
|
super(SubsSabBzSubtitle, self).__init__(language)
|
2019-04-07 23:31:39 +00:00
|
|
|
self.filename = filename
|
2019-04-08 17:58:13 +00:00
|
|
|
self.page_link = link
|
2019-04-07 23:31:39 +00:00
|
|
|
self.type = type
|
2019-04-08 17:58:13 +00:00
|
|
|
self.video = video
|
2020-04-25 17:24:58 +00:00
|
|
|
self.fps = fps
|
|
|
|
self.num_cds = num_cds
|
2020-08-16 08:01:21 +00:00
|
|
|
self.release_info = filename
|
|
|
|
if fps:
|
|
|
|
if video.fps and float(video.fps) == fps:
|
|
|
|
self.release_info += " <b>[{:.3f}]</b>".format(fps)
|
|
|
|
else:
|
|
|
|
self.release_info += " [{:.3f}]".format(fps)
|
2019-04-07 23:31:39 +00:00
|
|
|
|
|
|
|
@property
|
|
|
|
def id(self):
|
2020-04-25 17:24:58 +00:00
|
|
|
return self.page_link + self.filename
|
|
|
|
|
|
|
|
def get_fps(self):
|
|
|
|
return self.fps
|
2019-04-07 23:31:39 +00:00
|
|
|
|
2019-04-08 17:58:13 +00:00
|
|
|
def make_picklable(self):
|
2020-04-26 02:08:09 +00:00
|
|
|
self.content = None
|
2020-05-24 10:02:34 +00:00
|
|
|
self._is_valid = False
|
2019-04-08 17:58:13 +00:00
|
|
|
return self
|
|
|
|
|
2019-04-07 23:31:39 +00:00
|
|
|
def get_matches(self, video):
|
|
|
|
matches = set()
|
|
|
|
|
|
|
|
video_filename = video.name
|
|
|
|
video_filename = os.path.basename(video_filename)
|
|
|
|
video_filename, _ = os.path.splitext(video_filename)
|
2020-05-24 10:02:34 +00:00
|
|
|
video_filename = re.sub(r'\[\w+\]$', '', video_filename).strip().upper()
|
2019-04-07 23:31:39 +00:00
|
|
|
|
|
|
|
subtitle_filename = self.filename
|
|
|
|
subtitle_filename = os.path.basename(subtitle_filename)
|
|
|
|
subtitle_filename, _ = os.path.splitext(subtitle_filename)
|
2020-05-24 10:02:34 +00:00
|
|
|
subtitle_filename = re.sub(r'\[\w+\]$', '', subtitle_filename).strip().upper()
|
2019-04-07 23:31:39 +00:00
|
|
|
|
2020-05-24 10:02:34 +00:00
|
|
|
if ((video_filename == subtitle_filename) or
|
|
|
|
(self.single_file is True and video_filename in self.notes.upper())):
|
2020-05-31 15:07:59 +00:00
|
|
|
matches.add('hash')
|
2019-04-07 23:31:39 +00:00
|
|
|
|
2020-04-25 17:24:58 +00:00
|
|
|
if video.year and self.year == video.year:
|
|
|
|
matches.add('year')
|
|
|
|
|
|
|
|
if isinstance(video, Movie):
|
|
|
|
if video.imdb_id and self.imdb_id == video.imdb_id:
|
|
|
|
matches.add('imdb_id')
|
|
|
|
|
2020-05-24 10:02:34 +00:00
|
|
|
matches |= guess_matches(video, guessit(self.title, {'type': self.type}))
|
2020-05-31 15:07:59 +00:00
|
|
|
|
|
|
|
guess_filename = guessit(self.filename, video.hints)
|
|
|
|
matches |= guess_matches(video, guess_filename)
|
|
|
|
|
|
|
|
if isinstance(video, Movie) and (self.num_cds > 1 or 'cd' in guess_filename):
|
|
|
|
# reduce score of subtitles for multi-disc movie releases
|
|
|
|
return set()
|
|
|
|
|
2019-04-07 23:31:39 +00:00
|
|
|
return matches
|
|
|
|
|
|
|
|
|
|
|
|
class SubsSabBzProvider(Provider):
|
|
|
|
"""SubsSabBz Provider."""
|
2020-05-01 16:18:16 +00:00
|
|
|
languages = {Language(l) for l in [
|
2019-04-07 23:31:39 +00:00
|
|
|
'bul', 'eng'
|
|
|
|
]}
|
|
|
|
|
|
|
|
def initialize(self):
|
|
|
|
self.session = Session()
|
|
|
|
self.session.headers['User-Agent'] = AGENT_LIST[randint(0, len(AGENT_LIST) - 1)]
|
|
|
|
self.session.headers["Accept"] = "text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8"
|
|
|
|
self.session.headers["Accept-Language"] = "en-US,en;q=0.5"
|
|
|
|
self.session.headers["Accept-Encoding"] = "gzip, deflate, br"
|
|
|
|
self.session.headers["DNT"] = "1"
|
|
|
|
self.session.headers["Connection"] = "keep-alive"
|
|
|
|
self.session.headers["Upgrade-Insecure-Requests"] = "1"
|
|
|
|
self.session.headers["Cache-Control"] = "max-age=0"
|
|
|
|
|
|
|
|
def terminate(self):
|
|
|
|
self.session.close()
|
|
|
|
|
|
|
|
def query(self, language, video):
|
|
|
|
subtitles = []
|
|
|
|
isEpisode = isinstance(video, Episode)
|
|
|
|
|
|
|
|
params = {
|
|
|
|
'act': 'search',
|
|
|
|
'movie': '',
|
|
|
|
'select-language': '2',
|
|
|
|
'upldr': '',
|
|
|
|
'yr': '',
|
|
|
|
'release': ''
|
|
|
|
}
|
|
|
|
|
|
|
|
if isEpisode:
|
2020-03-20 07:56:18 +00:00
|
|
|
params['movie'] = "%s %02d %02d" % (sanitize(fix_tv_naming(video.series), {'\''}), video.season, video.episode)
|
2019-04-07 23:31:39 +00:00
|
|
|
else:
|
|
|
|
params['yr'] = video.year
|
2020-09-13 11:24:49 +00:00
|
|
|
params['movie'] = sanitize(fix_movie_naming(video.title), {'\''})
|
2019-04-07 23:31:39 +00:00
|
|
|
|
|
|
|
if language == 'en' or language == 'eng':
|
|
|
|
params['select-language'] = 1
|
|
|
|
|
|
|
|
logger.info('Searching subtitle %r', params)
|
|
|
|
response = self.session.post('http://subs.sab.bz/index.php?', params=params, allow_redirects=False, timeout=10, headers={
|
|
|
|
'Referer': 'http://subs.sab.bz/',
|
|
|
|
})
|
|
|
|
|
|
|
|
response.raise_for_status()
|
|
|
|
|
|
|
|
if response.status_code != 200:
|
|
|
|
logger.debug('No subtitles found')
|
|
|
|
return subtitles
|
|
|
|
|
2020-03-22 07:58:31 +00:00
|
|
|
soup = BeautifulSoup(response.content, 'lxml')
|
2019-04-07 23:31:39 +00:00
|
|
|
rows = soup.findAll('tr', {'class': 'subs-row'})
|
|
|
|
|
2020-04-25 17:24:58 +00:00
|
|
|
# Search on first 25 rows only
|
|
|
|
for row in rows[:25]:
|
2019-04-07 23:31:39 +00:00
|
|
|
a_element_wrapper = row.find('td', { 'class': 'c2field' })
|
|
|
|
if a_element_wrapper:
|
|
|
|
element = a_element_wrapper.find('a')
|
|
|
|
if element:
|
|
|
|
link = element.get('href')
|
2020-08-16 08:01:21 +00:00
|
|
|
notes = re.sub(r'ddrivetip\(\'<div.*/></div>(.*)\',\'#[0-9]+\'\)', r'\1', element.get('onmouseover'))
|
2020-04-25 17:24:58 +00:00
|
|
|
title = element.get_text()
|
|
|
|
|
|
|
|
try:
|
|
|
|
year = int(str(element.next_sibling).strip(' ()'))
|
|
|
|
except:
|
|
|
|
year = None
|
|
|
|
|
|
|
|
td = row.findAll('td')
|
|
|
|
|
|
|
|
try:
|
|
|
|
num_cds = int(td[6].get_text())
|
|
|
|
except:
|
|
|
|
num_cds = None
|
|
|
|
|
|
|
|
try:
|
|
|
|
fps = float(td[7].get_text())
|
|
|
|
except:
|
|
|
|
fps = None
|
|
|
|
|
|
|
|
try:
|
|
|
|
uploader = td[8].get_text()
|
|
|
|
except:
|
|
|
|
uploader = None
|
|
|
|
|
|
|
|
try:
|
|
|
|
imdb_id = re.findall(r'imdb.com/title/(tt\d+)/?$', td[9].find('a').get('href'))[0]
|
|
|
|
except:
|
|
|
|
imdb_id = None
|
|
|
|
|
2019-04-07 23:31:39 +00:00
|
|
|
logger.info('Found subtitle link %r', link)
|
2020-04-25 17:24:58 +00:00
|
|
|
sub = self.download_archive_and_add_subtitle_files(link, language, video, fps, num_cds)
|
|
|
|
for s in sub:
|
|
|
|
s.title = title
|
|
|
|
s.notes = notes
|
|
|
|
s.year = year
|
2020-03-22 07:58:31 +00:00
|
|
|
s.uploader = uploader
|
2020-04-25 17:24:58 +00:00
|
|
|
s.imdb_id = imdb_id
|
2020-05-24 10:02:34 +00:00
|
|
|
s.single_file = True if len(sub) == 1 and num_cds == 1 else False
|
2020-03-28 08:10:24 +00:00
|
|
|
subtitles = subtitles + sub
|
2019-04-07 23:31:39 +00:00
|
|
|
return subtitles
|
|
|
|
|
|
|
|
def list_subtitles(self, video, languages):
|
|
|
|
return [s for l in languages for s in self.query(l, video)]
|
|
|
|
|
|
|
|
def download_subtitle(self, subtitle):
|
2019-04-08 17:58:13 +00:00
|
|
|
if subtitle.content:
|
|
|
|
pass
|
|
|
|
else:
|
|
|
|
seeking_subtitle_file = subtitle.filename
|
2020-04-25 17:24:58 +00:00
|
|
|
arch = self.download_archive_and_add_subtitle_files(subtitle.page_link, subtitle.language, subtitle.video,
|
|
|
|
subtitle.fps, subtitle.num_cds)
|
2019-04-08 17:58:13 +00:00
|
|
|
for s in arch:
|
|
|
|
if s.filename == seeking_subtitle_file:
|
|
|
|
subtitle.content = s.content
|
2019-04-07 23:31:39 +00:00
|
|
|
|
2020-04-25 17:24:58 +00:00
|
|
|
def process_archive_subtitle_files(self, archiveStream, language, video, link, fps, num_cds):
|
2019-04-07 23:31:39 +00:00
|
|
|
subtitles = []
|
|
|
|
type = 'episode' if isinstance(video, Episode) else 'movie'
|
2020-05-01 16:18:16 +00:00
|
|
|
for file_name in sorted(archiveStream.namelist()):
|
2019-04-07 23:31:39 +00:00
|
|
|
if file_name.lower().endswith(('.srt', '.sub')):
|
|
|
|
logger.info('Found subtitle file %r', file_name)
|
2020-04-25 17:24:58 +00:00
|
|
|
subtitle = SubsSabBzSubtitle(language, file_name, type, video, link, fps, num_cds)
|
2020-05-01 16:18:16 +00:00
|
|
|
subtitle.content = fix_line_ending(archiveStream.read(file_name))
|
2019-04-07 23:31:39 +00:00
|
|
|
subtitles.append(subtitle)
|
|
|
|
return subtitles
|
|
|
|
|
2020-04-25 17:24:58 +00:00
|
|
|
def download_archive_and_add_subtitle_files(self, link, language, video, fps, num_cds):
|
2019-04-07 23:31:39 +00:00
|
|
|
logger.info('Downloading subtitle %r', link)
|
2020-05-24 10:02:34 +00:00
|
|
|
cache_key = sha1(link.encode("utf-8")).digest()
|
|
|
|
request = region.get(cache_key)
|
|
|
|
if request is NO_VALUE:
|
|
|
|
request = self.session.get(link, headers={
|
|
|
|
'Referer': 'http://subs.sab.bz/index.php?'
|
|
|
|
})
|
|
|
|
request.raise_for_status()
|
|
|
|
region.set(cache_key, request)
|
|
|
|
else:
|
|
|
|
logger.info('Cache file: %s', codecs.encode(cache_key, 'hex_codec').decode('utf-8'))
|
2019-04-07 23:31:39 +00:00
|
|
|
|
2020-08-16 08:01:21 +00:00
|
|
|
try:
|
|
|
|
archive_stream = io.BytesIO(request.content)
|
|
|
|
if is_rarfile(archive_stream):
|
|
|
|
return self.process_archive_subtitle_files(RarFile(archive_stream), language, video, link, fps, num_cds)
|
|
|
|
elif is_zipfile(archive_stream):
|
|
|
|
return self.process_archive_subtitle_files(ZipFile(archive_stream), language, video, link, fps, num_cds)
|
|
|
|
except:
|
|
|
|
pass
|
|
|
|
|
|
|
|
logger.error('Ignore unsupported archive %r', request.headers)
|
|
|
|
region.delete(cache_key)
|
|
|
|
return []
|