1
0
Fork 0
mirror of https://github.com/morpheus65535/bazarr synced 2024-12-23 00:03:33 +00:00
bazarr/custom_libs/subliminal_patch/pitcher.py

273 lines
8.9 KiB
Python
Raw Normal View History

2019-04-06 12:26:42 +00:00
# coding=utf-8
2019-09-17 02:04:27 +00:00
from __future__ import absolute_import
2019-04-07 23:31:39 +00:00
import os
2019-04-06 12:26:42 +00:00
import time
import logging
import json
2019-04-07 23:31:39 +00:00
from subliminal.cache import region
from dogpile.cache.api import NO_VALUE
from python_anticaptcha import AnticaptchaClient, NoCaptchaTaskProxylessTask, NoCaptchaTask, AnticaptchaException
2019-09-17 02:04:27 +00:00
from deathbycaptcha import SocketClient as DBCClient, DEFAULT_TOKEN_TIMEOUT
import six
from six.moves import range
from urllib import parse
from subliminal.exceptions import ConfigurationError
2019-04-06 12:26:42 +00:00
logger = logging.getLogger(__name__)
class PitcherRegistry(object):
2019-04-07 23:31:39 +00:00
pitchers = []
pitchers_by_key = {}
2019-04-06 12:26:42 +00:00
def register(self, cls):
2019-04-07 23:31:39 +00:00
idx = len(self.pitchers)
self.pitchers.append(cls)
key = "%s_%s" % (cls.name, cls.needs_proxy)
key_by_source = "%s_%s" % (cls.source, cls.needs_proxy)
self.pitchers_by_key[key] = idx
self.pitchers_by_key[key_by_source] = idx
2019-04-06 12:26:42 +00:00
return cls
2019-04-07 23:31:39 +00:00
def get_pitcher(self, name_or_site=None, with_proxy=False):
name_or_site = name_or_site or os.environ.get("ANTICAPTCHA_CLASS")
if not name_or_site:
raise ConfigurationError("AntiCaptcha class not given, exiting")
2019-04-07 23:31:39 +00:00
key = "%s_%s" % (name_or_site, with_proxy)
if key not in self.pitchers_by_key:
raise Exception("Pitcher %s not found (proxy: %s)" % (name_or_site, with_proxy))
return self.pitchers[self.pitchers_by_key.get(key)]
2019-04-06 12:26:42 +00:00
registry = pitchers = PitcherRegistry()
class Pitcher(object):
name = None
2019-04-07 23:31:39 +00:00
source = None
needs_proxy = False
2019-04-06 12:26:42 +00:00
tries = 3
job = None
client = None
2019-04-07 23:31:39 +00:00
client_key = None
2019-04-06 12:26:42 +00:00
website_url = None
website_key = None
website_name = None
solve_time = None
success = False
2019-04-07 23:31:39 +00:00
def __init__(self, website_name, website_url, website_key, tries=3, client_key=None, *args, **kwargs):
2019-04-06 12:26:42 +00:00
self.tries = tries
2019-04-07 23:31:39 +00:00
self.client_key = client_key or os.environ.get("ANTICAPTCHA_ACCOUNT_KEY")
if not self.client_key:
raise Exception("AntiCaptcha key not given, exiting")
2019-04-06 12:26:42 +00:00
self.website_name = website_name
self.website_key = website_key
self.website_url = website_url
self.success = False
self.solve_time = None
def get_client(self):
raise NotImplementedError
def get_job(self):
raise NotImplementedError
def _throw(self):
self.client = self.get_client()
self.job = self.get_job()
def throw(self):
t = time.time()
data = self._throw()
if self.success:
self.solve_time = time.time() - t
logger.info("%s: Solving took %ss", self.website_name, int(self.solve_time))
return data
@registry.register
class AntiCaptchaProxyLessPitcher(Pitcher):
name = "AntiCaptchaProxyLess"
2019-04-07 23:31:39 +00:00
source = "anti-captcha.com"
2019-04-06 12:26:42 +00:00
host = "api.anti-captcha.com"
language_pool = "en"
2019-04-07 23:31:39 +00:00
tries = 5
2019-04-06 12:26:42 +00:00
use_ssl = True
is_invisible = False
2019-04-07 23:31:39 +00:00
def __init__(self, website_name, website_url, website_key, tries=3, host=None, language_pool=None,
2019-04-06 12:26:42 +00:00
use_ssl=True, is_invisible=False, *args, **kwargs):
super(AntiCaptchaProxyLessPitcher, self).__init__(website_name, website_url, website_key, tries=tries, *args,
**kwargs)
self.host = host or self.host
self.language_pool = language_pool or self.language_pool
self.use_ssl = use_ssl
self.is_invisible = is_invisible
def get_client(self):
return AnticaptchaClient(self.client_key, self.language_pool, self.host, self.use_ssl)
def get_job(self):
task = NoCaptchaTaskProxylessTask(website_url=self.website_url, website_key=self.website_key,
is_invisible=self.is_invisible)
return self.client.createTask(task)
def _throw(self):
for i in range(self.tries):
try:
super(AntiCaptchaProxyLessPitcher, self)._throw()
self.job.join()
ret = self.job.get_solution_response()
if ret:
self.success = True
return ret
except AnticaptchaException as e:
if i >= self.tries - 1:
logger.error("%s: Captcha solving finally failed. Exiting", self.website_name)
return
if e.error_code == 'ERROR_ZERO_BALANCE':
logger.error("%s: No balance left on captcha solving service. Exiting", self.website_name)
return
elif e.error_code == 'ERROR_NO_SLOT_AVAILABLE':
logger.info("%s: No captcha solving slot available, retrying", self.website_name)
time.sleep(5.0)
continue
elif e.error_code == 'ERROR_KEY_DOES_NOT_EXIST':
logger.error("%s: Bad AntiCaptcha API key", self.website_name)
return
elif e.error_id is None and e.error_code == 250:
# timeout
if i < self.tries:
logger.info("%s: Captcha solving timed out, retrying", self.website_name)
time.sleep(1.0)
continue
else:
logger.error("%s: Captcha solving timed out three times; bailing out", self.website_name)
return
raise
@registry.register
class AntiCaptchaPitcher(AntiCaptchaProxyLessPitcher):
name = "AntiCaptcha"
proxy = None
2019-04-07 23:31:39 +00:00
needs_proxy = True
2019-04-06 12:26:42 +00:00
user_agent = None
cookies = None
def __init__(self, *args, **kwargs):
self.proxy = self.parse_url(kwargs.pop("proxy"))
2019-04-06 12:26:42 +00:00
self.user_agent = kwargs.pop("user_agent")
cookies = kwargs.pop("cookies", {})
if isinstance(cookies, dict):
2019-09-17 02:04:27 +00:00
self.cookies = ";".join(["%s=%s" % (k, v) for k, v in six.iteritems(cookies)])
2019-04-06 12:26:42 +00:00
super(AntiCaptchaPitcher, self).__init__(*args, **kwargs)
@staticmethod
def parse_url(url):
parsed = parse.urlparse(url)
return dict(
proxy_type=parsed.scheme,
proxy_address=parsed.hostname,
proxy_port=parsed.port,
proxy_login=parsed.username,
proxy_password=parsed.password,
)
2019-04-06 12:26:42 +00:00
def get_job(self):
task = NoCaptchaTask(website_url=self.website_url, website_key=self.website_key, proxy=self.proxy,
user_agent=self.user_agent, cookies=self.cookies, is_invisible=self.is_invisible)
return self.client.createTask(task)
@registry.register
class DBCProxyLessPitcher(Pitcher):
name = "DeathByCaptchaProxyLess"
2019-04-07 23:31:39 +00:00
source = "deathbycaptcha.com"
2019-04-06 12:26:42 +00:00
username = None
password = None
2019-04-07 23:31:39 +00:00
def __init__(self, website_name, website_url, website_key,
2019-09-17 02:04:27 +00:00
timeout=DEFAULT_TOKEN_TIMEOUT, tries=3, *args, **kwargs):
2019-04-06 12:26:42 +00:00
super(DBCProxyLessPitcher, self).__init__(website_name, website_url, website_key, tries=tries)
2019-04-07 23:31:39 +00:00
self.username, self.password = self.client_key.split(":", 1)
2019-04-06 12:26:42 +00:00
self.timeout = timeout
def get_client(self):
return DBCClient(self.username, self.password)
def get_job(self):
pass
@property
def payload_dict(self):
return {
"googlekey": self.website_key,
"pageurl": self.website_url
}
def _throw(self):
super(DBCProxyLessPitcher, self)._throw()
payload = json.dumps(self.payload_dict)
2019-04-07 23:31:39 +00:00
for i in range(self.tries):
try:
#balance = self.client.get_balance()
data = self.client.decode(timeout=self.timeout, type=4, token_params=payload)
if data and data["is_correct"] and data["text"]:
self.success = True
return data["text"]
except:
raise
2019-04-06 12:26:42 +00:00
@registry.register
class DBCPitcher(DBCProxyLessPitcher):
2019-04-07 23:31:39 +00:00
name = "DeathByCaptcha"
2019-04-06 12:26:42 +00:00
proxy = None
2019-04-07 23:31:39 +00:00
needs_proxy = True
2019-04-06 12:26:42 +00:00
proxy_type = "HTTP"
def __init__(self, *args, **kwargs):
self.proxy = kwargs.pop("proxy")
super(DBCPitcher, self).__init__(*args, **kwargs)
@property
def payload_dict(self):
payload = super(DBCPitcher, self).payload_dict
payload.update({
"proxytype": self.proxy_type,
"proxy": self.proxy
})
return payload
2019-04-07 23:31:39 +00:00
def load_verification(site_name, session, callback=lambda x: None):
2019-09-19 01:50:20 +00:00
ccks = region.get("%s_data" % site_name, expiration_time=15552000) # 6m
2019-04-07 23:31:39 +00:00
if ccks != NO_VALUE:
cookies, user_agent = ccks
logger.debug("%s: Re-using previous user agent: %s", site_name.capitalize(), user_agent)
session.headers["User-Agent"] = user_agent
try:
session.cookies._cookies.update(cookies)
return callback(region)
except:
return False
return False
def store_verification(site_name, session):
2019-09-20 21:56:33 +00:00
region.set("%s_data" % site_name, (session.cookies._cookies, session.headers["User-Agent"]))