mirror of
https://github.com/evilhero/mylar
synced 2025-03-19 10:05:25 +00:00
FIX:(#1468)(#1467)(#1465) Fixed problems with retrieving pullist when using alt_pull 0 method (the default method)
This commit is contained in:
parent
8e70faa2da
commit
8ad389fd8f
1 changed files with 25 additions and 10 deletions
|
@ -54,8 +54,6 @@ def pullit(forcecheck=None):
|
||||||
if pulldate is None: pulldate = '00000000'
|
if pulldate is None: pulldate = '00000000'
|
||||||
|
|
||||||
#only for pw-file or ALT_PULL = 1
|
#only for pw-file or ALT_PULL = 1
|
||||||
#PULLURL = 'http://www.previewsworld.com/shipping/prevues/newreleases.txt'
|
|
||||||
PULLURL = 'http://www.previewsworld.com/shipping/newreleases.txt'
|
|
||||||
newrl = os.path.join(mylar.CACHE_DIR, 'newreleases.txt')
|
newrl = os.path.join(mylar.CACHE_DIR, 'newreleases.txt')
|
||||||
mylar.PULLBYFILE = None
|
mylar.PULLBYFILE = None
|
||||||
|
|
||||||
|
@ -76,12 +74,10 @@ def pullit(forcecheck=None):
|
||||||
|
|
||||||
else:
|
else:
|
||||||
logger.info('[PULL-LIST] Unable to retrieve weekly pull-list. Dropping down to legacy method of PW-file')
|
logger.info('[PULL-LIST] Unable to retrieve weekly pull-list. Dropping down to legacy method of PW-file')
|
||||||
urllib.urlretrieve(PULLURL, newrl)
|
mylar.PULLBYFILE = pull_the_file(newrl)
|
||||||
mylar.PULLBYFILE = True
|
|
||||||
else:
|
else:
|
||||||
logger.info('[PULL-LIST] Populating & Loading pull-list data from file')
|
logger.info('[PULL-LIST] Populating & Loading pull-list data from file')
|
||||||
urllib.urlretrieve(PULLURL, newrl)
|
mylar.PULLBYFILE = pull_the_file(newrl)
|
||||||
mylar.PULLBYFILE = True
|
|
||||||
|
|
||||||
#set newrl to a manual file to pull in against that particular file
|
#set newrl to a manual file to pull in against that particular file
|
||||||
#newrl = '/mylar/tmp/newreleases.txt'
|
#newrl = '/mylar/tmp/newreleases.txt'
|
||||||
|
@ -100,9 +96,6 @@ def pullit(forcecheck=None):
|
||||||
else:
|
else:
|
||||||
pass
|
pass
|
||||||
|
|
||||||
#PULLURL = 'http://www.previewsworld.com/shipping/prevues/newreleases.txt'
|
|
||||||
PULLURL = 'http://www.previewsworld.com/shipping/newreleases.txt'
|
|
||||||
|
|
||||||
#Prepare the Substitute name switch for pulllist to comic vine conversion
|
#Prepare the Substitute name switch for pulllist to comic vine conversion
|
||||||
substitutes = os.path.join(mylar.DATA_DIR, "substitutes.csv")
|
substitutes = os.path.join(mylar.DATA_DIR, "substitutes.csv")
|
||||||
if not os.path.exists(substitutes):
|
if not os.path.exists(substitutes):
|
||||||
|
@ -966,7 +959,11 @@ def new_pullcheck(weeknumber, pullyear, comic1off_name=None, comic1off_id=None,
|
||||||
#if it's a name metch, it means that CV hasn't been populated yet with the necessary data
|
#if it's a name metch, it means that CV hasn't been populated yet with the necessary data
|
||||||
#do a quick issue check to see if the next issue number is in sequence and not a #1, or like #900
|
#do a quick issue check to see if the next issue number is in sequence and not a #1, or like #900
|
||||||
latestiss = namematch[0]['latestIssue'].strip()
|
latestiss = namematch[0]['latestIssue'].strip()
|
||||||
diff = int(week['Issue']) - int(latestiss)
|
try:
|
||||||
|
diff = int(week['Issue']) - int(latestiss)
|
||||||
|
except ValueError as e:
|
||||||
|
logger.warn('[WEEKLY-PULL] Invalid issue number detected. Skipping this entry for the time being.')
|
||||||
|
continue
|
||||||
if diff >= 0 and diff < 3:
|
if diff >= 0 and diff < 3:
|
||||||
comicname = namematch[0]['ComicName'].strip()
|
comicname = namematch[0]['ComicName'].strip()
|
||||||
comicid = namematch[0]['ComicID'].strip()
|
comicid = namematch[0]['ComicID'].strip()
|
||||||
|
@ -1226,6 +1223,24 @@ def checkthis(datecheck, datestatus, usedate):
|
||||||
|
|
||||||
return valid_check
|
return valid_check
|
||||||
|
|
||||||
|
def pull_the_file(newrl):
|
||||||
|
import requests
|
||||||
|
PULLURL = 'https://www.previewsworld.com/shipping/newreleases.txt'
|
||||||
|
PULL_AGENT = {'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/42.0.2311.135 Safari/537.36 Edge/12.246'}
|
||||||
|
try:
|
||||||
|
r = requests.get(PULLURL, verify=False, headers=PULL_AGENT, stream=True)
|
||||||
|
except requests.exceptions.RequestException as e:
|
||||||
|
logger.warn(e)
|
||||||
|
return False
|
||||||
|
|
||||||
|
with open(newrl, 'wb') as f:
|
||||||
|
for chunk in r.iter_content(chunk_size=1024):
|
||||||
|
if chunk: # filter out keep-alive new chunks
|
||||||
|
f.write(chunk)
|
||||||
|
f.flush()
|
||||||
|
|
||||||
|
return True
|
||||||
|
|
||||||
def weekly_check(comicid, issuenum, file=None, path=None, module=None, issueid=None):
|
def weekly_check(comicid, issuenum, file=None, path=None, module=None, issueid=None):
|
||||||
|
|
||||||
if module is None:
|
if module is None:
|
||||||
|
|
Loading…
Add table
Reference in a new issue