summaryrefslogtreecommitdiffstats
path: root/module/plugins
diff options
context:
space:
mode:
authorGravatar Walter Purcaro <vuolter@gmail.com> 2015-01-28 01:18:23 +0100
committerGravatar Walter Purcaro <vuolter@gmail.com> 2015-01-28 01:18:23 +0100
commite704a6cdd07c01a2f79eb851d7f2ecdf78da77fd (patch)
tree5abd1bb4599374afa7844c458984fbedd615be66 /module/plugins
parent[NetloadIn] Fix missing RECAPTCHA_KEY (diff)
downloadpyload-e704a6cdd07c01a2f79eb851d7f2ecdf78da77fd.tar.xz
[UploadedTo] Rewritten
Diffstat (limited to 'module/plugins')
-rw-r--r--module/plugins/hoster/UploadedTo.py263
-rw-r--r--module/plugins/internal/SimpleHoster.py6
2 files changed, 76 insertions, 193 deletions
diff --git a/module/plugins/hoster/UploadedTo.py b/module/plugins/hoster/UploadedTo.py
index d5c58fa6f..227e5e0d7 100644
--- a/module/plugins/hoster/UploadedTo.py
+++ b/module/plugins/hoster/UploadedTo.py
@@ -9,237 +9,120 @@ import re
from time import sleep
from module.network.RequestFactory import getURL
-from module.plugins.Hoster import Hoster
-from module.plugins.Plugin import chunks
from module.plugins.internal.CaptchaService import ReCaptcha
-from module.utils import html_unescape, parseFileSize
+from module.plugins.internal.SimpleHoster import SimpleHoster, create_getInfo
-key = "bGhGMkllZXByd2VEZnU5Y2NXbHhYVlZ5cEE1bkEzRUw=".decode('base64')
-
-
-def getID(url):
- """ returns id from file url"""
- m = re.match(UploadedTo.__pattern__, url)
- return m.group('ID')
-
-
-def getAPIData(urls):
- post = {"apikey": key}
-
- idMap = {}
-
- for i, url in enumerate(urls):
- id = getID(url)
- post['id_%s' % i] = id
- idMap[id] = url
-
- for _i in xrange(5):
- api = unicode(getURL("http://uploaded.net/api/filemultiple", post=post, decode=False), 'iso-8859-1')
- if api != "can't find request":
- break
- else:
- sleep(3)
-
- result = {}
-
- if api:
- for line in api.splitlines():
- data = line.split(",", 4)
- if data[1] in idMap:
- result[data[1]] = (data[0], data[2], data[4], data[3], idMap[data[1]])
-
- return result
-
-
-def parseFileInfo(self, url='', html=''):
- if not html and hasattr(self, "html"):
- html = self.html
+class UploadedTo(SimpleHoster):
+ __name__ = "UploadedTo"
+ __type__ = "hoster"
+ __version__ = "0.77"
- name = url
- size = 0
- fileid = None
+ __pattern__ = r'https?://(?:www\.)?(uploaded\.(to|net)|ul\.to)(/file/|/?\?id=|.*?&id=|/)(?P<ID>\w+)'
- if re.search(self.OFFLINE_PATTERN, html):
- # File offline
- status = 1
- else:
- m = re.search(self.INFO_PATTERN, html)
- if m:
- name, fileid = html_unescape(m.group('N')), m.group('ID')
- size = parseFileSize(m.group('S'))
- status = 2
- else:
- status = 3
+ __description__ = """Uploaded.net hoster plugin"""
+ __license__ = "GPLv3"
+ __authors__ = [("Walter Purcaro", "vuolter@gmail.com")]
- return name, size, status, fileid
+ API_KEY = "bGhGMkllZXByd2VEZnU5Y2NXbHhYVlZ5cEE1bkEzRUw=" #@NOTE: base64 encoded
-def getInfo(urls):
- for chunk in chunks(urls, 80):
- result = []
+ URL_REPLACEMENTS = [(__pattern__ + ".*", r'http://uploaded.net/file/\g<ID>')]
- api = getAPIData(chunk)
+ INFO_PATTERN = r'<a href="file/(?P<ID>\w+)" id="filename">(?P<N>[^<]+)</a> &nbsp;\s*<small[^>]*>(?P<S>[^<]+)</small>'
+ OFFLINE_PATTERN = r'<small class="cL">Error: 404'
- for data in api.itervalues():
- if data[0] == "online":
- result.append((html_unescape(data[2]), data[1], 2, data[4]))
+ LINK_PREMIUM_PATTERN = r'<div class="tfree".*\s*<form method="post" action="(.+?)"'
- elif data[0] == "offline":
- result.append((data[4], 0, 1, data[4]))
+ DL_LIMIT_ERROR = r'You have reached the max. number of possible free downloads for this hour'
- yield result
+ @classmethod
+ def apiInfo(cls, url="", get={}, post={}):
+ info = super(UploadedTo, cls).apiInfo(url)
-class UploadedTo(Hoster):
- __name__ = "UploadedTo"
- __type__ = "hoster"
- __version__ = "0.76"
+ for _i in xrange(5):
+ api = getURL("http://uploaded.net/api/filemultiple",
+ post={"apikey": cls.API_KEY.decode('base64'), 'id_0': re.match(cls.__pattern__, url).group('ID')},
+ decode=True)
+ if api != "can't find request":
+ api = api.splitlines()[0].split(",", 4)
- __pattern__ = r'https?://(?:www\.)?(uploaded\.(to|net)|ul\.to)(/file/|/?\?id=|.*?&id=|/)(?P<ID>\w+)'
+ if api[0] == "online":
+ info.update({'name': api[4], 'size': api[2], 'status': 2})
- __description__ = """Uploaded.net hoster plugin"""
- __license__ = "GPLv3"
- __authors__ = [("spoob", "spoob@pyload.org"),
- ("mkaay", "mkaay@mkaay.de"),
- ("zoidberg", "zoidberg@mujmail.cz"),
- ("netpok", "netpok@gmail.com"),
- ("stickell", "l.stickell@yahoo.it")]
+ elif api[0] == "offline":
+ info['status'] = 1
+ break
+ else:
+ sleep(3)
- INFO_PATTERN = r'<a href="file/(?P<ID>\w+)" id="filename">(?P<N>[^<]+)</a> &nbsp;\s*<small[^>]*>(?P<S>[^<]+)</small>'
- OFFLINE_PATTERN = r'<small class="cL">Error: 404</small>'
- DL_LIMIT_PATTERN = r'You have reached the max. number of possible free downloads for this hour'
+ return info
def setup(self):
self.multiDL = self.resumeDownload = self.premium
self.chunkLimit = 1 # critical problems with more chunks
-
- self.fileID = getID(self.pyfile.url)
- self.pyfile.url = "http://uploaded.net/file/%s" % self.fileID
-
-
- def process(self, pyfile):
self.load("http://uploaded.net/language/en", just_header=True)
- api = getAPIData([pyfile.url])
- # TODO: fallback to parse from site, because api sometimes delivers wrong status codes
+ def checkErrors(self):
+ if 'var free_enabled = false;' in self.html:
+ self.logError(_("Free-download capacities exhausted"))
+ self.retry(24, 5 * 60)
- if not api:
- self.logWarning(_("No response for API call"))
+ m = re.search(r"Current waiting period: <span>(\d+)</span> seconds", self.html)
+ if m:
+ self.wait(m.group(1))
+ else:
+ self.fail(_("File not downloadable for free users"))
- self.html = unicode(self.load(pyfile.url, decode=False), 'iso-8859-1')
- name, size, status, self.fileID = parseFileInfo(self)
- self.logDebug(name, size, status, self.fileID)
- if status == 1:
- self.offline()
- elif status == 2:
- pyfile.name, pyfile.size = name, size
- else:
- self.error(_("file info"))
+ if "limit-size" in self.html:
+ self.fail(_("File too big for free download"))
- elif api == 'Access denied':
- self.fail(_("API key invalid"))
+ elif "limit-slot" in self.html: # Temporary restriction so just wait a bit
+ self.wait(30 * 60, True)
+ self.retry()
- else:
- if self.fileID not in api:
- self.offline()
+ elif "limit-parallel" in self.html:
+ self.fail(_("Cannot download in parallel"))
- self.data = api[self.fileID]
- if self.data[0] != "online":
- self.offline()
+ elif "limit-dl" in self.html or self.DL_LIMIT_ERROR in self.html: # limit-dl
+ self.wait(3 * 60 * 60, True)
+ self.retry()
- pyfile.name = html_unescape(self.data[2])
+ elif '"err":"captcha"' in self.html:
+ self.invalidCaptcha()
- # pyfile.name = self.get_file_name()
- if self.premium:
- self.handlePremium()
- else:
- self.handleFree()
-
-
- def handlePremium(self):
- info = self.account.getAccountInfo(self.user, True)
- self.logDebug("%(name)s: Use Premium Account (%(left)sGB left)" % {"name": self.__name__,
- "left": info['trafficleft'] / 1024 / 1024})
- if int(self.data[1]) / 1024 > info['trafficleft']:
- self.logInfo(_("Not enough traffic left"))
- self.account.empty(self.user)
- self.resetAccount()
- self.fail(_("Traffic exceeded"))
-
- header = self.load("http://uploaded.net/file/%s" % self.fileID, just_header=True)
- if 'location' in header:
- #Direct download
- self.logDebug("Direct download link detected")
- self.download(header['location'])
- else:
- #Indirect download
- self.html = self.load("http://uploaded.net/file/%s" % self.fileID)
- m = re.search(r'<div class="tfree".*\s*<form method="post" action="(.*?)"', self.html)
- if m is None:
- self.fail(_("Download URL not m. Try to enable direct downloads"))
- url = m.group(1)
- self.download(url, post={})
+ def handleFree(self, pyfile):
+ self.html = self.load("http://uploaded.net/js/download.js", decode=True)
+ recaptcha = ReCaptcha(self)
+ response, challenge = recaptcha.challenge()
- def handleFree(self):
- self.html = self.load(self.pyfile.url, decode=True)
+ self.html = self.load("http://uploaded.net/io/ticket/captcha/%s" % self.info['pattern']['ID'],
+ post={'recaptcha_challenge_field': challenge,
+ 'recaptcha_response_field' : response})
- if 'var free_enabled = false;' in self.html:
- self.logError(_("Free-download capacities exhausted"))
- self.retry(24, 5 * 60)
+ if "type:'download'" in self.html:
+ self.correctCaptcha()
+ try:
+ self.link = re.search("url:'([^']+)", self.html).group(1)
- m = re.search(r"Current waiting period: <span>(\d+)</span> seconds", self.html)
- if m is None:
- self.fail(_("File not downloadable for free users"))
- self.setWait(int(m.group(1)))
+ except Exception:
+ pass
- self.html = self.load("http://uploaded.net/js/download.js", decode=True)
+ self.checkErrors()
- url = "http://uploaded.net/io/ticket/captcha/%s" % self.fileID
- downloadURL = ""
- recaptcha = ReCaptcha(self)
+ def checkFile():
+ if self.checkDownload({'limit-dl': self.DL_LIMIT_ERROR}):
+ self.wait(3 * 60 * 60, True)
+ self.retry()
- for _i in xrange(5):
- response, challenge = recaptcha.challenge()
- options = {"recaptcha_challenge_field": challenge, "recaptcha_response_field": response}
- self.wait()
-
- result = self.load(url, post=options)
- self.logDebug("Result: %s" % result)
-
- if "limit-size" in result:
- self.fail(_("File too big for free download"))
- elif "limit-slot" in result: # Temporary restriction so just wait a bit
- self.setWait(30 * 60, True)
- self.wait()
- self.retry()
- elif "limit-parallel" in result:
- self.fail(_("Cannot download in parallel"))
- elif "limit-dl" in result or self.DL_LIMIT_PATTERN in result: # limit-dl
- self.setWait(3 * 60 * 60, True)
- self.wait()
- self.retry()
- elif '"err":"captcha"' in result:
- self.invalidCaptcha()
- elif "type:'download'" in result:
- self.correctCaptcha()
- downloadURL = re.search("url:'([^']+)", result).group(1)
- break
- else:
- self.error(_("Unknown error: %s") % result)
+ return super(UploadedTo, self).checkFile()
- if not downloadURL:
- self.fail(_("No Download url retrieved/all captcha attempts failed"))
- self.download(downloadURL, disposition=True)
- check = self.checkDownload({"limit-dl": self.DL_LIMIT_PATTERN})
- if check == "limit-dl":
- self.setWait(3 * 60 * 60, True)
- self.wait()
- self.retry()
+getInfo = create_getInfo(UploadedTo)
diff --git a/module/plugins/internal/SimpleHoster.py b/module/plugins/internal/SimpleHoster.py
index 5c9d993e7..30add3f41 100644
--- a/module/plugins/internal/SimpleHoster.py
+++ b/module/plugins/internal/SimpleHoster.py
@@ -245,7 +245,7 @@ def secondsToMidnight(gmt=0):
class SimpleHoster(Hoster):
__name__ = "SimpleHoster"
__type__ = "hoster"
- __version__ = "1.10"
+ __version__ = "1.11"
__pattern__ = r'^unmatchable$'
@@ -333,7 +333,7 @@ class SimpleHoster(Hoster):
@classmethod
def getInfo(cls, url="", html=""):
info = cls.apiInfo(url)
- online = False
+ online = False if info['status'] != 2 else True
try:
info['pattern'] = re.match(cls.__pattern__, url).groupdict() #: pattern groups will be saved here
@@ -341,7 +341,7 @@ class SimpleHoster(Hoster):
except Exception:
info['pattern'] = {}
- if not html:
+ if not html and not online:
if not url:
info['error'] = "missing url"
info['status'] = 1