summaryrefslogtreecommitdiffstats
path: root/module/plugins/hoster/UploadedTo.py
diff options
context:
space:
mode:
Diffstat (limited to 'module/plugins/hoster/UploadedTo.py')
-rw-r--r--module/plugins/hoster/UploadedTo.py126
1 files changed, 70 insertions, 56 deletions
diff --git a/module/plugins/hoster/UploadedTo.py b/module/plugins/hoster/UploadedTo.py
index 6d17bcbd9..833468a80 100644
--- a/module/plugins/hoster/UploadedTo.py
+++ b/module/plugins/hoster/UploadedTo.py
@@ -1,18 +1,19 @@
# -*- coding: utf-8 -*-
-
-# Test links (random.bin):
+#
+# Test links:
# http://ul.to/044yug9o
# http://ul.to/gzfhd0xs
import re
-from time import sleep
-from module.utils import html_unescape, parseFileSize
+from time import sleep
-from module.plugins.Hoster import Hoster
from module.network.RequestFactory import getURL
+from module.plugins.Hoster import Hoster
from module.plugins.Plugin import chunks
from module.plugins.internal.CaptchaService import ReCaptcha
+from module.utils import html_unescape, parseFileSize
+
key = "bGhGMkllZXByd2VEZnU5Y2NXbHhYVlZ5cEE1bkEzRUw=".decode('base64')
@@ -30,10 +31,10 @@ def getAPIData(urls):
for i, url in enumerate(urls):
id = getID(url)
- post["id_%s" % i] = id
+ post['id_%s' % i] = id
idMap[id] = url
- for _ in xrange(5):
+ for _i in xrange(5):
api = unicode(getURL("http://uploaded.net/api/filemultiple", post=post, decode=False), 'iso-8859-1')
if api != "can't find request":
break
@@ -54,17 +55,22 @@ def getAPIData(urls):
def parseFileInfo(self, url='', html=''):
if not html and hasattr(self, "html"):
html = self.html
- name, size, status, found, fileid = url, 0, 3, None, None
- if re.search(self.FILE_OFFLINE_PATTERN, html):
+ name = url
+ size = 0
+ fileid = None
+
+ if re.search(self.OFFLINE_PATTERN, html):
# File offline
status = 1
else:
- found = re.search(self.FILE_INFO_PATTERN, html)
- if found:
- name, fileid = html_unescape(found.group('N')), found.group('ID')
- size = parseFileSize(found.group('S'))
+ m = re.search(self.INFO_PATTERN, html)
+ if m:
+ name, fileid = html_unescape(m.group('N')), m.group('ID')
+ size = parseFileSize(m.group('S'))
status = 2
+ else:
+ status = 3
return name, size, status, fileid
@@ -86,26 +92,34 @@ def getInfo(urls):
class UploadedTo(Hoster):
- __name__ = "UploadedTo"
- __type__ = "hoster"
+ __name__ = "UploadedTo"
+ __type__ = "hoster"
+ __version__ = "0.75"
+
__pattern__ = r'https?://(?:www\.)?(uploaded\.(to|net)|ul\.to)(/file/|/?\?id=|.*?&id=|/)(?P<ID>\w+)'
- __version__ = "0.72"
+
__description__ = """Uploaded.net hoster plugin"""
- __author_name__ = ("spoob", "mkaay", "zoidberg", "netpok", "stickell")
- __author_mail__ = ("spoob@pyload.org", "mkaay@mkaay.de", "zoidberg@mujmail.cz",
- "netpok@gmail.com", "l.stickell@yahoo.it")
+ __license__ = "GPLv3"
+ __authors__ = [("spoob", "spoob@pyload.org"),
+ ("mkaay", "mkaay@mkaay.de"),
+ ("zoidberg", "zoidberg@mujmail.cz"),
+ ("netpok", "netpok@gmail.com"),
+ ("stickell", "l.stickell@yahoo.it")]
+
+
+ INFO_PATTERN = r'<a href="file/(?P<ID>\w+)" id="filename">(?P<N>[^<]+)</a> &nbsp;\s*<small[^>]*>(?P<S>[^<]+)</small>'
+ OFFLINE_PATTERN = r'<small class="cL">Error: 404</small>'
+ DL_LIMIT_PATTERN = r'You have reached the max. number of possible free downloads for this hour'
- FILE_INFO_PATTERN = r'<a href="file/(?P<ID>\w+)" id="filename">(?P<N>[^<]+)</a> &nbsp;\s*<small[^>]*>(?P<S>[^<]+)</small>'
- FILE_OFFLINE_PATTERN = r'<small class="cL">Error: 404</small>'
- DL_LIMIT_PATTERN = "You have reached the max. number of possible free downloads for this hour"
def setup(self):
- self.multiDL = self.resumeDownload = self.premium
+ self.multiDL = self.resumeDownload = self.premium
self.chunkLimit = 1 # critical problems with more chunks
self.fileID = getID(self.pyfile.url)
self.pyfile.url = "http://uploaded.net/file/%s" % self.fileID
+
def process(self, pyfile):
self.load("http://uploaded.net/language/en", just_header=True)
@@ -114,7 +128,7 @@ class UploadedTo(Hoster):
# TODO: fallback to parse from site, because api sometimes delivers wrong status codes
if not api:
- self.logWarning("No response for API call")
+ self.logWarning(_("No response for API call"))
self.html = unicode(self.load(pyfile.url, decode=False), 'iso-8859-1')
name, size, status, self.fileID = parseFileInfo(self)
@@ -124,7 +138,8 @@ class UploadedTo(Hoster):
elif status == 2:
pyfile.name, pyfile.size = name, size
else:
- self.fail('Parse error - file info')
+ self.error(_("file info"))
+
elif api == 'Access denied':
self.fail(_("API key invalid"))
@@ -145,83 +160,82 @@ class UploadedTo(Hoster):
else:
self.handleFree()
+
def handlePremium(self):
info = self.account.getAccountInfo(self.user, True)
self.logDebug("%(name)s: Use Premium Account (%(left)sGB left)" % {"name": self.__name__,
- "left": info["trafficleft"] / 1024 / 1024})
- if int(self.data[1]) / 1024 > info["trafficleft"]:
- self.logInfo(_("%s: Not enough traffic left" % self.__name__))
+ "left": info['trafficleft'] / 1024 / 1024})
+ if int(self.data[1]) / 1024 > info['trafficleft']:
+ self.logInfo(_("Not enough traffic left"))
self.account.empty(self.user)
self.resetAccount()
self.fail(_("Traffic exceeded"))
header = self.load("http://uploaded.net/file/%s" % self.fileID, just_header=True)
- if "location" in header:
+ if 'location' in header:
#Direct download
- print "Direct Download: " + header['location']
+ self.logDebug("Direct download link detected")
self.download(header['location'])
else:
#Indirect download
self.html = self.load("http://uploaded.net/file/%s" % self.fileID)
- found = re.search(r'<div class="tfree".*\s*<form method="post" action="(.*?)"', self.html)
- if not found:
- self.fail("Download URL not found. Try to enable direct downloads.")
- url = found.group(1)
- print "Premium URL: " + url
+ m = re.search(r'<div class="tfree".*\s*<form method="post" action="(.*?)"', self.html)
+ if m is None:
+ self.fail(_("Download URL not m. Try to enable direct downloads"))
+ url = m.group(1)
self.download(url, post={})
+
def handleFree(self):
self.html = self.load(self.pyfile.url, decode=True)
if 'var free_enabled = false;' in self.html:
- self.logError("Free-download capacities exhausted.")
- self.retry(max_tries=24, wait_time=5 * 60)
+ self.logError(_("Free-download capacities exhausted"))
+ self.retry(24, 5 * 60)
- found = re.search(r"Current waiting period: <span>(\d+)</span> seconds", self.html)
- if not found:
- self.fail("File not downloadable for free users")
- self.setWait(int(found.group(1)))
+ m = re.search(r"Current waiting period: <span>(\d+)</span> seconds", self.html)
+ if m is None:
+ self.fail(_("File not downloadable for free users"))
+ self.setWait(int(m.group(1)))
- js = self.load("http://uploaded.net/js/download.js", decode=True)
-
- challengeId = re.search(r'Recaptcha\.create\("([^"]+)', js)
+ self.html = self.load("http://uploaded.net/js/download.js", decode=True)
url = "http://uploaded.net/io/ticket/captcha/%s" % self.fileID
downloadURL = ""
- for _ in xrange(5):
- re_captcha = ReCaptcha(self)
- challenge, result = re_captcha.challenge(challengeId.group(1))
- options = {"recaptcha_challenge_field": challenge, "recaptcha_response_field": result}
+ recaptcha = ReCaptcha(self)
+
+ for _i in xrange(5):
+ challenge, response = recaptcha.challenge()
+ options = {"recaptcha_challenge_field": challenge, "recaptcha_response_field": response}
self.wait()
result = self.load(url, post=options)
- self.logDebug("result: %s" % result)
+ self.logDebug("Result: %s" % result)
if "limit-size" in result:
- self.fail("File too big for free download")
+ self.fail(_("File too big for free download"))
elif "limit-slot" in result: # Temporary restriction so just wait a bit
self.setWait(30 * 60, True)
self.wait()
self.retry()
elif "limit-parallel" in result:
- self.fail("Cannot download in parallel")
- elif self.DL_LIMIT_PATTERN in result: # limit-dl
+ self.fail(_("Cannot download in parallel"))
+ elif "limit-dl" in result or self.DL_LIMIT_PATTERN in result: # limit-dl
self.setWait(3 * 60 * 60, True)
self.wait()
self.retry()
- elif 'err:"captcha"' in result:
- self.logError("ul.net captcha is disabled")
+ elif '"err":"captcha"' in result:
self.invalidCaptcha()
elif "type:'download'" in result:
self.correctCaptcha()
downloadURL = re.search("url:'([^']+)", result).group(1)
break
else:
- self.fail("Unknown error '%s'")
+ self.error(_("Unknown error: %s") % result)
if not downloadURL:
- self.fail("No Download url retrieved/all captcha attempts failed")
+ self.fail(_("No Download url retrieved/all captcha attempts failed"))
self.download(downloadURL, disposition=True)
check = self.checkDownload({"limit-dl": self.DL_LIMIT_PATTERN})