# -*- coding: utf-8 -*- """ This program is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation; either version 3 of the License, or (at your option) any later version. This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with this program; if not, see . """ import re from module.plugins.Hoster import Hoster from module.network.RequestFactory import getURL from module.plugins.internal.CaptchaService import ReCaptcha from module.common.json_layer import json_loads from module.utils import parseFileSize from module.plugins.Plugin import chunks def checkFile(plugin, urls): html = getURL(plugin.URLS[1], post={"urls": "\n".join(urls)}, decode=True) file_info = [] for li in re.finditer(plugin.LINKCHECK_TR, html, re.DOTALL): try: cols = re.findall(plugin.LINKCHECK_TD, li.group(1)) if cols: file_info.append(( cols[1] if cols[1] != '--' else cols[0], parseFileSize(cols[2]) if cols[2] != '--' else 0, 2 if cols[3].startswith('Available') else 1, cols[0])) except Exception, e: continue return file_info class FileserveCom(Hoster): __name__ = "FileserveCom" __type__ = "hoster" __pattern__ = r"http://(?:www\.)?fileserve\.com/file/(?P[^/]+).*" __version__ = "0.51" __description__ = """Fileserve.com hoster plugin""" __author_name__ = ("jeix", "mkaay", "Paul King", "zoidberg") __author_mail__ = ("jeix@hasnomail.de", "mkaay@mkaay.de", "", "zoidberg@mujmail.cz") URLS = ['http://www.fileserve.com/file/', 'http://www.fileserve.com/link-checker.php', 'http://www.fileserve.com/checkReCaptcha.php'] LINKCHECK_TR = r'\s*(http://www.fileserve\.com/file/.*?)' LINKCHECK_TD = r'(?:<[^>]*>| )*([^<]*)' CAPTCHA_KEY_PATTERN = r"var reCAPTCHA_publickey='(?P[^']+)'" LONG_WAIT_PATTERN = r'
  • You need to wait (\d+) (\w+) to start another download\.
  • ' LINK_EXPIRED_PATTERN = "Your download link has expired" DAILY_LIMIT_PATTERN = "Your daily download limit has been reached" NOT_LOGGED_IN_PATTERN = '
    Login' # shares code with FilejungleCom and UploadstationCom def setup(self): self.resumeDownload = self.multiDL = self.premium self.file_id = re.search(self.__pattern__, self.pyfile.url).group('id') self.url = "%s%s" % (self.URLS[0], self.file_id) self.logDebug("File ID: %s URL: %s" % (self.file_id, self.url)) def process(self, pyfile): pyfile.name, pyfile.size, status, self.url = checkFile(self, [self.url])[0] if status != 2: self.offline() self.logDebug("File Name: %s Size: %d" % (pyfile.name, pyfile.size)) if self.premium: self.handlePremium() else: self.handleFree() def handleFree(self): self.html = self.load(self.url) action = self.load(self.url, post={"checkDownload": "check"}, decode=True) action = json_loads(action) self.logDebug(action) if "fail" in action: if action["fail"] == "timeLimit": self.html = self.load(self.url,post={"checkDownload": "showError", "errorType": "timeLimit"}, decode=True) self.doLongWait(re.search(self.LONG_WAIT_PATTERN, self.html)) elif action["fail"] == "parallelDownload": self.logWarning(_("Parallel download error, now waiting 60s.")) self.retry(wait_time=60, reason="parallelDownload") else: self.fail("Download check returned %s" % action["fail"]) elif "success" in action: if action["success"] == "showCaptcha": self.doCaptcha() self.doTimmer() elif action["success"] == "showTimmer": self.doTimmer() else: self.fail("Unknown server response") # show download link response = self.load(self.url, post={"downloadLink": "show"}, decode=True) self.logDebug("show downloadLink response : %s" % response) if "fail" in response: self.fail("Couldn't retrieve download url") # this may either download our file or forward us to an error page self.download(self.url, post={"download": "normal"}) self.logDebug(self.req.http.lastEffectiveURL) check = self.checkDownload({"expired": self.LINK_EXPIRED_PATTERN, "wait": re.compile(self.LONG_WAIT_PATTERN), "limit": self.DAILY_LIMIT_PATTERN}) if check == "expired": self.logDebug("Download link was expired") self.retry() elif check == "wait": self.doLongWait(self.lastCheck) elif check == "limit": #download limited reached for today (not a exact time known) self.setWait(3 * 60 * 60, True) # wait 3 hours #TO-DO: resolve waittime using UnrestrictLi's secondsToMidnight self.wait() self.retry() self.thread.m.reconnecting.wait(3) # Ease issue with later downloads appearing to be in parallel def doTimmer(self): response = self.load(self.url, post={"downloadLink": "wait"}, decode=True) self.logDebug("wait response : %s" % response[:80]) if "fail" in response: self.fail("Failed getting wait time") if self.__name__ == "FilejungleCom": found = re.search(r'"waitTime":(\d+)', response) if not found: self.fail("Cannot get wait time") wait_time = int(found.group(1)) else: wait_time = int(response) + 3 self.setWait(wait_time) self.wait() def doCaptcha(self): captcha_key = re.search(self.CAPTCHA_KEY_PATTERN, self.html).group("key") recaptcha = ReCaptcha(self) for _ in xrange(5): challenge, code = recaptcha.challenge(captcha_key) response = json_loads(self.load(self.URLS[2], post={'recaptcha_challenge_field': challenge, 'recaptcha_response_field': code, 'recaptcha_shortencode_field': self.file_id})) self.logDebug("reCaptcha response : %s" % response) if not response["success"]: self.invalidCaptcha() else: self.correctCaptcha() break else: self.fail("Invalid captcha") def doLongWait(self, m): wait_time = (int(m.group(1)) * {'seconds': 1, 'minutes': 60, 'hours': 3600}[m.group(2)]) if m else 12 * 60 self.setWait(wait_time, True) self.wait() self.retry() def handlePremium(self): premium_url = None if self.__name__ == "FileserveCom": #try api download response = self.load("http://app.fileserve.com/api/download/premium/", post={"username": self.user, "password": self.account.getAccountData(self.user)["password"], "shorten": self.file_id}, decode=True) if response: response = json_loads(response) if response['error_code'] == "302": premium_url = response['next'] elif response['error_code'] in ["305", "500"]: self.tempOffline() elif response['error_code'] in ["403", "605"]: self.resetAccount() elif response['error_code'] in ["606", "607", "608"]: self.offline() else: self.logError(response['error_code'], response['error_message']) self.download(premium_url or self.pyfile.url) if not premium_url: check = self.checkDownload({"login": re.compile(self.NOT_LOGGED_IN_PATTERN)}) if check == "login": self.account.relogin(self.user) self.retry(reason=_("Not logged in.")) def getInfo(urls): for chunk in chunks(urls, 100): yield checkFile(FileserveCom, chunk)