diff options
Diffstat (limited to 'module/plugins/hoster/FileserveCom.py')
| -rw-r--r-- | module/plugins/hoster/FileserveCom.py | 211 | 
1 files changed, 0 insertions, 211 deletions
| diff --git a/module/plugins/hoster/FileserveCom.py b/module/plugins/hoster/FileserveCom.py deleted file mode 100644 index 04c297661..000000000 --- a/module/plugins/hoster/FileserveCom.py +++ /dev/null @@ -1,211 +0,0 @@ -# -*- coding: utf-8 -*- -""" -    This program is free software; you can redistribute it and/or modify -    it under the terms of the GNU General Public License as published by -    the Free Software Foundation; either version 3 of the License, -    or (at your option) any later version. - -    This program is distributed in the hope that it will be useful, -    but WITHOUT ANY WARRANTY; without even the implied warranty of -    MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. -    See the GNU General Public License for more details. - -    You should have received a copy of the GNU General Public License -    along with this program; if not, see <http://www.gnu.org/licenses/>. -""" - -import re -from module.plugins.Hoster import Hoster -from module.network.RequestFactory import getURL -from module.plugins.ReCaptcha import ReCaptcha -from module.common.json_layer import json_loads -from module.utils import parseFileSize -from module.plugins.Plugin import chunks - -def checkFile(plugin, urls): -    html = getURL(plugin.URLS[1], post = {"urls": "\n".join(urls)}, decode=True) - -    file_info = []     -    for li in re.finditer(plugin.LINKCHECK_TR, html, re.DOTALL): -        try: -            cols = re.findall(plugin.LINKCHECK_TD, li.group(1)) -            if cols: -                file_info.append(( -                    cols[1] if cols[1] != '--' else cols[0], -                    parseFileSize(cols[2]) if cols[2] != '--' else 0,  -                    2 if cols[3].startswith('Available') else 1, -                    cols[0])) -        except Exception, e: -            continue -     -    return file_info - -class FileserveCom(Hoster): -    __name__ = "FileserveCom" -    __type__ = "hoster" -    __pattern__ = r"http://(?:www\.)?fileserve\.com/file/(?P<id>[^/]+).*" -    __version__ = "0.51" -    __description__ = """Fileserve.Com File Download Hoster""" -    __author_name__ = ("jeix", "mkaay", "paul king", "zoidberg") -    __author_mail__ = ("jeix@hasnomail.de", "mkaay@mkaay.de", "", "zoidberg@mujmail.cz") -    -    URLS = ['http://www.fileserve.com/file/', 'http://www.fileserve.com/link-checker.php', 'http://www.fileserve.com/checkReCaptcha.php'] -    LINKCHECK_TR = r'<tr>\s*(<td>http://www.fileserve\.com/file/.*?)</tr>' -    LINKCHECK_TD = r'<td>(?:<[^>]*>| )*([^<]*)' -     -    CAPTCHA_KEY_PATTERN = r"var reCAPTCHA_publickey='(?P<key>[^']+)'" -    LONG_WAIT_PATTERN = r'<li class="title">You need to wait (\d+) (\w+) to start another download\.</li>' -    LINK_EXPIRED_PATTERN = "Your download link has expired" -    DAILY_LIMIT_PATTERN = "Your daily download limit has been reached" -    NOT_LOGGED_IN_PATTERN = '<form (name="loginDialogBoxForm"|id="login_form")|<li><a href="/login.php">Login</a></li>' -     -    # shares code with FilejungleCom and UploadstationCom -        -    def setup(self): -        self.resumeDownload = self.multiDL = True if self.premium else False -         -        self.file_id = re.search(self.__pattern__, self.pyfile.url).group('id') -        self.url = "%s%s" % (self.URLS[0], self.file_id) -        self.logDebug("File ID: %s URL: %s" % (self.file_id, self.url)) - -    def process(self, pyfile): -        pyfile.name, pyfile.size, status, self.url = checkFile(self, [self.url])[0]         -        if status != 2: self.offline() -        self.logDebug("File Name: %s Size: %d" % (pyfile.name, pyfile.size))  -         -        if self.premium: -            self.handlePremium() -        else: -            self.handleFree() -     -    def handleFree(self): -        self.html = self.load(self.url)              -        action = self.load(self.url, post={"checkDownload": "check"}, decode=True) -        action = json_loads(action) -        self.logDebug(action) -              -        if "fail" in action: -            if action["fail"] == "timeLimit": -                self.html = self.load(self.url, -                                 post={"checkDownload": "showError", -                                       "errorType": "timeLimit"}, -                                 decode=True) -                                  -                self.doLongWait(re.search(self.LONG_WAIT_PATTERN, self.html)) - -            elif action["fail"] == "parallelDownload": -                self.logWarning(_("Parallel download error, now waiting 60s.")) -                self.retry(wait_time=60, reason="parallelDownload") - -            else: -                self.fail("Download check returned %s" % action["fail"]) -         -        elif "success" in action:            -            if action["success"] == "showCaptcha": -                self.doCaptcha() -                self.doTimmer() -            elif action["success"] == "showTimmer": -                self.doTimmer() -         -        else: -            self.fail("Unknown server response")                                         -         -        # show download link -        response = self.load(self.url, post={"downloadLink": "show"}, decode=True) -        self.logDebug("show downloadLink response : %s" % response) -        if "fail" in response: -            self.fail("Couldn't retrieve download url") - -        # this may either download our file or forward us to an error page -        self.download(self.url, post = {"download": "normal"}) -        self.logDebug(self.req.http.lastEffectiveURL) - -        check = self.checkDownload({"expired": self.LINK_EXPIRED_PATTERN, -                                    "wait": re.compile(self.LONG_WAIT_PATTERN), -                                    "limit": self.DAILY_LIMIT_PATTERN}) - -        if check == "expired": -            self.logDebug("Download link was expired") -            self.retry() -        elif check == "wait": -            self.doLongWait(self.lastCheck) -        elif check == "limit": -            #download limited reached for today (not a exact time known) -            self.setWait(180 * 60, True) # wait 3 hours -            self.wait() -            self.retry(max_tries=0) - -        self.thread.m.reconnecting.wait(3) # Ease issue with later downloads appearing to be in parallel -     -    def doTimmer(self): -        response = self.load(self.url, -                         post={"downloadLink": "wait"}, -                         decode=True) -        self.logDebug("wait response : %s" % response[:80]) - -        if "fail" in response: -            self.fail("Failed getting wait time") - -        if self.__name__ == "FilejungleCom":    -            found = re.search(r'"waitTime":(\d+)', response) -            if not found: self.fail("Cannot get wait time") -            wait_time = int(found.group(1)) -        else: -            wait_time = int(response) + 3 -             -        self.setWait(wait_time) -        self.wait() - -    def doCaptcha(self): -        captcha_key = re.search(self.CAPTCHA_KEY_PATTERN, self.html).group("key") -        recaptcha = ReCaptcha(self) - -        for i in range(5): -            challenge, code = recaptcha.challenge(captcha_key) - -            response = json_loads(self.load(self.URLS[2], -                            post={'recaptcha_challenge_field': challenge, -                                  'recaptcha_response_field': code, -                                  'recaptcha_shortencode_field': self.file_id})) -            self.logDebug("reCaptcha response : %s" % response) -            if not response["success"]: -                self.invalidCaptcha() -            else: -                self.correctCaptcha() -                break -        else: self.fail("Invalid captcha") -         -    def doLongWait(self, m): -        wait_time = (int(m.group(1)) * {'seconds':1, 'minutes':60, 'hours':3600}[m.group(2)]) if m else 720  -        self.setWait(wait_time, True) -        self.wait() -        self.retry() -     -    def handlePremium(self): -        premium_url = None  -        if self.__name__ == "FileserveCom": -            #try api download -            response = self.load("http://app.fileserve.com/api/download/premium/", -                            post = {"username": self.user, -                                    "password": self.account.getAccountData(self.user)["password"], -                                    "shorten": self.file_id},  -                            decode = True) -            if response: -                response = json_loads(response) -                if response['error_code'] == "302": premium_url = response['next'] -                elif response['error_code'] in ["305", "500"]: self.tempOffline() -                elif response['error_code'] in ["403", "605"]: self.resetAccount() -                elif response['error_code'] in ["606", "607", "608"]: self.offline() -                else: self.logError(response['error_code'], response['error_message']) -                -        self.download(premium_url or self.pyfile.url) -         -        if not premium_url:    -            check = self.checkDownload({"login": re.compile(self.NOT_LOGGED_IN_PATTERN)}) -     -            if check == "login": -                self.account.relogin(self.user) -                self.retry(reason=_("Not logged in.")) - -def getInfo(urls): -    for chunk in chunks(urls, 100): yield checkFile(FileserveCom, chunk)
\ No newline at end of file | 
