diff options
| author | 2014-09-26 15:39:33 +0200 | |
|---|---|---|
| committer | 2014-09-26 15:39:33 +0200 | |
| commit | 76c67b22c0b2dbd275fae77e8a3121acb2d4301b (patch) | |
| tree | 1b5db66ef14830ceaadeddd7ac0c84904f7647c9 /module/plugins/internal | |
| parent | Remove ReCaptcha (diff) | |
| download | pyload-76c67b22c0b2dbd275fae77e8a3121acb2d4301b.tar.xz | |
Fix cookie deletion + Account support for SimpleCrypter
Fix: https://github.com/pyload/pyload/issues/749
Diffstat (limited to 'module/plugins/internal')
| -rw-r--r-- | module/plugins/internal/SimpleCrypter.py | 17 | ||||
| -rw-r--r-- | module/plugins/internal/SimpleHoster.py | 23 | 
2 files changed, 31 insertions, 9 deletions
| diff --git a/module/plugins/internal/SimpleCrypter.py b/module/plugins/internal/SimpleCrypter.py index 3cd52d9f1..bacfce1dd 100644 --- a/module/plugins/internal/SimpleCrypter.py +++ b/module/plugins/internal/SimpleCrypter.py @@ -10,7 +10,7 @@ from module.utils import html_unescape  class SimpleCrypter(Crypter):      __name__ = "SimpleCrypter"      __type__ = "crypter" -    __version__ = "0.10" +    __version__ = "0.11"      __pattern__ = None @@ -48,9 +48,14 @@ class SimpleCrypter(Crypter):            return the html of the page number page_n      """ +      URL_REPLACEMENTS = [] -    SH_COOKIES = True  # or False or list of tuples [(domain, name, value)] +    SH_BROKEN_ENCODING = False  #: Set to True or encoding name if encoding in http header is not correct +    SH_COOKIES = True  #: or False or list of tuples [(domain, name, value)] + +    LOGIN_ACCOUNT = False +    LOGIN_PREMIUM = False      def setup(self): @@ -59,9 +64,15 @@ class SimpleCrypter(Crypter):      def decrypt(self, pyfile): +        if self.LOGIN_ACCOUNT and not self.account: +            self.fail('Required account not found!') + +        if self.LOGIN_PREMIUM and not self.premium: +            self.fail('Required premium account not found!') +          pyfile.url = replace_patterns(pyfile.url, self.URL_REPLACEMENTS) -        self.html = self.load(pyfile.url, decode=True) +        self.html = self.load(pyfile.url, decode=not self.SH_BROKEN_ENCODING, cookies=self.SH_COOKIES)          self.checkOnline() diff --git a/module/plugins/internal/SimpleHoster.py b/module/plugins/internal/SimpleHoster.py index ec9cf1b70..b2db56f3f 100644 --- a/module/plugins/internal/SimpleHoster.py +++ b/module/plugins/internal/SimpleHoster.py @@ -187,24 +187,26 @@ class SimpleHoster(Hoster):      FILE_SIZE_REPLACEMENTS = []      FILE_URL_REPLACEMENTS = [] -    SH_BROKEN_ENCODING = False  # Set to True or encoding name if encoding in http header is not correct -    SH_COOKIES = True  # or False or list of tuples [(domain, name, value)] -    SH_CHECK_TRAFFIC = False  # True = force check traffic left for a premium account +    SH_BROKEN_ENCODING = False  #: Set to True or encoding name if encoding in http header is not correct +    SH_COOKIES = True  #: or False or list of tuples [(domain, name, value)] +    SH_CHECK_TRAFFIC = False  #: True = force check traffic left for a premium account      def init(self):          self.file_info = {} +      def setup(self):          self.resumeDownload = self.multiDL = self.premium          if isinstance(self.SH_COOKIES, list):              set_cookies(self.req.cj, self.SH_COOKIES) +      def process(self, pyfile):          pyfile.url = replace_patterns(pyfile.url, self.FILE_URL_REPLACEMENTS)          self.req.setOption("timeout", 120)          # Due to a 0.4.9 core bug self.load would keep previous cookies even if overridden by cookies parameter. -        # Workaround using getURL. Can be reverted in 0.5 as the cookies bug has been fixed. +        # Workaround using getURL. Can be reverted in 0.4.10 as the cookies bug has been fixed.          self.html = getURL(pyfile.url, decode=not self.SH_BROKEN_ENCODING, cookies=self.SH_COOKIES)          premium_only = hasattr(self, 'PREMIUM_ONLY_PATTERN') and re.search(self.PREMIUM_ONLY_PATTERN, self.html)          if not premium_only:  # Usually premium only pages doesn't show the file information @@ -215,16 +217,18 @@ class SimpleHoster(Hoster):          elif premium_only:              self.fail("This link require a premium account")          else: -            # This line is required due to the getURL workaround. Can be removed in 0.5 +            # This line is required due to the getURL workaround. Can be removed in 0.4.10              self.html = self.load(pyfile.url, decode=not self.SH_BROKEN_ENCODING, cookies=self.SH_COOKIES)              self.handleFree() +      def load(self, url, get={}, post={}, ref=True, cookies=True, just_header=False, decode=False):          if type(url) == unicode:              url = url.encode('utf8')          return Hoster.load(self, url=url, get=get, post=post, ref=ref, cookies=cookies,                             just_header=just_header, decode=decode) +      def getFileInfo(self):          self.logDebug("URL: %s" % self.pyfile.url) @@ -251,15 +255,19 @@ class SimpleHoster(Hoster):          self.logDebug("FILE NAME: %s FILE SIZE: %s" % (self.pyfile.name, self.pyfile.size))          return self.file_info +      def handleFree(self):          self.fail("Free download not implemented") +      def handlePremium(self):          self.fail("Premium download not implemented") +      def parseError(self, msg):          raise PluginParseError(msg) +      def longWait(self, wait_time=None, max_tries=3):          if wait_time and isinstance(wait_time, (int, long, float)):              time_str = "%dh %dm" % divmod(wait_time / 60, 60) @@ -274,9 +282,11 @@ class SimpleHoster(Hoster):          self.wait()          self.retry(max_tries=max_tries, reason="Download limit reached") +      def parseHtmlForm(self, attr_str='', input_names=None):          return parseHtmlForm(attr_str, self.html, input_names) +      def checkTrafficLeft(self):          traffic = self.account.getAccountInfo(self.user, True)['trafficleft']          if traffic == -1: @@ -285,7 +295,8 @@ class SimpleHoster(Hoster):          self.logInfo("Filesize: %i KiB, Traffic left for user %s: %i KiB" % (size, self.user, traffic))          return size <= traffic -    # TODO: Remove in 0.5 + +    #@TODO: Remove in 0.4.10      def wait(self, seconds=False, reconnect=False):          if seconds:              self.setWait(seconds, reconnect) | 
