From 5ffe48238d4e6664e1a1c87505e5d3387ad4215a Mon Sep 17 00:00:00 2001 From: EstrellaXD Date: Tue, 19 Jul 2022 22:17:33 +0800 Subject: [PATCH] =?UTF-8?q?2.5.15=20-=20=E4=BF=AE=E5=A4=8D=E5=B0=91?= =?UTF-8?q?=E9=87=8F=E5=8F=98=E9=87=8F=20-=20=E5=BC=80=E5=90=AF=20?= =?UTF-8?q?=E5=88=B7=E6=96=B0=20RSS=20=E9=80=89=E9=A1=B9?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- src/conf/const.py | 5 +++-- src/network/__init__.py | 18 +++++++++--------- src/network/request.py | 1 + 3 files changed, 13 insertions(+), 11 deletions(-) diff --git a/src/conf/const.py b/src/conf/const.py index 18957870..0da258e2 100644 --- a/src/conf/const.py +++ b/src/conf/const.py @@ -28,7 +28,7 @@ DEFAULT_SETTINGS = { "enable_rename": True, "reset_folder": False, "log_path": "/config/log.txt", - "refresh_rss": True, + "refresh_rss": False, } ENV_TO_ATTR = { @@ -55,7 +55,8 @@ ENV_TO_ATTR = { "AB_SOCKS": "socks", "AB_RENAME": ("enable_rename", lambda e: e.lower() in ("true", "1", "t")), "AB_RSS_COLLECTOR": ("enable_rss_collector", lambda e: e.lower() in ("true", "1", "t")), - "AB_RESET_FOLDER": ("reset_folder", lambda e: e.lower() in ("true", "1", "t")) + "AB_RESET_FOLDER": ("reset_folder", lambda e: e.lower() in ("true", "1", "t")), + "AB_REFRESH_RSS": ("refresh_rss", lambda e: e.lower() in ("true", "1", "t")), } diff --git a/src/network/__init__.py b/src/network/__init__.py index 95d7d06c..be49175c 100644 --- a/src/network/__init__.py +++ b/src/network/__init__.py @@ -17,25 +17,25 @@ class RequestContent: self._req = RequestURL() # Mikanani RSS - def get_torrents(self, url: str) -> [TorrentInfo]: - soup = self._req.get_content(url) + def get_torrents(self, _url: str) -> [TorrentInfo]: + soup = self._req.get_content(_url) torrent_titles = [item.title.string for item in soup.find_all("item")] torrent_urls = [item.get("url") for item in soup.find_all("enclosure")] torrents = [] - for title, torrent_url in zip(torrent_titles, torrent_urls): - if re.search(settings.not_contain, title) is None: - torrents.append(TorrentInfo(title, torrent_url)) + for _title, torrent_url in zip(torrent_titles, torrent_urls): + if re.search(settings.not_contain, _title) is None: + torrents.append(TorrentInfo(_title, torrent_url)) return torrents - def get_torrent(self, url) -> TorrentInfo: - soup = self._req.get_content(url) + def get_torrent(self, _url) -> TorrentInfo: + soup = self._req.get_content(_url) item = soup.find("item") enclosure = item.find("enclosure") return TorrentInfo(item.title.string, enclosure["url"]) # API JSON - def get_json(self, url) -> dict: - return self._req.get_content(url, content="json") + def get_json(self, _url) -> dict: + return self._req.get_content(_url, content="json") def close_session(self): self._req.close() diff --git a/src/network/request.py b/src/network/request.py index 313ec4d4..f15b82d9 100644 --- a/src/network/request.py +++ b/src/network/request.py @@ -38,6 +38,7 @@ class RequestURL: return req except Exception as e: logger.debug(f"URL: {url}") + logger.debug(e) logger.warning("ERROR with Connection.Please check DNS/Connection settings") time.sleep(settings.connect_retry_interval) times += 1