import datetime import re import traceback from typing import Any, Optional from typing import List from urllib.parse import quote, urlencode, urlparse, parse_qs from fastapi.concurrency import run_in_threadpool from jinja2 import Template from pyquery import PyQuery from app.core.config import settings from app.log import logger from app.schemas.types import MediaType from app.utils.http import RequestUtils, AsyncRequestUtils from app.utils.string import StringUtils class SiteSpider: """ 站点爬虫 """ @property def __class__(self): return object @property def __dict__(self): return {} @property def __dir__(self): raise AttributeError(f"Cannot read protected attribute!") def __init__(self, indexer: dict, keyword: Optional[str] = None, mtype: MediaType = None, cat: Optional[str] = None, page: Optional[int] = 0, referer: Optional[str] = None): """ 设置查询参数 :param indexer: 索引器 :param keyword: 搜索关键字,如果数组则为批量搜索 :param mtype: 媒体类型 :param cat: 搜索分类 :param page: 页码 :param referer: Referer """ if not indexer: return self.keyword = keyword self.cat = cat self.mtype = mtype self.indexerid = indexer.get('id') self.indexername = indexer.get('name') self.search = indexer.get('search') self.batch = indexer.get('batch') self.browse = indexer.get('browse') self.category = indexer.get('category') self.list = indexer.get('torrents').get('list', {}) self.fields = indexer.get('torrents').get('fields') if not keyword and self.browse: self.list = self.browse.get('list') or self.list self.fields = self.browse.get('fields') or self.fields self.domain = indexer.get('domain') self.result_num = int(indexer.get('result_num') or 100) self._timeout = int(indexer.get('timeout') or 15) self.page = page if self.domain and not str(self.domain).endswith("/"): self.domain = self.domain + "/" self.ua = indexer.get('ua') or settings.USER_AGENT self.proxies = settings.PROXY if indexer.get('proxy') else None self.proxy_server = settings.PROXY_SERVER if indexer.get('proxy') else None self.cookie = indexer.get('cookie') self.referer = referer # 初始化属性 self.is_error = False self.torrents_info = {} self.torrents_info_array = [] def __get_search_url(self): """ 获取搜索URL """ # 种子搜索相对路径 paths = self.search.get('paths', []) torrentspath = "" if len(paths) == 1: torrentspath = paths[0].get('path', '') else: for path in paths: if path.get("type") == "all" and not self.mtype: torrentspath = path.get('path') break elif path.get("type") == "movie" and self.mtype == MediaType.MOVIE: torrentspath = path.get('path') break elif path.get("type") == "tv" and self.mtype == MediaType.TV: torrentspath = path.get('path') break # 精确搜索 if self.keyword: if isinstance(self.keyword, list): # 批量查询 if self.batch: delimiter = self.batch.get('delimiter') or ' ' space_replace = self.batch.get('space_replace') or ' ' search_word = delimiter.join([str(k).replace(' ', space_replace) for k in self.keyword]) else: search_word = " ".join(self.keyword) # 查询模式:或 search_mode = "1" else: # 单个查询 search_word = self.keyword # 查询模式与 search_mode = "0" # 搜索URL indexer_params = self.search.get("params", {}).copy() if indexer_params: search_area = indexer_params.get('search_area') # search_area非0表示支持imdbid搜索 if (search_area and (not self.keyword or not self.keyword.startswith('tt'))): # 支持imdbid搜索,但关键字不是imdbid时,不启用imdbid搜索 indexer_params.pop('search_area') # 变量字典 inputs_dict = { "keyword": search_word } # 查询参数,默认查询标题 params = { "search_mode": search_mode, "search_area": 0, "page": self.page or 0, "notnewword": 1 } # 额外参数 for key, value in indexer_params.items(): params.update({ "%s" % key: str(value).format(**inputs_dict) }) # 分类条件 if self.category: if self.mtype == MediaType.TV: cats = self.category.get("tv") or [] elif self.mtype == MediaType.MOVIE: cats = self.category.get("movie") or [] else: cats = (self.category.get("movie") or []) + (self.category.get("tv") or []) allowed_cats = set(self.cat.split(',')) if self.cat else None for cat in cats: if allowed_cats and str(cat.get('id')) not in allowed_cats: continue if self.category.get("field"): value = params.get(self.category.get("field"), "") params.update({ "%s" % self.category.get("field"): value + self.category.get("delimiter", ' ') + cat.get("id") }) else: params.update({ "cat%s" % cat.get("id"): 1 }) searchurl = self.domain + torrentspath + "?" + urlencode(params) else: # 变量字典 inputs_dict = { "keyword": quote(search_word), "page": self.page or 0 } # 无额外参数 searchurl = self.domain + str(torrentspath).format(**inputs_dict) # 列表浏览 else: # 变量字典 inputs_dict = { "page": self.page or 0, "keyword": "" } # 有单独浏览路径 if self.browse: torrentspath = self.browse.get("path") if self.browse.get("start"): start_page = int(self.browse.get("start")) + int(self.page or 0) inputs_dict.update({ "page": start_page }) elif self.page: torrentspath = torrentspath + f"?page={self.page}" # 搜索Url searchurl = self.domain + str(torrentspath).format(**inputs_dict) return searchurl def get_torrents(self) -> List[dict]: """ 开始请求 """ if not self.search or not self.domain: return [] # 获取搜索URL searchurl = self.__get_search_url() logger.info(f"开始请求:{searchurl}") # requests请求 ret = RequestUtils( ua=self.ua, cookies=self.cookie, timeout=self._timeout, referer=self.referer, proxies=self.proxies ).get_res(searchurl, allow_redirects=True) # 解析返回 return self.parse( RequestUtils.get_decoded_html_content( ret, performance_mode=settings.ENCODING_DETECTION_PERFORMANCE_MODE, confidence_threshold=settings.ENCODING_DETECTION_MIN_CONFIDENCE ) ) async def async_get_torrents(self) -> List[dict]: """ 异步请求 """ if not self.search or not self.domain: return [] # 获取搜索URL searchurl = self.__get_search_url() logger.info(f"开始异步请求:{searchurl}") # httpx请求 ret = await AsyncRequestUtils( ua=self.ua, cookies=self.cookie, timeout=self._timeout, referer=self.referer, proxies=self.proxies ).get_res(searchurl, allow_redirects=True) # 解析返回 return await run_in_threadpool( self.parse, RequestUtils.get_decoded_html_content( ret, performance_mode=settings.ENCODING_DETECTION_PERFORMANCE_MODE, confidence_threshold=settings.ENCODING_DETECTION_MIN_CONFIDENCE ) ) def __get_title(self, torrent: Any): # title default text if 'title' not in self.fields: return selector = self.fields.get('title', {}) if 'selector' in selector: self.torrents_info['title'] = self._safe_query(torrent, selector) elif 'text' in selector: render_dict = {} if "title_default" in self.fields: title_default_selector = self.fields.get('title_default', {}) title_default = self._safe_query(torrent, title_default_selector) render_dict.update({'title_default': title_default}) if "title_optional" in self.fields: title_optional_selector = self.fields.get('title_optional', {}) title_optional = self._safe_query(torrent, title_optional_selector) render_dict.update({'title_optional': title_optional}) self.torrents_info['title'] = Template(selector.get('text')).render(fields=render_dict) self.torrents_info['title'] = self.__filter_text(self.torrents_info.get('title'), selector.get('filters')) def __get_description(self, torrent: Any): # description text if 'description' not in self.fields: return selector = self.fields.get('description', {}) if "selector" in selector or "selectors" in selector: # 对于selectors情况,需要特殊处理selector_config desc_selector = selector.copy() if "selectors" in selector and "selector" not in selector: desc_selector["selector"] = selector.get("selectors", "") self.torrents_info['description'] = self._safe_query(torrent, desc_selector) elif "text" in selector: render_dict = {} if "tags" in self.fields: tags_selector = self.fields.get('tags', {}) tag = self._safe_query(torrent, tags_selector) render_dict.update({'tags': tag}) if "subject" in self.fields: subject_selector = self.fields.get('subject', {}) subject = self._safe_query(torrent, subject_selector) render_dict.update({'subject': subject}) if "description_free_forever" in self.fields: description_free_forever_selector = self.fields.get("description_free_forever", {}) description_free_forever = self._safe_query(torrent, description_free_forever_selector) render_dict.update({"description_free_forever": description_free_forever}) if "description_normal" in self.fields: description_normal_selector = self.fields.get("description_normal", {}) description_normal = self._safe_query(torrent, description_normal_selector) render_dict.update({"description_normal": description_normal}) self.torrents_info['description'] = Template(selector.get('text')).render(fields=render_dict) self.torrents_info['description'] = self.__filter_text(self.torrents_info.get('description'), selector.get('filters')) def __get_detail(self, torrent: Any): # details page text if 'details' not in self.fields: return selector = self.fields.get('details', {}) item = self._safe_query(torrent, selector) detail_link = self.__filter_text(item, selector.get('filters')) if detail_link: if not detail_link.startswith("http"): if detail_link.startswith("//"): self.torrents_info['page_url'] = self.domain.split(":")[0] + ":" + detail_link elif detail_link.startswith("/"): self.torrents_info['page_url'] = self.domain + detail_link[1:] else: self.torrents_info['page_url'] = self.domain + detail_link else: self.torrents_info['page_url'] = detail_link def __get_download(self, torrent: Any): # download link text if 'download' not in self.fields: return selector = self.fields.get('download', {}) item = self._safe_query(torrent, selector) download_link = self.__filter_text(item, selector.get('filters')) if download_link: if not download_link.startswith("http") \ and not download_link.startswith("magnet"): _scheme, _domain = StringUtils.get_url_netloc(self.domain) if _domain in download_link: if download_link.startswith("/"): self.torrents_info['enclosure'] = f"{_scheme}:{download_link}" else: self.torrents_info['enclosure'] = f"{_scheme}://{download_link}" else: if download_link.startswith("/"): self.torrents_info['enclosure'] = f"{self.domain}{download_link[1:]}" else: self.torrents_info['enclosure'] = f"{self.domain}{download_link}" else: self.torrents_info['enclosure'] = download_link def __get_imdbid(self, torrent: Any): # imdbid if "imdbid" not in self.fields: return selector = self.fields.get('imdbid', {}) item = self._safe_query(torrent, selector) self.torrents_info['imdbid'] = self.__filter_text(item, selector.get('filters')) def __get_size(self, torrent: Any): # torrent size int if 'size' not in self.fields: return selector = self.fields.get('size', {}) item = self._safe_query(torrent, selector) if item: size_val = item.replace("\n", "").strip() size_val = self.__filter_text(size_val, selector.get('filters')) self.torrents_info['size'] = StringUtils.num_filesize(size_val) else: self.torrents_info['size'] = 0 def __get_leechers(self, torrent: Any): # torrent leechers int if 'leechers' not in self.fields: return selector = self.fields.get('leechers', {}) item = self._safe_query(torrent, selector) if item: peers_val = item.split("/")[0] peers_val = peers_val.replace(",", "") peers_val = self.__filter_text(peers_val, selector.get('filters')) self.torrents_info['peers'] = int(peers_val) if peers_val and peers_val.isdigit() else 0 else: self.torrents_info['peers'] = 0 def __get_seeders(self, torrent: Any): # torrent seeders int if 'seeders' not in self.fields: return selector = self.fields.get('seeders', {}) item = self._safe_query(torrent, selector) if item: seeders_val = item.split("/")[0] seeders_val = seeders_val.replace(",", "") seeders_val = self.__filter_text(seeders_val, selector.get('filters')) self.torrents_info['seeders'] = int(seeders_val) if seeders_val and seeders_val.isdigit() else 0 else: self.torrents_info['seeders'] = 0 def __get_grabs(self, torrent: Any): # torrent grabs int if 'grabs' not in self.fields: return selector = self.fields.get('grabs', {}) item = self._safe_query(torrent, selector) if item: grabs_val = item.split("/")[0] grabs_val = grabs_val.replace(",", "") grabs_val = self.__filter_text(grabs_val, selector.get('filters')) self.torrents_info['grabs'] = int(grabs_val) if grabs_val and grabs_val.isdigit() else 0 else: self.torrents_info['grabs'] = 0 def __get_pubdate(self, torrent: Any): # torrent pubdate yyyy-mm-dd hh:mm:ss if 'date_added' not in self.fields: return selector = self.fields.get('date_added', {}) pubdate_str = self._safe_query(torrent, selector) if pubdate_str: pubdate_str = pubdate_str.replace('\n', ' ').strip() self.torrents_info['pubdate'] = self.__filter_text(pubdate_str, selector.get('filters')) def __get_date_elapsed(self, torrent: Any): # torrent date elapsed text if 'date_elapsed' not in self.fields: return selector = self.fields.get('date_elapsed', {}) date_elapsed = self._safe_query(torrent, selector) self.torrents_info['date_elapsed'] = self.__filter_text(date_elapsed, selector.get('filters')) def __get_downloadvolumefactor(self, torrent: Any): # downloadvolumefactor int selector = self.fields.get('downloadvolumefactor', {}) if not selector: return self.torrents_info['downloadvolumefactor'] = 1 if 'case' in selector: for downloadvolumefactorselector in list(selector.get('case', {}).keys()): downloadvolumefactor = torrent(downloadvolumefactorselector) try: if len(downloadvolumefactor) > 0: self.torrents_info['downloadvolumefactor'] = selector.get('case', {}).get( downloadvolumefactorselector) break finally: downloadvolumefactor.clear() del downloadvolumefactor elif "selector" in selector: item = self._safe_query(torrent, selector) if item: downloadvolumefactor = re.search(r'(\d+\.?\d*)', item) if downloadvolumefactor: self.torrents_info['downloadvolumefactor'] = int(downloadvolumefactor.group(1)) def __get_uploadvolumefactor(self, torrent: Any): # uploadvolumefactor int selector = self.fields.get('uploadvolumefactor', {}) if not selector: return self.torrents_info['uploadvolumefactor'] = 1 if 'case' in selector: for uploadvolumefactorselector in list(selector.get('case', {}).keys()): uploadvolumefactor = torrent(uploadvolumefactorselector) try: if len(uploadvolumefactor) > 0: self.torrents_info['uploadvolumefactor'] = selector.get('case', {}).get( uploadvolumefactorselector) break finally: uploadvolumefactor.clear() del uploadvolumefactor elif "selector" in selector: item = self._safe_query(torrent, selector) if item: uploadvolumefactor = re.search(r'(\d+\.?\d*)', item) if uploadvolumefactor: self.torrents_info['uploadvolumefactor'] = int(uploadvolumefactor.group(1)) def __get_labels(self, torrent: Any): # labels ['label1', 'label2'] if 'labels' not in self.fields: return selector = self.fields.get('labels', {}) if not selector.get('selector'): self.torrents_info['labels'] = [] return # labels需要特殊处理,因为它返回的是列表 labels = torrent(selector.get("selector", "")).clone() try: self.__remove(labels, selector) items = self.__attribute_or_text(labels, selector) if items: self.torrents_info['labels'] = [item for item in items if item] else: self.torrents_info['labels'] = [] finally: labels.clear() del labels def __get_free_date(self, torrent: Any): # free date yyyy-mm-dd hh:mm:ss if 'freedate' not in self.fields: return selector = self.fields.get('freedate', {}) freedate = self._safe_query(torrent, selector) self.torrents_info['freedate'] = self.__filter_text(freedate, selector.get('filters')) def __get_hit_and_run(self, torrent: Any): # hitandrun True/False if 'hr' not in self.fields: return selector = self.fields.get('hr', {}) hit_and_run = torrent(selector.get('selector', '')) try: if hit_and_run: self.torrents_info['hit_and_run'] = True else: self.torrents_info['hit_and_run'] = False finally: hit_and_run.clear() del hit_and_run def __get_category(self, torrent: Any): # category 电影/电视剧 if 'category' not in self.fields: return selector = self.fields.get('category', {}) category_value = self._safe_query(torrent, selector) category_value = self.__filter_text(category_value, selector.get('filters')) if category_value and self.category: tv_cats = [str(cat.get("id")) for cat in self.category.get("tv") or []] movie_cats = [str(cat.get("id")) for cat in self.category.get("movie") or []] if category_value in tv_cats \ and category_value not in movie_cats: self.torrents_info['category'] = MediaType.TV.value elif category_value in movie_cats: self.torrents_info['category'] = MediaType.MOVIE.value else: self.torrents_info['category'] = MediaType.UNKNOWN.value else: self.torrents_info['category'] = MediaType.UNKNOWN.value def _safe_query(self, torrent: Any, selector_config: Optional[dict]) -> Optional[str]: """ 安全地执行PyQuery查询并自动清理资源 :param torrent: PyQuery对象 :param selector_config: 选择器配置 :return: 处理后的结果 """ if not selector_config or not selector_config.get('selector'): return None query_obj = torrent(selector_config.get('selector', '')).clone() try: self.__remove(query_obj, selector_config) items = self.__attribute_or_text(query_obj, selector_config) return self.__index(items, selector_config) finally: query_obj.clear() del query_obj def get_info(self, torrent: Any) -> dict: """ 解析单条种子数据 """ # 每次调用时重新初始化,避免数据累积 self.torrents_info = {} try: # 标题 self.__get_title(torrent) # 描述 self.__get_description(torrent) # 详情页面 self.__get_detail(torrent) # 下载链接 self.__get_download(torrent) # 完成数 self.__get_grabs(torrent) # 下载数 self.__get_leechers(torrent) # 做种数 self.__get_seeders(torrent) # 大小 self.__get_size(torrent) # IMDBID self.__get_imdbid(torrent) # 下载系数 self.__get_downloadvolumefactor(torrent) # 上传系数 self.__get_uploadvolumefactor(torrent) # 发布时间 self.__get_pubdate(torrent) # 已发布时间 self.__get_date_elapsed(torrent) # 免费载止时间 self.__get_free_date(torrent) # 标签 self.__get_labels(torrent) # HR self.__get_hit_and_run(torrent) # 分类 self.__get_category(torrent) # 返回当前种子信息的副本,而不是引用 return self.torrents_info.copy() if self.torrents_info else {} except Exception as err: logger.error("%s 搜索出现错误:%s" % (self.indexername, str(err))) return {} finally: self.torrents_info.clear() @staticmethod def __filter_text(text: Optional[str], filters: Optional[List[dict]]) -> str: """ 对文件进行处理 """ if not text or not filters or not isinstance(filters, list): return text if not isinstance(text, str): text = str(text) for filter_item in filters: if not text: break method_name = filter_item.get("name") try: args = filter_item.get("args") if method_name == "re_search" and isinstance(args, list): rematch = re.search(r"%s" % args[0], text) if rematch: text = rematch.group(args[-1]) elif method_name == "split" and isinstance(args, list): text = text.split(r"%s" % args[0])[args[-1]] elif method_name == "replace" and isinstance(args, list): text = text.replace(r"%s" % args[0], r"%s" % args[-1]) elif method_name == "dateparse" and isinstance(args, str): text = text.replace("\n", " ").strip() text = datetime.datetime.strptime(text, r"%s" % args) elif method_name == "strip": text = text.strip() elif method_name == "appendleft": text = f"{args}{text}" elif method_name == "querystring": parsed_url = urlparse(str(text)) query_params = parse_qs(parsed_url.query) param_value = query_params.get(args) text = param_value[0] if param_value else '' except Exception as err: logger.debug(f'过滤器 {method_name} 处理失败:{str(err)} - {traceback.format_exc()}') return text.strip() @staticmethod def __remove(item: Any, selector: Optional[dict]): """ 移除元素 """ if selector and "remove" in selector: removelist = selector.get('remove', '').split(', ') for v in removelist: item.remove(v) @staticmethod def __attribute_or_text(item: Any, selector: Optional[dict]) -> list: if not selector: return item if not item: return [] if 'attribute' in selector: items = [i.attr(selector.get('attribute')) for i in item.items() if i] else: items = [i.text() for i in item.items() if i] return items @staticmethod def __index(items: Optional[list], selector: Optional[dict]) -> Optional[str]: if not items: return None if selector: if "contents" in selector \ and len(items) > int(selector.get("contents")): item = items[0].split("\n")[selector.get("contents")] elif "index" in selector \ and len(items) > int(selector.get("index")): item = items[int(selector.get("index"))] else: item = items[0] else: item = items[0] return item def parse(self, html_text: str) -> List[dict]: """ 解析整个页面 """ if not html_text: self.is_error = True return [] # 清空旧结果 self.torrents_info_array = [] html_doc = None try: # 解析站点文本对象 html_doc = PyQuery(html_text) # 种子筛选器 torrents_selector = self.list.get('selector', '') # 遍历种子html列表 for i, torn in enumerate(html_doc(torrents_selector)): if i >= int(self.result_num): break # 创建临时PyQuery对象进行解析 torrent_query = PyQuery(torn) try: # 直接获取种子信息,避免深拷贝 torrent_info = self.get_info(torrent_query) if torrent_info: # 浅拷贝即可,减少内存使用 self.torrents_info_array.append(torrent_info) finally: # 显式删除临时PyQuery对象 torrent_query.clear() del torrent_query # 返回数组的副本,防止被后续清理操作影响 return self.torrents_info_array.copy() except Exception as err: self.is_error = True logger.warn(f"错误:{self.indexername} {str(err)}") return [] finally: # 清理种子缓存 self.torrents_info_array.clear() # 清理HTML文档对象 if html_doc is not None: html_doc.clear() del html_doc # 清理html_text引用 del html_text