searxng/searx/engines/kickass.py

64 lines
2.0 KiB
Python

# SPDX-License-Identifier: AGPL-3.0-or-later
# lint: pylint
"""Kickass Torrent (Videos, Music, Files)"""
import random
from operator import itemgetter
from urllib.parse import quote
from lxml import html
from searx.utils import (
eval_xpath,
eval_xpath_getindex,
eval_xpath_list,
extract_text,
get_torrent_size,
int_or_zero,
)
about = {
"website": 'https://kickasstorrents.to',
"wikidata_id": 'Q17062285',
"official_api_documentation": None,
"use_official_api": False,
"require_api_key": False,
"results": 'HTML',
}
categories = ['files']
paging = True
# base_url can be overwritten by a list of URLs in the settings.yml
base_url = 'https://kickasstorrents.to'
def request(query, params):
params['base_url'] = random.choice(base_url) if isinstance(base_url, list) else base_url
params['url'] = params['base_url'] + f'/usearch/{quote(query)}/{params["pageno"]}/'
return params
def response(resp):
results = []
dom = html.fromstring(resp.text)
search_res = eval_xpath_list(dom, '//table[contains(@class, "data")]//tr', None)
if search_res is None:
return []
for tag in search_res[1:]:
result = {'template': 'torrent.html'}
url = eval_xpath_getindex(tag, './/a[contains(@class, "cellMainLink")]/@href', 0, None)
result['url'] = resp.search_params['base_url'] + url
result['title'] = extract_text(eval_xpath(tag, './/a[contains(@class, "cellMainLink")]'))
result['content'] = extract_text(eval_xpath(tag, './/span[@class="font11px lightgrey block"]'))
result['seed'] = int_or_zero(extract_text(eval_xpath(tag, './/td[contains(@class, "green")]')))
result['leech'] = int_or_zero(extract_text(eval_xpath(tag, './/td[contains(@class, "red")]')))
result['filesize'] = get_torrent_size(*extract_text(eval_xpath(tag, './/td[contains(@class, "nobr")]')).split())
results.append(result)
# results sorted by seeder count
return sorted(results, key=itemgetter('seed'), reverse=True)