forked from Archives/searxng
fd65c12921
languages.py can change, so users may query on a language that is not on the list anymore, even if it is still recognized by a few engines. also made no and nb the same because they seem to return the same, though most engines will only support one or the other.
129 lines
4.1 KiB
Python
129 lines
4.1 KiB
Python
"""
|
|
Qwant (Web, Images, News, Social)
|
|
|
|
@website https://qwant.com/
|
|
@provide-api not officially (https://api.qwant.com/api/search/)
|
|
|
|
@using-api yes
|
|
@results JSON
|
|
@stable yes
|
|
@parse url, title, content
|
|
"""
|
|
|
|
from datetime import datetime
|
|
from json import loads
|
|
from urllib import urlencode
|
|
|
|
from searx.utils import html_to_text
|
|
|
|
# engine dependent config
|
|
categories = None
|
|
paging = True
|
|
language_support = True
|
|
supported_languages_url = 'https://qwant.com/region'
|
|
|
|
category_to_keyword = {'general': 'web',
|
|
'images': 'images',
|
|
'news': 'news',
|
|
'social media': 'social'}
|
|
|
|
# search-url
|
|
url = 'https://api.qwant.com/api/search/{keyword}?count=10&offset={offset}&f=&{query}'
|
|
|
|
|
|
# do search-request
|
|
def request(query, params):
|
|
offset = (params['pageno'] - 1) * 10
|
|
|
|
if categories[0] and categories[0] in category_to_keyword:
|
|
|
|
params['url'] = url.format(keyword=category_to_keyword[categories[0]],
|
|
query=urlencode({'q': query}),
|
|
offset=offset)
|
|
else:
|
|
params['url'] = url.format(keyword='web',
|
|
query=urlencode({'q': query}),
|
|
offset=offset)
|
|
|
|
# add language tag if specified
|
|
if params['language'] != 'all':
|
|
if params['language'] == 'no' or params['language'].startswith('no-'):
|
|
params['language'] = params['language'].replace('no', 'nb', 1)
|
|
if params['language'].find('-') < 0:
|
|
# tries to get a country code from language
|
|
for lang in supported_languages:
|
|
lc = lang.split('-')
|
|
if params['language'] == lc[0]:
|
|
params['language'] = lang
|
|
break
|
|
params['url'] += '&locale=' + params['language'].replace('-', '_').lower()
|
|
|
|
return params
|
|
|
|
|
|
# get response from search-request
|
|
def response(resp):
|
|
results = []
|
|
|
|
search_results = loads(resp.text)
|
|
|
|
# return empty array if there are no results
|
|
if 'data' not in search_results:
|
|
return []
|
|
|
|
data = search_results.get('data', {})
|
|
|
|
res = data.get('result', {})
|
|
|
|
# parse results
|
|
for result in res.get('items', {}):
|
|
|
|
title = html_to_text(result['title'])
|
|
res_url = result['url']
|
|
content = html_to_text(result['desc'])
|
|
|
|
if category_to_keyword.get(categories[0], '') == 'web':
|
|
results.append({'title': title,
|
|
'content': content,
|
|
'url': res_url})
|
|
|
|
elif category_to_keyword.get(categories[0], '') == 'images':
|
|
thumbnail_src = result['thumbnail']
|
|
img_src = result['media']
|
|
results.append({'template': 'images.html',
|
|
'url': res_url,
|
|
'title': title,
|
|
'content': '',
|
|
'thumbnail_src': thumbnail_src,
|
|
'img_src': img_src})
|
|
|
|
elif (category_to_keyword.get(categories[0], '') == 'news' or
|
|
category_to_keyword.get(categories[0], '') == 'social'):
|
|
published_date = datetime.fromtimestamp(result['date'], None)
|
|
|
|
results.append({'url': res_url,
|
|
'title': title,
|
|
'publishedDate': published_date,
|
|
'content': content})
|
|
|
|
return results
|
|
|
|
|
|
# get supported languages from their site
|
|
def _fetch_supported_languages(resp):
|
|
# list of regions is embedded in page as a js object
|
|
response_text = resp.text
|
|
response_text = response_text[response_text.find('regionalisation'):]
|
|
response_text = response_text[response_text.find('{'):response_text.find(');')]
|
|
|
|
regions_json = loads(response_text)
|
|
|
|
supported_languages = []
|
|
for lang in regions_json['languages'].values():
|
|
if lang['code'] == 'nb':
|
|
lang['code'] = 'no'
|
|
for country in lang['countries']:
|
|
supported_languages.append(lang['code'] + '-' + country)
|
|
|
|
return supported_languages
|