mirror of
https://github.com/searxng/searxng
synced 2024-11-09 01:10:26 +00:00
3d96a9839a
This patch was generated by black [1]:: make format.python [1] https://github.com/psf/black Signed-off-by: Markus Heiser <markus.heiser@darmarit.de>
82 lines
2.0 KiB
Python
82 lines
2.0 KiB
Python
# SPDX-License-Identifier: AGPL-3.0-or-later
|
|
# lint: pylint
|
|
"""CORE (science)
|
|
|
|
"""
|
|
|
|
from json import loads
|
|
from datetime import datetime
|
|
from urllib.parse import urlencode
|
|
|
|
from searx.exceptions import SearxEngineAPIException
|
|
|
|
about = {
|
|
"website": 'https://core.ac.uk',
|
|
"wikidata_id": 'Q22661180',
|
|
"official_api_documentation": 'https://core.ac.uk/documentation/api/',
|
|
"use_official_api": True,
|
|
"require_api_key": True,
|
|
"results": 'JSON',
|
|
}
|
|
|
|
categories = ['science']
|
|
paging = True
|
|
nb_per_page = 10
|
|
|
|
api_key = 'unset'
|
|
|
|
base_url = 'https://core.ac.uk:443/api-v2/search/'
|
|
search_string = '{query}?page={page}&pageSize={nb_per_page}&apiKey={apikey}'
|
|
|
|
|
|
def request(query, params):
|
|
|
|
if api_key == 'unset':
|
|
raise SearxEngineAPIException('missing CORE API key')
|
|
|
|
search_path = search_string.format(
|
|
query=urlencode({'q': query}),
|
|
nb_per_page=nb_per_page,
|
|
page=params['pageno'],
|
|
apikey=api_key,
|
|
)
|
|
params['url'] = base_url + search_path
|
|
|
|
logger.debug("query_url --> %s", params['url'])
|
|
return params
|
|
|
|
|
|
def response(resp):
|
|
results = []
|
|
json_data = loads(resp.text)
|
|
|
|
for result in json_data['data']:
|
|
|
|
source = result['_source']
|
|
time = source['publishedDate'] or source['depositedDate']
|
|
if time:
|
|
date = datetime.fromtimestamp(time / 1000)
|
|
else:
|
|
date = None
|
|
|
|
metadata = []
|
|
if source['publisher'] and len(source['publisher']) > 3:
|
|
metadata.append(source['publisher'])
|
|
if source['topics']:
|
|
metadata.append(source['topics'][0])
|
|
if source['doi']:
|
|
metadata.append(source['doi'])
|
|
metadata = ' / '.join(metadata)
|
|
|
|
results.append(
|
|
{
|
|
'url': source['urls'][0].replace('http://', 'https://', 1),
|
|
'title': source['title'],
|
|
'content': source['description'],
|
|
'publishedDate': date,
|
|
'metadata': metadata,
|
|
}
|
|
)
|
|
|
|
return results
|