[fix] engine: core.ac.uk implement API v3 / v2 is no longer supported
This commit is contained in:
parent
82b9e533b2
commit
40feede51e
13
docs/dev/engines/online/core.rst
Normal file
13
docs/dev/engines/online/core.rst
Normal file
@ -0,0 +1,13 @@
|
|||||||
|
.. _core engine:
|
||||||
|
|
||||||
|
====
|
||||||
|
CORE
|
||||||
|
====
|
||||||
|
|
||||||
|
.. contents::
|
||||||
|
:depth: 2
|
||||||
|
:local:
|
||||||
|
:backlinks: entry
|
||||||
|
|
||||||
|
.. automodule:: searx.engines.core
|
||||||
|
:members:
|
@ -1,7 +1,33 @@
|
|||||||
# SPDX-License-Identifier: AGPL-3.0-or-later
|
# SPDX-License-Identifier: AGPL-3.0-or-later
|
||||||
"""CORE (science)
|
"""CORE_ (COnnecting REpositories) provides a comprehensive bibliographic
|
||||||
|
database of the world’s scholarly literature, collecting and indexing
|
||||||
|
research from repositories and journals.
|
||||||
|
|
||||||
|
.. _CORE: https://core.ac.uk/about
|
||||||
|
|
||||||
|
.. _core engine config:
|
||||||
|
|
||||||
|
Configuration
|
||||||
|
=============
|
||||||
|
|
||||||
|
The engine has the following additional settings:
|
||||||
|
|
||||||
|
- :py:obj:`api_key`
|
||||||
|
|
||||||
|
.. code:: yaml
|
||||||
|
|
||||||
|
- name: core.ac.uk
|
||||||
|
engine: core
|
||||||
|
categories: science
|
||||||
|
shortcut: cor
|
||||||
|
api_key: "..."
|
||||||
|
timeout: 5
|
||||||
|
|
||||||
|
Implementations
|
||||||
|
===============
|
||||||
|
|
||||||
"""
|
"""
|
||||||
|
# pylint: disable=too-many-branches
|
||||||
|
|
||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
from urllib.parse import urlencode
|
from urllib.parse import urlencode
|
||||||
@ -11,34 +37,36 @@ from searx.exceptions import SearxEngineAPIException
|
|||||||
about = {
|
about = {
|
||||||
"website": 'https://core.ac.uk',
|
"website": 'https://core.ac.uk',
|
||||||
"wikidata_id": 'Q22661180',
|
"wikidata_id": 'Q22661180',
|
||||||
"official_api_documentation": 'https://core.ac.uk/documentation/api/',
|
"official_api_documentation": 'https://api.core.ac.uk/docs/v3',
|
||||||
"use_official_api": True,
|
"use_official_api": True,
|
||||||
"require_api_key": True,
|
"require_api_key": True,
|
||||||
"results": 'JSON',
|
"results": 'JSON',
|
||||||
}
|
}
|
||||||
|
|
||||||
|
api_key = 'unset'
|
||||||
|
"""For an API key register at https://core.ac.uk/services/api and insert
|
||||||
|
the API key in the engine :ref:`core engine config`."""
|
||||||
|
|
||||||
categories = ['science', 'scientific publications']
|
categories = ['science', 'scientific publications']
|
||||||
paging = True
|
paging = True
|
||||||
nb_per_page = 10
|
nb_per_page = 10
|
||||||
|
base_url = 'https://api.core.ac.uk/v3/search/works/'
|
||||||
api_key = 'unset'
|
|
||||||
|
|
||||||
base_url = 'https://core.ac.uk:443/api-v2/search/'
|
|
||||||
search_string = '{query}?page={page}&pageSize={nb_per_page}&apiKey={apikey}'
|
|
||||||
|
|
||||||
|
|
||||||
def request(query, params):
|
def request(query, params):
|
||||||
|
|
||||||
if api_key == 'unset':
|
if api_key == 'unset':
|
||||||
raise SearxEngineAPIException('missing CORE API key')
|
raise SearxEngineAPIException('missing CORE API key')
|
||||||
|
|
||||||
search_path = search_string.format(
|
# API v3 uses different parameters
|
||||||
query=urlencode({'q': query}),
|
search_params = {
|
||||||
nb_per_page=nb_per_page,
|
'q': query,
|
||||||
page=params['pageno'],
|
'offset': (params['pageno'] - 1) * nb_per_page,
|
||||||
apikey=api_key,
|
'limit': nb_per_page,
|
||||||
)
|
'sort': 'relevance',
|
||||||
params['url'] = base_url + search_path
|
}
|
||||||
|
|
||||||
|
params['url'] = base_url + '?' + urlencode(search_params)
|
||||||
|
params['headers'] = {'Authorization': f'Bearer {api_key}'}
|
||||||
|
|
||||||
return params
|
return params
|
||||||
|
|
||||||
@ -47,68 +75,76 @@ def response(resp):
|
|||||||
results = []
|
results = []
|
||||||
json_data = resp.json()
|
json_data = resp.json()
|
||||||
|
|
||||||
for result in json_data['data']:
|
for result in json_data.get('results', []):
|
||||||
source = result['_source']
|
# Get title
|
||||||
url = None
|
if not result.get('title'):
|
||||||
if source.get('urls'):
|
|
||||||
url = source['urls'][0].replace('http://', 'https://', 1)
|
|
||||||
|
|
||||||
if url is None and source.get('doi'):
|
|
||||||
# use the DOI reference
|
|
||||||
url = 'https://doi.org/' + source['doi']
|
|
||||||
|
|
||||||
if url is None and source.get('downloadUrl'):
|
|
||||||
# use the downloadUrl
|
|
||||||
url = source['downloadUrl']
|
|
||||||
|
|
||||||
if url is None and source.get('identifiers'):
|
|
||||||
# try to find an ark id, see
|
|
||||||
# https://www.wikidata.org/wiki/Property:P8091
|
|
||||||
# and https://en.wikipedia.org/wiki/Archival_Resource_Key
|
|
||||||
arkids = [
|
|
||||||
identifier[5:] # 5 is the length of "ark:/"
|
|
||||||
for identifier in source.get('identifiers')
|
|
||||||
if isinstance(identifier, str) and identifier.startswith('ark:/')
|
|
||||||
]
|
|
||||||
if len(arkids) > 0:
|
|
||||||
url = 'https://n2t.net/' + arkids[0]
|
|
||||||
|
|
||||||
if url is None:
|
|
||||||
continue
|
continue
|
||||||
|
|
||||||
publishedDate = None
|
# Get URL - try different options
|
||||||
time = source['publishedDate'] or source['depositedDate']
|
url = None
|
||||||
if time:
|
|
||||||
publishedDate = datetime.fromtimestamp(time / 1000)
|
|
||||||
|
|
||||||
# sometimes the 'title' is None / filter None values
|
# Try DOI first
|
||||||
journals = [j['title'] for j in (source.get('journals') or []) if j['title']]
|
doi = result.get('doi')
|
||||||
|
if doi:
|
||||||
|
url = f'https://doi.org/{doi}'
|
||||||
|
|
||||||
publisher = source['publisher']
|
if url is None and result.get('doi'):
|
||||||
|
# use the DOI reference
|
||||||
|
url = 'https://doi.org/' + str(result['doi'])
|
||||||
|
elif result.get('id'):
|
||||||
|
url = 'https://core.ac.uk/works/' + str(result['id'])
|
||||||
|
elif result.get('downloadUrl'):
|
||||||
|
url = result['downloadUrl']
|
||||||
|
elif result.get('sourceFulltextUrls'):
|
||||||
|
url = result['sourceFulltextUrls']
|
||||||
|
else:
|
||||||
|
continue
|
||||||
|
|
||||||
|
# Published date
|
||||||
|
published_date = None
|
||||||
|
|
||||||
|
raw_date = result.get('publishedDate') or result.get('depositedDate')
|
||||||
|
if raw_date:
|
||||||
|
try:
|
||||||
|
published_date = datetime.fromisoformat(result['publishedDate'].replace('Z', '+00:00'))
|
||||||
|
except (ValueError, AttributeError):
|
||||||
|
pass
|
||||||
|
|
||||||
|
# Handle journals
|
||||||
|
journals = []
|
||||||
|
if result.get('journals'):
|
||||||
|
journals = [j.get('title') for j in result['journals'] if j.get('title')]
|
||||||
|
|
||||||
|
# Handle publisher
|
||||||
|
publisher = result.get('publisher', '').strip("'")
|
||||||
if publisher:
|
if publisher:
|
||||||
publisher = source['publisher'].strip("'")
|
publisher = publisher.strip("'")
|
||||||
|
|
||||||
|
# Handle authors
|
||||||
|
authors = set()
|
||||||
|
for i in result.get('authors', []):
|
||||||
|
name = i.get("name")
|
||||||
|
if name:
|
||||||
|
authors.add(name)
|
||||||
|
|
||||||
results.append(
|
results.append(
|
||||||
{
|
{
|
||||||
'template': 'paper.html',
|
'template': 'paper.html',
|
||||||
'title': source['title'],
|
'title': result.get('title'),
|
||||||
'url': url,
|
'url': url,
|
||||||
'content': source['description'] or '',
|
'content': result.get('fullText', '') or '',
|
||||||
# 'comments': '',
|
# 'comments': '',
|
||||||
'tags': source['topics'],
|
'tags': result.get('fieldOfStudy', []),
|
||||||
'publishedDate': publishedDate,
|
'publishedDate': published_date,
|
||||||
'type': (source['types'] or [None])[0],
|
'type': result.get('documentType', '') or '',
|
||||||
'authors': source['authors'],
|
'authors': authors,
|
||||||
'editor': ', '.join(source['contributors'] or []),
|
'editor': ', '.join(result.get('contributors', [])),
|
||||||
'publisher': publisher,
|
'publisher': publisher,
|
||||||
'journal': ', '.join(journals),
|
'journal': ', '.join(journals),
|
||||||
# 'volume': '',
|
'doi': result.get('doi'),
|
||||||
# 'pages' : '',
|
# 'issn' : ''
|
||||||
# 'number': '',
|
# 'isbn' : ''
|
||||||
'doi': source['doi'],
|
'pdf_url': result.get('downloadUrl', {}) or result.get("sourceFulltextUrls", {}),
|
||||||
'issn': [x for x in [source.get('issn')] if x],
|
|
||||||
'isbn': [x for x in [source.get('isbn')] if x], # exists in the rawRecordXml
|
|
||||||
'pdf_url': source.get('repositoryDocument', {}).get('pdfOrigin'),
|
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -29,7 +29,7 @@ Implementations
|
|||||||
from typing import TYPE_CHECKING
|
from typing import TYPE_CHECKING
|
||||||
|
|
||||||
try:
|
try:
|
||||||
import mariadb
|
import mariadb # pyright: ignore [reportMissingImports]
|
||||||
except ImportError:
|
except ImportError:
|
||||||
# import error is ignored because the admin has to install mysql manually to use
|
# import error is ignored because the admin has to install mysql manually to use
|
||||||
# the engine
|
# the engine
|
||||||
|
Loading…
x
Reference in New Issue
Block a user