mirror of
				https://github.com/searxng/searxng.git
				synced 2025-11-03 19:17:07 -05:00 
			
		
		
		
	
		
			
				
	
	
		
			117 lines
		
	
	
		
			3.5 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			117 lines
		
	
	
		
			3.5 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
# SPDX-License-Identifier: AGPL-3.0-or-later
 | 
						|
# lint: pylint
 | 
						|
"""CORE (science)
 | 
						|
 | 
						|
"""
 | 
						|
 | 
						|
from datetime import datetime
 | 
						|
from urllib.parse import urlencode
 | 
						|
 | 
						|
from searx.exceptions import SearxEngineAPIException
 | 
						|
 | 
						|
about = {
 | 
						|
    "website": 'https://core.ac.uk',
 | 
						|
    "wikidata_id": 'Q22661180',
 | 
						|
    "official_api_documentation": 'https://core.ac.uk/documentation/api/',
 | 
						|
    "use_official_api": True,
 | 
						|
    "require_api_key": True,
 | 
						|
    "results": 'JSON',
 | 
						|
}
 | 
						|
 | 
						|
categories = ['science', 'scientific publications']
 | 
						|
paging = True
 | 
						|
nb_per_page = 10
 | 
						|
 | 
						|
api_key = 'unset'
 | 
						|
 | 
						|
base_url = 'https://core.ac.uk:443/api-v2/search/'
 | 
						|
search_string = '{query}?page={page}&pageSize={nb_per_page}&apiKey={apikey}'
 | 
						|
 | 
						|
 | 
						|
def request(query, params):
 | 
						|
 | 
						|
    if api_key == 'unset':
 | 
						|
        raise SearxEngineAPIException('missing CORE API key')
 | 
						|
 | 
						|
    search_path = search_string.format(
 | 
						|
        query=urlencode({'q': query}),
 | 
						|
        nb_per_page=nb_per_page,
 | 
						|
        page=params['pageno'],
 | 
						|
        apikey=api_key,
 | 
						|
    )
 | 
						|
    params['url'] = base_url + search_path
 | 
						|
 | 
						|
    return params
 | 
						|
 | 
						|
 | 
						|
def response(resp):
 | 
						|
    results = []
 | 
						|
    json_data = resp.json()
 | 
						|
 | 
						|
    for result in json_data['data']:
 | 
						|
        source = result['_source']
 | 
						|
        url = None
 | 
						|
        if source.get('urls'):
 | 
						|
            url = source['urls'][0].replace('http://', 'https://', 1)
 | 
						|
 | 
						|
        if url is None and source.get('doi'):
 | 
						|
            # use the DOI reference
 | 
						|
            url = 'https://doi.org/' + source['doi']
 | 
						|
 | 
						|
        if url is None and source.get('downloadUrl'):
 | 
						|
            # use the downloadUrl
 | 
						|
            url = source['downloadUrl']
 | 
						|
 | 
						|
        if url is None and source.get('identifiers'):
 | 
						|
            # try to find an ark id, see
 | 
						|
            # https://www.wikidata.org/wiki/Property:P8091
 | 
						|
            # and https://en.wikipedia.org/wiki/Archival_Resource_Key
 | 
						|
            arkids = [
 | 
						|
                identifier[5:]  # 5 is the length of "ark:/"
 | 
						|
                for identifier in source.get('identifiers')
 | 
						|
                if isinstance(identifier, str) and identifier.startswith('ark:/')
 | 
						|
            ]
 | 
						|
            if len(arkids) > 0:
 | 
						|
                url = 'https://n2t.net/' + arkids[0]
 | 
						|
 | 
						|
        if url is None:
 | 
						|
            continue
 | 
						|
 | 
						|
        publishedDate = None
 | 
						|
        time = source['publishedDate'] or source['depositedDate']
 | 
						|
        if time:
 | 
						|
            publishedDate = datetime.fromtimestamp(time / 1000)
 | 
						|
 | 
						|
        # sometimes the 'title' is None / filter None values
 | 
						|
        journals = [j['title'] for j in (source.get('journals') or []) if j['title']]
 | 
						|
 | 
						|
        publisher = source['publisher']
 | 
						|
        if publisher:
 | 
						|
            publisher = source['publisher'].strip("'")
 | 
						|
 | 
						|
        results.append(
 | 
						|
            {
 | 
						|
                'template': 'paper.html',
 | 
						|
                'title': source['title'],
 | 
						|
                'url': url,
 | 
						|
                'content': source['description'] or '',
 | 
						|
                # 'comments': '',
 | 
						|
                'tags': source['topics'],
 | 
						|
                'publishedDate': publishedDate,
 | 
						|
                'type': (source['types'] or [None])[0],
 | 
						|
                'authors': source['authors'],
 | 
						|
                'editor': ', '.join(source['contributors'] or []),
 | 
						|
                'publisher': publisher,
 | 
						|
                'journal': ', '.join(journals),
 | 
						|
                # 'volume': '',
 | 
						|
                # 'pages' : '',
 | 
						|
                # 'number': '',
 | 
						|
                'doi': source['doi'],
 | 
						|
                'issn': [x for x in [source.get('issn')] if x],
 | 
						|
                'isbn': [x for x in [source.get('isbn')] if x],  # exists in the rawRecordXml
 | 
						|
                'pdf_url': source.get('repositoryDocument', {}).get('pdfOrigin'),
 | 
						|
            }
 | 
						|
        )
 | 
						|
 | 
						|
    return results
 |