mirror of
				https://github.com/searxng/searxng.git
				synced 2025-10-30 18:22:31 -04:00 
			
		
		
		
	Suggested-by: @dalf https://github.com/searxng/searxng/issues/102#issuecomment-914168470 Signed-off-by: Markus Heiser <markus.heiser@darmarit.de>
		
			
				
	
	
		
			78 lines
		
	
	
		
			2.0 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			78 lines
		
	
	
		
			2.0 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
| # SPDX-License-Identifier: AGPL-3.0-or-later
 | |
| # lint: pylint
 | |
| """CORE (science)
 | |
| 
 | |
| """
 | |
| 
 | |
| from json import loads
 | |
| from datetime import datetime
 | |
| from urllib.parse import urlencode
 | |
| 
 | |
| from searx.exceptions import SearxEngineAPIException
 | |
| 
 | |
| about = {
 | |
|     "website": 'https://core.ac.uk',
 | |
|     "wikidata_id": 'Q22661180',
 | |
|     "official_api_documentation": 'https://core.ac.uk/documentation/api/',
 | |
|     "use_official_api": True,
 | |
|     "require_api_key": True,
 | |
|     "results": 'JSON',
 | |
| }
 | |
| 
 | |
| categories = ['science']
 | |
| paging = True
 | |
| nb_per_page = 10
 | |
| 
 | |
| api_key = 'unset'
 | |
| 
 | |
| base_url = 'https://core.ac.uk:443/api-v2/search/'
 | |
| search_string = '{query}?page={page}&pageSize={nb_per_page}&apiKey={apikey}'
 | |
| 
 | |
| def request(query, params):
 | |
| 
 | |
|     if api_key == 'unset':
 | |
|         raise SearxEngineAPIException('missing CORE API key')
 | |
| 
 | |
|     search_path = search_string.format(
 | |
|         query = urlencode({'q': query}),
 | |
|         nb_per_page = nb_per_page,
 | |
|         page = params['pageno'],
 | |
|         apikey = api_key,
 | |
|     )
 | |
|     params['url'] = base_url + search_path
 | |
| 
 | |
|     logger.debug("query_url --> %s", params['url'])
 | |
|     return params
 | |
| 
 | |
| def response(resp):
 | |
|     results = []
 | |
|     json_data = loads(resp.text)
 | |
| 
 | |
|     for result in json_data['data']:
 | |
| 
 | |
|         source = result['_source']
 | |
|         time = source['publishedDate'] or source['depositedDate']
 | |
|         if time :
 | |
|             date = datetime.fromtimestamp(time / 1000)
 | |
|         else:
 | |
|             date = None
 | |
| 
 | |
|         metadata = []
 | |
|         if source['publisher'] and len(source['publisher']) > 3:
 | |
|             metadata.append(source['publisher'])
 | |
|         if source['topics']:
 | |
|             metadata.append(source['topics'][0])
 | |
|         if source['doi']:
 | |
|             metadata.append(source['doi'])
 | |
|         metadata = ' / '.join(metadata)
 | |
| 
 | |
|         results.append({
 | |
|             'url': source['urls'][0].replace('http://', 'https://', 1),
 | |
|             'title': source['title'],
 | |
|             'content': source['description'],
 | |
|             'publishedDate': date,
 | |
|             'metadata' : metadata,
 | |
|         })
 | |
| 
 | |
|     return results
 |