mirror of
				https://github.com/searxng/searxng.git
				synced 2025-11-03 19:17:07 -05:00 
			
		
		
		
	
		
			
				
	
	
		
			68 lines
		
	
	
		
			1.7 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			68 lines
		
	
	
		
			1.7 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
"""
 | 
						|
 Google (News)
 | 
						|
 | 
						|
 @website     https://www.google.com
 | 
						|
 @provide-api yes (https://developers.google.com/web-search/docs/),
 | 
						|
              deprecated!
 | 
						|
 | 
						|
 @using-api   yes
 | 
						|
 @results     JSON
 | 
						|
 @stable      yes (but deprecated)
 | 
						|
 @parse       url, title, content, publishedDate
 | 
						|
"""
 | 
						|
 | 
						|
from urllib import urlencode
 | 
						|
from json import loads
 | 
						|
from dateutil import parser
 | 
						|
 | 
						|
# search-url
 | 
						|
categories = ['news']
 | 
						|
paging = True
 | 
						|
language_support = True
 | 
						|
 | 
						|
# engine dependent config
 | 
						|
url = 'https://ajax.googleapis.com/'
 | 
						|
search_url = url + 'ajax/services/search/news?v=2.0&start={offset}&rsz=large&safe=off&filter=off&{query}&hl={lang}'
 | 
						|
 | 
						|
 | 
						|
# do search-request
 | 
						|
def request(query, params):
 | 
						|
    offset = (params['pageno'] - 1) * 8
 | 
						|
 | 
						|
    language = 'en-US'
 | 
						|
    if params['language'] != 'all':
 | 
						|
        language = params['language'].replace('_', '-')
 | 
						|
 | 
						|
    params['url'] = search_url.format(offset=offset,
 | 
						|
                                      query=urlencode({'q': query}),
 | 
						|
                                      lang=language)
 | 
						|
 | 
						|
    return params
 | 
						|
 | 
						|
 | 
						|
# get response from search-request
 | 
						|
def response(resp):
 | 
						|
    results = []
 | 
						|
 | 
						|
    search_res = loads(resp.text)
 | 
						|
 | 
						|
    # return empty array if there are no results
 | 
						|
    if not search_res.get('responseData', {}).get('results'):
 | 
						|
        return []
 | 
						|
 | 
						|
    # parse results
 | 
						|
    for result in search_res['responseData']['results']:
 | 
						|
        # parse publishedDate
 | 
						|
        publishedDate = parser.parse(result['publishedDate'])
 | 
						|
        if 'url' not in result:
 | 
						|
            continue
 | 
						|
 | 
						|
        # append result
 | 
						|
        results.append({'url': result['unescapedUrl'],
 | 
						|
                        'title': result['titleNoFormatting'],
 | 
						|
                        'publishedDate': publishedDate,
 | 
						|
                        'content': result['content']})
 | 
						|
 | 
						|
    # return results
 | 
						|
    return results
 |