mirror of
				https://github.com/searxng/searxng.git
				synced 2025-11-04 03:27:06 -05:00 
			
		
		
		
	
		
			
				
	
	
		
			77 lines
		
	
	
		
			2.0 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			77 lines
		
	
	
		
			2.0 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
"""
 | 
						|
 Reddit
 | 
						|
 | 
						|
 @website      https://www.reddit.com/
 | 
						|
 @provide-api  yes (https://www.reddit.com/dev/api)
 | 
						|
 | 
						|
 @using-api    yes
 | 
						|
 @results      JSON
 | 
						|
 @stable       yes
 | 
						|
 @parse        url, title, content, thumbnail, publishedDate
 | 
						|
"""
 | 
						|
 | 
						|
import json
 | 
						|
from datetime import datetime
 | 
						|
from searx.url_utils import urlencode, urljoin, urlparse
 | 
						|
 | 
						|
# engine dependent config
 | 
						|
categories = ['general', 'images', 'news', 'social media']
 | 
						|
page_size = 25
 | 
						|
 | 
						|
# search-url
 | 
						|
base_url = 'https://www.reddit.com/'
 | 
						|
search_url = base_url + 'search.json?{query}'
 | 
						|
 | 
						|
 | 
						|
# do search-request
 | 
						|
def request(query, params):
 | 
						|
    query = urlencode({'q': query, 'limit': page_size})
 | 
						|
    params['url'] = search_url.format(query=query)
 | 
						|
 | 
						|
    return params
 | 
						|
 | 
						|
 | 
						|
# get response from search-request
 | 
						|
def response(resp):
 | 
						|
    img_results = []
 | 
						|
    text_results = []
 | 
						|
 | 
						|
    search_results = json.loads(resp.text)
 | 
						|
 | 
						|
    # return empty array if there are no results
 | 
						|
    if 'data' not in search_results:
 | 
						|
        return []
 | 
						|
 | 
						|
    posts = search_results.get('data', {}).get('children', [])
 | 
						|
 | 
						|
    # process results
 | 
						|
    for post in posts:
 | 
						|
        data = post['data']
 | 
						|
 | 
						|
        # extract post information
 | 
						|
        params = {
 | 
						|
            'url': urljoin(base_url, data['permalink']),
 | 
						|
            'title': data['title']
 | 
						|
        }
 | 
						|
 | 
						|
        # if thumbnail field contains a valid URL, we need to change template
 | 
						|
        thumbnail = data['thumbnail']
 | 
						|
        url_info = urlparse(thumbnail)
 | 
						|
        # netloc & path
 | 
						|
        if url_info[1] != '' and url_info[2] != '':
 | 
						|
            params['img_src'] = data['url']
 | 
						|
            params['thumbnail_src'] = thumbnail
 | 
						|
            params['template'] = 'images.html'
 | 
						|
            img_results.append(params)
 | 
						|
        else:
 | 
						|
            created = datetime.fromtimestamp(data['created_utc'])
 | 
						|
            content = data['selftext']
 | 
						|
            if len(content) > 500:
 | 
						|
                content = content[:500] + '...'
 | 
						|
            params['content'] = content
 | 
						|
            params['publishedDate'] = created
 | 
						|
            text_results.append(params)
 | 
						|
 | 
						|
    # show images first and text results second
 | 
						|
    return img_results + text_results
 |