mirror of
				https://github.com/searxng/searxng.git
				synced 2025-11-04 03:27:06 -05:00 
			
		
		
		
	
		
			
				
	
	
		
			143 lines
		
	
	
		
			4.2 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			143 lines
		
	
	
		
			4.2 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
# SPDX-License-Identifier: AGPL-3.0-or-later
 | 
						|
"""Flickr (Images)
 | 
						|
 | 
						|
"""
 | 
						|
 | 
						|
from typing import TYPE_CHECKING
 | 
						|
 | 
						|
import json
 | 
						|
from time import time
 | 
						|
import re
 | 
						|
from urllib.parse import urlencode
 | 
						|
from searx.utils import ecma_unescape, html_to_text
 | 
						|
 | 
						|
if TYPE_CHECKING:
 | 
						|
    import logging
 | 
						|
 | 
						|
    logger: logging.Logger
 | 
						|
 | 
						|
# about
 | 
						|
about = {
 | 
						|
    "website": 'https://www.flickr.com',
 | 
						|
    "wikidata_id": 'Q103204',
 | 
						|
    "official_api_documentation": 'https://secure.flickr.com/services/api/flickr.photos.search.html',
 | 
						|
    "use_official_api": False,
 | 
						|
    "require_api_key": False,
 | 
						|
    "results": 'HTML',
 | 
						|
}
 | 
						|
 | 
						|
# engine dependent config
 | 
						|
categories = ['images']
 | 
						|
paging = True
 | 
						|
time_range_support = True
 | 
						|
safesearch = False
 | 
						|
 | 
						|
time_range_dict = {
 | 
						|
    'day': 60 * 60 * 24,
 | 
						|
    'week': 60 * 60 * 24 * 7,
 | 
						|
    'month': 60 * 60 * 24 * 7 * 4,
 | 
						|
    'year': 60 * 60 * 24 * 7 * 52,
 | 
						|
}
 | 
						|
image_sizes = ('o', 'k', 'h', 'b', 'c', 'z', 'm', 'n', 't', 'q', 's')
 | 
						|
 | 
						|
search_url = 'https://www.flickr.com/search?{query}&page={page}'
 | 
						|
time_range_url = '&min_upload_date={start}&max_upload_date={end}'
 | 
						|
photo_url = 'https://www.flickr.com/photos/{userid}/{photoid}'
 | 
						|
modelexport_re = re.compile(r"^\s*modelExport:\s*({.*}),$", re.M)
 | 
						|
 | 
						|
 | 
						|
def build_flickr_url(user_id, photo_id):
 | 
						|
    return photo_url.format(userid=user_id, photoid=photo_id)
 | 
						|
 | 
						|
 | 
						|
def _get_time_range_url(time_range):
 | 
						|
    if time_range in time_range_dict:
 | 
						|
        return time_range_url.format(start=time(), end=str(int(time()) - time_range_dict[time_range]))
 | 
						|
    return ''
 | 
						|
 | 
						|
 | 
						|
def request(query, params):
 | 
						|
    params['url'] = search_url.format(query=urlencode({'text': query}), page=params['pageno']) + _get_time_range_url(
 | 
						|
        params['time_range']
 | 
						|
    )
 | 
						|
    return params
 | 
						|
 | 
						|
 | 
						|
def response(resp):  # pylint: disable=too-many-branches
 | 
						|
    results = []
 | 
						|
 | 
						|
    matches = modelexport_re.search(resp.text)
 | 
						|
    if matches is None:
 | 
						|
        return results
 | 
						|
 | 
						|
    match = matches.group(1)
 | 
						|
    model_export = json.loads(match)
 | 
						|
 | 
						|
    if 'legend' not in model_export:
 | 
						|
        return results
 | 
						|
    legend = model_export['legend']
 | 
						|
 | 
						|
    # handle empty page
 | 
						|
    if not legend or not legend[0]:
 | 
						|
        return results
 | 
						|
 | 
						|
    for x, index in enumerate(legend):
 | 
						|
        if len(index) != 8:
 | 
						|
            logger.debug("skip legend enty %s : %s", x, index)
 | 
						|
            continue
 | 
						|
 | 
						|
        photo = model_export['main'][index[0]][int(index[1])][index[2]][index[3]][index[4]][index[5]][int(index[6])][
 | 
						|
            index[7]
 | 
						|
        ]
 | 
						|
        author = ecma_unescape(photo.get('realname', ''))
 | 
						|
        source = ecma_unescape(photo.get('username', ''))
 | 
						|
        if source:
 | 
						|
            source += ' @ Flickr'
 | 
						|
        title = ecma_unescape(photo.get('title', ''))
 | 
						|
        content = html_to_text(ecma_unescape(photo.get('description', '')))
 | 
						|
        img_src = None
 | 
						|
 | 
						|
        # From the biggest to the lowest format
 | 
						|
        size_data = None
 | 
						|
        for image_size in image_sizes:
 | 
						|
            if image_size in photo['sizes']['data']:
 | 
						|
                size_data = photo['sizes']['data'][image_size]['data']
 | 
						|
                break
 | 
						|
 | 
						|
        if not size_data:
 | 
						|
            logger.debug('cannot find valid image size: {0}'.format(repr(photo['sizes']['data'])))
 | 
						|
            continue
 | 
						|
 | 
						|
        img_src = size_data['url']
 | 
						|
        resolution = f"{size_data['width']} x {size_data['height']}"
 | 
						|
 | 
						|
        # For a bigger thumbnail, keep only the url_z, not the url_n
 | 
						|
        if 'n' in photo['sizes']['data']:
 | 
						|
            thumbnail_src = photo['sizes']['data']['n']['data']['url']
 | 
						|
        elif 'z' in photo['sizes']['data']:
 | 
						|
            thumbnail_src = photo['sizes']['data']['z']['data']['url']
 | 
						|
        else:
 | 
						|
            thumbnail_src = img_src
 | 
						|
 | 
						|
        if 'ownerNsid' not in photo:
 | 
						|
            # should not happen, disowned photo? Show it anyway
 | 
						|
            url = img_src
 | 
						|
        else:
 | 
						|
            url = build_flickr_url(photo['ownerNsid'], photo['id'])
 | 
						|
 | 
						|
        result = {
 | 
						|
            'url': url,
 | 
						|
            'img_src': img_src,
 | 
						|
            'thumbnail_src': thumbnail_src,
 | 
						|
            'source': source,
 | 
						|
            'resolution': resolution,
 | 
						|
            'template': 'images.html',
 | 
						|
        }
 | 
						|
        result['author'] = author.encode(errors='ignore').decode()
 | 
						|
        result['source'] = source.encode(errors='ignore').decode()
 | 
						|
        result['title'] = title.encode(errors='ignore').decode()
 | 
						|
        result['content'] = content.encode(errors='ignore').decode()
 | 
						|
        results.append(result)
 | 
						|
 | 
						|
    return results
 |