mirror of
				https://github.com/searxng/searxng.git
				synced 2025-11-03 11:07:07 -05:00 
			
		
		
		
	
		
			
				
	
	
		
			154 lines
		
	
	
		
			4.1 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			154 lines
		
	
	
		
			4.1 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
# SPDX-License-Identifier: AGPL-3.0-or-later
 | 
						|
"""This is the implementation of the Google Videos engine.
 | 
						|
 | 
						|
.. admonition:: Content-Security-Policy (CSP)
 | 
						|
 | 
						|
   This engine needs to allow images from the `data URLs`_ (prefixed with the
 | 
						|
   ``data:`` scheme)::
 | 
						|
 | 
						|
     Header set Content-Security-Policy "img-src 'self' data: ;"
 | 
						|
 | 
						|
.. _data URLs:
 | 
						|
   https://developer.mozilla.org/en-US/docs/Web/HTTP/Basics_of_HTTP/Data_URIs
 | 
						|
 | 
						|
"""
 | 
						|
from __future__ import annotations
 | 
						|
 | 
						|
from typing import TYPE_CHECKING
 | 
						|
 | 
						|
from urllib.parse import urlencode
 | 
						|
from lxml import html
 | 
						|
 | 
						|
from searx.utils import (
 | 
						|
    eval_xpath,
 | 
						|
    eval_xpath_list,
 | 
						|
    eval_xpath_getindex,
 | 
						|
    extract_text,
 | 
						|
)
 | 
						|
 | 
						|
from searx.engines.google import fetch_traits  # pylint: disable=unused-import
 | 
						|
from searx.engines.google import (
 | 
						|
    get_google_info,
 | 
						|
    time_range_dict,
 | 
						|
    filter_mapping,
 | 
						|
    suggestion_xpath,
 | 
						|
    detect_google_sorry,
 | 
						|
    ui_async,
 | 
						|
    parse_data_images,
 | 
						|
)
 | 
						|
from searx.enginelib.traits import EngineTraits
 | 
						|
from searx.utils import get_embeded_stream_url
 | 
						|
 | 
						|
if TYPE_CHECKING:
 | 
						|
    import logging
 | 
						|
 | 
						|
    logger: logging.Logger
 | 
						|
 | 
						|
traits: EngineTraits
 | 
						|
 | 
						|
# about
 | 
						|
about = {
 | 
						|
    "website": 'https://www.google.com',
 | 
						|
    "wikidata_id": 'Q219885',
 | 
						|
    "official_api_documentation": 'https://developers.google.com/custom-search',
 | 
						|
    "use_official_api": False,
 | 
						|
    "require_api_key": False,
 | 
						|
    "results": 'HTML',
 | 
						|
}
 | 
						|
 | 
						|
# engine dependent config
 | 
						|
 | 
						|
categories = ['videos', 'web']
 | 
						|
paging = True
 | 
						|
max_page = 50
 | 
						|
"""`Google: max 50 pages`
 | 
						|
 | 
						|
.. _Google: max 50 pages: https://github.com/searxng/searxng/issues/2982
 | 
						|
"""
 | 
						|
language_support = True
 | 
						|
time_range_support = True
 | 
						|
safesearch = True
 | 
						|
 | 
						|
 | 
						|
def request(query, params):
 | 
						|
    """Google-Video search request"""
 | 
						|
 | 
						|
    google_info = get_google_info(params, traits)
 | 
						|
    start = (params['pageno'] - 1) * 10
 | 
						|
 | 
						|
    query_url = (
 | 
						|
        'https://'
 | 
						|
        + google_info['subdomain']
 | 
						|
        + '/search'
 | 
						|
        + "?"
 | 
						|
        + urlencode(
 | 
						|
            {
 | 
						|
                'q': query,
 | 
						|
                'tbm': "vid",
 | 
						|
                'start': 10 * params['pageno'],
 | 
						|
                **google_info['params'],
 | 
						|
                'asearch': 'arc',
 | 
						|
                'async': ui_async(start),
 | 
						|
            }
 | 
						|
        )
 | 
						|
    )
 | 
						|
 | 
						|
    if params['time_range'] in time_range_dict:
 | 
						|
        query_url += '&' + urlencode({'tbs': 'qdr:' + time_range_dict[params['time_range']]})
 | 
						|
    if 'safesearch' in params:
 | 
						|
        query_url += '&' + urlencode({'safe': filter_mapping[params['safesearch']]})
 | 
						|
    params['url'] = query_url
 | 
						|
 | 
						|
    params['cookies'] = google_info['cookies']
 | 
						|
    params['headers'].update(google_info['headers'])
 | 
						|
    return params
 | 
						|
 | 
						|
 | 
						|
def response(resp):
 | 
						|
    """Get response from google's search request"""
 | 
						|
    results = []
 | 
						|
 | 
						|
    detect_google_sorry(resp)
 | 
						|
    data_image_map = parse_data_images(resp.text)
 | 
						|
 | 
						|
    # convert the text to dom
 | 
						|
    dom = html.fromstring(resp.text)
 | 
						|
 | 
						|
    # parse results
 | 
						|
    for result in eval_xpath_list(dom, '//div[contains(@class, "g ")]'):
 | 
						|
 | 
						|
        thumbnail = eval_xpath_getindex(result, './/img/@src', 0, None)
 | 
						|
        if thumbnail:
 | 
						|
            if thumbnail.startswith('data:image'):
 | 
						|
                img_id = eval_xpath_getindex(result, './/img/@id', 0, None)
 | 
						|
                if img_id:
 | 
						|
                    thumbnail = data_image_map.get(img_id)
 | 
						|
        else:
 | 
						|
            thumbnail = None
 | 
						|
 | 
						|
        title = extract_text(eval_xpath_getindex(result, './/a/h3[1]', 0))
 | 
						|
        url = eval_xpath_getindex(result, './/a/h3[1]/../@href', 0)
 | 
						|
 | 
						|
        c_node = eval_xpath_getindex(result, './/div[contains(@class, "ITZIwc")]', 0)
 | 
						|
        content = extract_text(c_node)
 | 
						|
        pub_info = extract_text(eval_xpath(result, './/div[contains(@class, "gqF9jc")]'))
 | 
						|
 | 
						|
        results.append(
 | 
						|
            {
 | 
						|
                'url': url,
 | 
						|
                'title': title,
 | 
						|
                'content': content,
 | 
						|
                'author': pub_info,
 | 
						|
                'thumbnail': thumbnail,
 | 
						|
                'iframe_src': get_embeded_stream_url(url),
 | 
						|
                'template': 'videos.html',
 | 
						|
            }
 | 
						|
        )
 | 
						|
 | 
						|
    # parse suggestion
 | 
						|
    for suggestion in eval_xpath_list(dom, suggestion_xpath):
 | 
						|
        # append suggestion
 | 
						|
        results.append({'suggestion': extract_text(suggestion)})
 | 
						|
 | 
						|
    return results
 |