mirror of
				https://github.com/searxng/searxng.git
				synced 2025-10-30 18:22:31 -04:00 
			
		
		
		
	In 25.1.0 [2] an old bug has been fixed: "Docstring formatting does not apply to module docstrings" [3]. [1] https://github.com/psf/black/blob/main/CHANGES.md#2590 [2] https://github.com/psf/black/blob/main/CHANGES.md#2510 [3] https://github.com/psf/black/issues/4094 Signed-off-by: Markus Heiser <markus.heiser@darmarit.de>
		
			
				
	
	
		
			373 lines
		
	
	
		
			11 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			373 lines
		
	
	
		
			11 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
| # SPDX-License-Identifier: AGPL-3.0-or-later
 | |
| """This module implements functions needed for the autocompleter."""
 | |
| # pylint: disable=use-dict-literal
 | |
| 
 | |
| import json
 | |
| import html
 | |
| import typing as t
 | |
| from urllib.parse import urlencode, quote_plus
 | |
| 
 | |
| import lxml.etree
 | |
| import lxml.html
 | |
| from httpx import HTTPError
 | |
| 
 | |
| from searx import settings
 | |
| from searx.engines import (
 | |
|     engines,
 | |
|     google,
 | |
| )
 | |
| from searx.network import get as http_get, post as http_post  # pyright: ignore[reportUnknownVariableType]
 | |
| from searx.exceptions import SearxEngineResponseException
 | |
| from searx.utils import extr, gen_useragent
 | |
| 
 | |
| if t.TYPE_CHECKING:
 | |
|     from searx.extended_types import SXNG_Response
 | |
| 
 | |
| 
 | |
| def update_kwargs(**kwargs) -> None:  # type: ignore
 | |
|     if 'timeout' not in kwargs:
 | |
|         kwargs['timeout'] = settings['outgoing']['request_timeout']
 | |
|     kwargs['raise_for_httperror'] = True
 | |
| 
 | |
| 
 | |
| def get(*args, **kwargs) -> "SXNG_Response":  # type: ignore
 | |
|     update_kwargs(**kwargs)  # pyright: ignore[reportUnknownArgumentType]
 | |
|     return http_get(*args, **kwargs)  # pyright: ignore[reportUnknownArgumentType]
 | |
| 
 | |
| 
 | |
| def post(*args, **kwargs) -> "SXNG_Response":  # type: ignore
 | |
|     update_kwargs(**kwargs)  # pyright: ignore[reportUnknownArgumentType]
 | |
|     return http_post(*args, **kwargs)  # pyright: ignore[reportUnknownArgumentType]
 | |
| 
 | |
| 
 | |
| def baidu(query: str, _sxng_locale: str) -> list[str]:
 | |
|     # baidu search autocompleter
 | |
|     base_url = "https://www.baidu.com/sugrec?"
 | |
|     response = get(base_url + urlencode({'ie': 'utf-8', 'json': 1, 'prod': 'pc', 'wd': query}))
 | |
|     results: list[str] = []
 | |
| 
 | |
|     if response.ok:
 | |
|         data: dict[str, t.Any] = response.json()
 | |
|         if 'g' in data:
 | |
|             for item in data['g']:
 | |
|                 results.append(item['q'])
 | |
|     return results
 | |
| 
 | |
| 
 | |
| def brave(query: str, _sxng_locale: str) -> list[str]:
 | |
|     # brave search autocompleter
 | |
|     url = 'https://search.brave.com/api/suggest?'
 | |
|     url += urlencode({'q': query})
 | |
|     country = 'all'
 | |
|     kwargs = {'cookies': {'country': country}}
 | |
|     resp = get(url, **kwargs)
 | |
|     results: list[str] = []
 | |
| 
 | |
|     if resp.ok:
 | |
|         data: list[list[str]] = resp.json()
 | |
|         for item in data[1]:
 | |
|             results.append(item)
 | |
|     return results
 | |
| 
 | |
| 
 | |
| def dbpedia(query: str, _sxng_locale: str) -> list[str]:
 | |
|     autocomplete_url = 'https://lookup.dbpedia.org/api/search.asmx/KeywordSearch?'
 | |
|     resp = get(autocomplete_url + urlencode(dict(QueryString=query)))
 | |
|     results: list[str] = []
 | |
| 
 | |
|     if resp.ok:
 | |
|         dom = lxml.etree.fromstring(resp.content)
 | |
|         results = [str(x) for x in dom.xpath('//Result/Label//text()')]
 | |
| 
 | |
|     return results
 | |
| 
 | |
| 
 | |
| def duckduckgo(query: str, sxng_locale: str) -> list[str]:
 | |
|     """Autocomplete from DuckDuckGo. Supports DuckDuckGo's languages"""
 | |
| 
 | |
|     traits = engines['duckduckgo'].traits
 | |
|     args: dict[str, str] = {
 | |
|         'q': query,
 | |
|         'kl': traits.get_region(sxng_locale, traits.all_locale),
 | |
|     }
 | |
| 
 | |
|     url = 'https://duckduckgo.com/ac/?type=list&' + urlencode(args)
 | |
|     resp = get(url)
 | |
|     results: list[str] = []
 | |
| 
 | |
|     if resp.ok:
 | |
|         j = resp.json()
 | |
|         if len(j) > 1:
 | |
|             results = j[1]
 | |
|     return results
 | |
| 
 | |
| 
 | |
| def google_complete(query: str, sxng_locale: str) -> list[str]:
 | |
|     """Autocomplete from Google.  Supports Google's languages and subdomains
 | |
|     (:py:obj:`searx.engines.google.get_google_info`) by using the async REST
 | |
|     API::
 | |
| 
 | |
|         https://{subdomain}/complete/search?{args}
 | |
| 
 | |
|     """
 | |
| 
 | |
|     google_info: dict[str, t.Any] = google.get_google_info({'searxng_locale': sxng_locale}, engines['google'].traits)
 | |
|     url = 'https://{subdomain}/complete/search?{args}'
 | |
|     args = urlencode(
 | |
|         {
 | |
|             'q': query,
 | |
|             'client': 'gws-wiz',
 | |
|             'hl': google_info['params']['hl'],
 | |
|         }
 | |
|     )
 | |
|     results: list[str] = []
 | |
| 
 | |
|     resp = get(url.format(subdomain=google_info['subdomain'], args=args))
 | |
|     if resp and resp.ok:
 | |
|         json_txt = resp.text[resp.text.find('[') : resp.text.find(']', -3) + 1]
 | |
|         data = json.loads(json_txt)
 | |
|         for item in data[0]:
 | |
|             results.append(lxml.html.fromstring(item[0]).text_content())
 | |
|     return results
 | |
| 
 | |
| 
 | |
| def mwmbl(query: str, _sxng_locale: str) -> list[str]:
 | |
|     """Autocomplete from Mwmbl_."""
 | |
| 
 | |
|     # mwmbl autocompleter
 | |
|     url = 'https://api.mwmbl.org/search/complete?{query}'
 | |
| 
 | |
|     results: list[str] = get(url.format(query=urlencode({'q': query}))).json()[1]
 | |
| 
 | |
|     # results starting with `go:` are direct urls and not useful for auto completion
 | |
|     return [result for result in results if not result.startswith("go: ") and not result.startswith("search: ")]
 | |
| 
 | |
| 
 | |
| def naver(query: str, _sxng_locale: str) -> list[str]:
 | |
|     # Naver search autocompleter
 | |
|     url = f"https://ac.search.naver.com/nx/ac?{urlencode({'q': query, 'r_format': 'json', 'st': 0})}"
 | |
|     response = get(url)
 | |
|     results: list[str] = []
 | |
| 
 | |
|     if response.ok:
 | |
|         data: dict[str, t.Any] = response.json()
 | |
|         if data.get('items'):
 | |
|             for item in data['items'][0]:
 | |
|                 results.append(item[0])
 | |
|     return results
 | |
| 
 | |
| 
 | |
| def qihu360search(query: str, _sxng_locale: str) -> list[str]:
 | |
|     # 360Search search autocompleter
 | |
|     url = f"https://sug.so.360.cn/suggest?{urlencode({'format': 'json', 'word': query})}"
 | |
|     response = get(url)
 | |
|     results: list[str] = []
 | |
| 
 | |
|     if response.ok:
 | |
|         data: dict[str, t.Any] = response.json()
 | |
|         if 'result' in data:
 | |
|             for item in data['result']:
 | |
|                 results.append(item['word'])
 | |
|     return results
 | |
| 
 | |
| 
 | |
| def quark(query: str, _sxng_locale: str) -> list[str]:
 | |
|     # Quark search autocompleter
 | |
|     url = f"https://sugs.m.sm.cn/web?{urlencode({'q': query})}"
 | |
|     response = get(url)
 | |
|     results: list[str] = []
 | |
| 
 | |
|     if response.ok:
 | |
|         data = response.json()
 | |
|         for item in data.get('r', []):
 | |
|             results.append(item['w'])
 | |
|     return results
 | |
| 
 | |
| 
 | |
| def seznam(query: str, _sxng_locale: str) -> list[str]:
 | |
|     # seznam search autocompleter
 | |
|     url = 'https://suggest.seznam.cz/fulltext/cs?{query}'
 | |
|     resp = get(
 | |
|         url.format(
 | |
|             query=urlencode(
 | |
|                 {'phrase': query, 'cursorPosition': len(query), 'format': 'json-2', 'highlight': '1', 'count': '6'}
 | |
|             )
 | |
|         )
 | |
|     )
 | |
|     results: list[str] = []
 | |
| 
 | |
|     if resp.ok:
 | |
|         data = resp.json()
 | |
|         results = [
 | |
|             ''.join([part.get('text', '') for part in item.get('text', [])])
 | |
|             for item in data.get('result', [])
 | |
|             if item.get('itemType', None) == 'ItemType.TEXT'
 | |
|         ]
 | |
|     return results
 | |
| 
 | |
| 
 | |
| def sogou(query: str, _sxng_locale: str) -> list[str]:
 | |
|     # Sogou search autocompleter
 | |
|     base_url = "https://sor.html5.qq.com/api/getsug?"
 | |
|     resp = get(base_url + urlencode({'m': 'searxng', 'key': query}))
 | |
|     results: list[str] = []
 | |
| 
 | |
|     if resp.ok:
 | |
|         raw_json = extr(resp.text, "[", "]", default="")
 | |
|         try:
 | |
|             data = json.loads(f"[{raw_json}]]")
 | |
|             results = data[1]
 | |
|         except json.JSONDecodeError:
 | |
|             pass
 | |
|     return results
 | |
| 
 | |
| 
 | |
| def startpage(query: str, sxng_locale: str) -> list[str]:
 | |
|     """Autocomplete from Startpage's Firefox extension.
 | |
|     Supports the languages specified in lang_map.
 | |
|     """
 | |
| 
 | |
|     lang_map = {
 | |
|         'da': 'dansk',
 | |
|         'de': 'deutsch',
 | |
|         'en': 'english',
 | |
|         'es': 'espanol',
 | |
|         'fr': 'francais',
 | |
|         'nb': 'norsk',
 | |
|         'nl': 'nederlands',
 | |
|         'pl': 'polski',
 | |
|         'pt': 'portugues',
 | |
|         'sv': 'svenska',
 | |
|     }
 | |
| 
 | |
|     base_lang = sxng_locale.split('-')[0]
 | |
|     lui = lang_map.get(base_lang, 'english')
 | |
| 
 | |
|     url_params = {
 | |
|         'q': query,
 | |
|         'format': 'opensearch',
 | |
|         'segment': 'startpage.defaultffx',
 | |
|         'lui': lui,
 | |
|     }
 | |
|     url = f'https://www.startpage.com/suggestions?{urlencode(url_params)}'
 | |
| 
 | |
|     # Needs user agent, returns a 204 otherwise
 | |
|     h = {'User-Agent': gen_useragent()}
 | |
| 
 | |
|     resp = get(url, headers=h)
 | |
|     results: list[str] = []
 | |
| 
 | |
|     if resp.ok:
 | |
|         try:
 | |
|             data = resp.json()
 | |
|             if len(data) >= 2 and isinstance(data[1], list):
 | |
|                 results = data[1]
 | |
|         except json.JSONDecodeError:
 | |
|             pass
 | |
| 
 | |
|     return results
 | |
| 
 | |
| 
 | |
| def stract(query: str, _sxng_locale: str) -> list[str]:
 | |
|     # stract autocompleter (beta)
 | |
|     url = f"https://stract.com/beta/api/autosuggest?q={quote_plus(query)}"
 | |
|     resp = post(url)
 | |
|     results: list[str] = []
 | |
| 
 | |
|     if resp.ok:
 | |
|         results = [html.unescape(suggestion['raw']) for suggestion in resp.json()]
 | |
| 
 | |
|     return results
 | |
| 
 | |
| 
 | |
| def swisscows(query: str, _sxng_locale: str) -> list[str]:
 | |
|     # swisscows autocompleter
 | |
|     url = 'https://swisscows.ch/api/suggest?{query}&itemsCount=5'
 | |
|     results: list[str] = json.loads(get(url.format(query=urlencode({'query': query}))).text)
 | |
|     return results
 | |
| 
 | |
| 
 | |
| def qwant(query: str, sxng_locale: str) -> list[str]:
 | |
|     """Autocomplete from Qwant. Supports Qwant's regions."""
 | |
|     locale = engines['qwant'].traits.get_region(sxng_locale, 'en_US')
 | |
|     url = 'https://api.qwant.com/v3/suggest?{query}'
 | |
|     resp = get(url.format(query=urlencode({'q': query, 'locale': locale, 'version': '2'})))
 | |
|     results: list[str] = []
 | |
| 
 | |
|     if resp.ok:
 | |
|         data = resp.json()
 | |
|         if data['status'] == 'success':
 | |
|             for item in data['data']['items']:
 | |
|                 results.append(item['value'])
 | |
| 
 | |
|     return results
 | |
| 
 | |
| 
 | |
| def wikipedia(query: str, sxng_locale: str) -> list[str]:
 | |
|     """Autocomplete from Wikipedia. Supports Wikipedia's languages (aka netloc)."""
 | |
|     eng_traits = engines['wikipedia'].traits
 | |
|     wiki_lang = eng_traits.get_language(sxng_locale, 'en')
 | |
|     wiki_netloc: str = eng_traits.custom['wiki_netloc'].get(wiki_lang, 'en.wikipedia.org')  # type: ignore
 | |
| 
 | |
|     args = urlencode(
 | |
|         {
 | |
|             'action': 'opensearch',
 | |
|             'format': 'json',
 | |
|             'formatversion': '2',
 | |
|             'search': query,
 | |
|             'namespace': '0',
 | |
|             'limit': '10',
 | |
|         }
 | |
|     )
 | |
|     resp = get(f'https://{wiki_netloc}/w/api.php?{args}')
 | |
|     results: list[str] = []
 | |
| 
 | |
|     if resp.ok:
 | |
|         data = resp.json()
 | |
|         if len(data) > 1:
 | |
|             results = data[1]
 | |
| 
 | |
|     return results
 | |
| 
 | |
| 
 | |
| def yandex(query: str, _sxng_locale: str) -> list[str]:
 | |
|     # yandex autocompleter
 | |
|     url = "https://suggest.yandex.com/suggest-ff.cgi?{0}"
 | |
|     resp = json.loads(get(url.format(urlencode(dict(part=query)))).text)
 | |
|     results: list[str] = []
 | |
| 
 | |
|     if len(resp) > 1:
 | |
|         results = resp[1]
 | |
|     return results
 | |
| 
 | |
| 
 | |
| backends: dict[str, t.Callable[[str, str], list[str]]] = {
 | |
|     '360search': qihu360search,
 | |
|     'baidu': baidu,
 | |
|     'brave': brave,
 | |
|     'dbpedia': dbpedia,
 | |
|     'duckduckgo': duckduckgo,
 | |
|     'google': google_complete,
 | |
|     'mwmbl': mwmbl,
 | |
|     'naver': naver,
 | |
|     'quark': quark,
 | |
|     'qwant': qwant,
 | |
|     'seznam': seznam,
 | |
|     'sogou': sogou,
 | |
|     'startpage': startpage,
 | |
|     'stract': stract,
 | |
|     'swisscows': swisscows,
 | |
|     'wikipedia': wikipedia,
 | |
|     'yandex': yandex,
 | |
| }
 | |
| 
 | |
| 
 | |
| def search_autocomplete(backend_name: str, query: str, sxng_locale: str) -> list[str]:
 | |
|     backend = backends.get(backend_name)
 | |
|     if backend is None:
 | |
|         return []
 | |
|     try:
 | |
|         return backend(query, sxng_locale)
 | |
|     except (HTTPError, SearxEngineResponseException):
 | |
|         return []
 |