mirror of
				https://github.com/searxng/searxng.git
				synced 2025-11-04 03:27:06 -05:00 
			
		
		
		
	
		
			
				
	
	
		
			258 lines
		
	
	
		
			8.5 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			258 lines
		
	
	
		
			8.5 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
# SPDX-License-Identifier: AGPL-3.0-or-later
 | 
						|
# lint: pylint
 | 
						|
# pylint: disable=missing-module-docstring, global-statement
 | 
						|
 | 
						|
import asyncio
 | 
						|
import logging
 | 
						|
import threading
 | 
						|
 | 
						|
import anyio
 | 
						|
import httpcore
 | 
						|
import httpx
 | 
						|
from httpx_socks import AsyncProxyTransport
 | 
						|
from python_socks import (
 | 
						|
    parse_proxy_url,
 | 
						|
    ProxyConnectionError,
 | 
						|
    ProxyTimeoutError,
 | 
						|
    ProxyError
 | 
						|
)
 | 
						|
 | 
						|
from searx import logger
 | 
						|
 | 
						|
# Optional uvloop (support Python 3.6)
 | 
						|
try:
 | 
						|
    import uvloop
 | 
						|
except ImportError:
 | 
						|
    pass
 | 
						|
else:
 | 
						|
    uvloop.install()
 | 
						|
 | 
						|
 | 
						|
logger = logger.getChild('searx.network.client')
 | 
						|
LOOP = None
 | 
						|
SSLCONTEXTS = {}
 | 
						|
TRANSPORT_KWARGS = {
 | 
						|
    # use anyio :
 | 
						|
    # * https://github.com/encode/httpcore/issues/344
 | 
						|
    # * https://github.com/encode/httpx/discussions/1511
 | 
						|
    'backend': 'anyio',
 | 
						|
    'trust_env': False,
 | 
						|
}
 | 
						|
 | 
						|
 | 
						|
# pylint: disable=protected-access
 | 
						|
async def close_connections_for_url(
 | 
						|
    connection_pool: httpcore.AsyncConnectionPool, url: httpcore._utils.URL
 | 
						|
):
 | 
						|
 | 
						|
    origin = httpcore._utils.url_to_origin(url)
 | 
						|
    logger.debug('Drop connections for %r', origin)
 | 
						|
    connections_to_close = connection_pool._connections_for_origin(origin)
 | 
						|
    for connection in connections_to_close:
 | 
						|
        await connection_pool._remove_from_pool(connection)
 | 
						|
        try:
 | 
						|
            await connection.aclose()
 | 
						|
        except httpx.NetworkError as e:
 | 
						|
            logger.warning('Error closing an existing connection', exc_info=e)
 | 
						|
# pylint: enable=protected-access
 | 
						|
 | 
						|
 | 
						|
def get_sslcontexts(proxy_url=None, cert=None, verify=True, trust_env=True, http2=False):
 | 
						|
    key = (proxy_url, cert, verify, trust_env, http2)
 | 
						|
    if key not in SSLCONTEXTS:
 | 
						|
        SSLCONTEXTS[key] = httpx.create_ssl_context(cert, verify, trust_env, http2)
 | 
						|
    return SSLCONTEXTS[key]
 | 
						|
 | 
						|
 | 
						|
class AsyncHTTPTransportNoHttp(httpx.AsyncHTTPTransport):
 | 
						|
    """Block HTTP request"""
 | 
						|
 | 
						|
    async def handle_async_request(
 | 
						|
        self, method, url, headers=None, stream=None, extensions=None
 | 
						|
    ):
 | 
						|
        raise httpx.UnsupportedProtocol('HTTP protocol is disabled')
 | 
						|
 | 
						|
 | 
						|
class AsyncProxyTransportFixed(AsyncProxyTransport):
 | 
						|
    """Fix httpx_socks.AsyncProxyTransport
 | 
						|
 | 
						|
    Map python_socks exceptions to httpx.ProxyError / httpx.ConnectError
 | 
						|
 | 
						|
    Map socket.gaierror to httpx.ConnectError
 | 
						|
 | 
						|
    Note: AsyncProxyTransport inherit from AsyncConnectionPool
 | 
						|
    """
 | 
						|
 | 
						|
    async def handle_async_request(
 | 
						|
        self, method, url, headers=None, stream=None, extensions=None
 | 
						|
    ):
 | 
						|
        retry = 2
 | 
						|
        while retry > 0:
 | 
						|
            retry -= 1
 | 
						|
            try:
 | 
						|
                return await super().handle_async_request(
 | 
						|
                    method, url, headers=headers, stream=stream, extensions=extensions
 | 
						|
                )
 | 
						|
            except (ProxyConnectionError, ProxyTimeoutError, ProxyError) as e:
 | 
						|
                raise httpx.ProxyError from e
 | 
						|
            except OSError as e:
 | 
						|
                # socket.gaierror when DNS resolution fails
 | 
						|
                raise httpx.ConnectError from e
 | 
						|
            except httpx.NetworkError as e:
 | 
						|
                # httpx.WriteError on HTTP/2 connection leaves a new opened stream
 | 
						|
                # then each new request creates a new stream and raise the same WriteError
 | 
						|
                await close_connections_for_url(self, url)
 | 
						|
                raise e
 | 
						|
            except anyio.ClosedResourceError as e:
 | 
						|
                await close_connections_for_url(self, url)
 | 
						|
                raise httpx.CloseError from e
 | 
						|
            except httpx.RemoteProtocolError as e:
 | 
						|
                # in case of httpx.RemoteProtocolError: Server disconnected
 | 
						|
                await close_connections_for_url(self, url)
 | 
						|
                logger.warning('httpx.RemoteProtocolError: retry', exc_info=e)
 | 
						|
                # retry
 | 
						|
 | 
						|
 | 
						|
class AsyncHTTPTransportFixed(httpx.AsyncHTTPTransport):
 | 
						|
    """Fix httpx.AsyncHTTPTransport"""
 | 
						|
 | 
						|
    async def handle_async_request(
 | 
						|
        self, method, url, headers=None, stream=None, extensions=None
 | 
						|
    ):
 | 
						|
        retry = 2
 | 
						|
        while retry > 0:
 | 
						|
            retry -= 1
 | 
						|
            try:
 | 
						|
                return await super().handle_async_request(
 | 
						|
                    method, url, headers=headers, stream=stream, extensions=extensions
 | 
						|
                )
 | 
						|
            except OSError as e:
 | 
						|
                # socket.gaierror when DNS resolution fails
 | 
						|
                raise httpx.ConnectError from e
 | 
						|
            except httpx.NetworkError as e:
 | 
						|
                # httpx.WriteError on HTTP/2 connection leaves a new opened stream
 | 
						|
                # then each new request creates a new stream and raise the same WriteError
 | 
						|
                await close_connections_for_url(self._pool, url)
 | 
						|
                raise e
 | 
						|
            except anyio.ClosedResourceError as e:
 | 
						|
                await close_connections_for_url(self._pool, url)
 | 
						|
                raise httpx.CloseError from e
 | 
						|
            except httpx.RemoteProtocolError as e:
 | 
						|
                # in case of httpx.RemoteProtocolError: Server disconnected
 | 
						|
                await close_connections_for_url(self._pool, url)
 | 
						|
                logger.warning('httpx.RemoteProtocolError: retry', exc_info=e)
 | 
						|
                # retry
 | 
						|
 | 
						|
 | 
						|
def get_transport_for_socks_proxy(verify, http2, local_address, proxy_url, limit, retries):
 | 
						|
    # support socks5h (requests compatibility):
 | 
						|
    # https://requests.readthedocs.io/en/master/user/advanced/#socks
 | 
						|
    # socks5://   hostname is resolved on client side
 | 
						|
    # socks5h://  hostname is resolved on proxy side
 | 
						|
    rdns = False
 | 
						|
    socks5h = 'socks5h://'
 | 
						|
    if proxy_url.startswith(socks5h):
 | 
						|
        proxy_url = 'socks5://' + proxy_url[len(socks5h):]
 | 
						|
        rdns = True
 | 
						|
 | 
						|
    proxy_type, proxy_host, proxy_port, proxy_username, proxy_password = parse_proxy_url(proxy_url)
 | 
						|
    verify = get_sslcontexts(proxy_url, None, True, False, http2) if verify is True else verify
 | 
						|
    return AsyncProxyTransportFixed(
 | 
						|
        proxy_type=proxy_type, proxy_host=proxy_host, proxy_port=proxy_port,
 | 
						|
        username=proxy_username, password=proxy_password,
 | 
						|
        rdns=rdns,
 | 
						|
        loop=get_loop(),
 | 
						|
        verify=verify,
 | 
						|
        http2=http2,
 | 
						|
        local_address=local_address,
 | 
						|
        max_connections=limit.max_connections,
 | 
						|
        max_keepalive_connections=limit.max_keepalive_connections,
 | 
						|
        keepalive_expiry=limit.keepalive_expiry,
 | 
						|
        retries=retries,
 | 
						|
        **TRANSPORT_KWARGS
 | 
						|
    )
 | 
						|
 | 
						|
 | 
						|
def get_transport(verify, http2, local_address, proxy_url, limit, retries):
 | 
						|
    verify = get_sslcontexts(None, None, True, False, http2) if verify is True else verify
 | 
						|
    return AsyncHTTPTransportFixed(
 | 
						|
        # pylint: disable=protected-access
 | 
						|
        verify=verify,
 | 
						|
        http2=http2,
 | 
						|
        local_address=local_address,
 | 
						|
        proxy=httpx._config.Proxy(proxy_url) if proxy_url else None,
 | 
						|
        limits=limit,
 | 
						|
        retries=retries,
 | 
						|
        **TRANSPORT_KWARGS
 | 
						|
    )
 | 
						|
 | 
						|
 | 
						|
def new_client(
 | 
						|
        # pylint: disable=too-many-arguments
 | 
						|
        enable_http, verify, enable_http2,
 | 
						|
        max_connections, max_keepalive_connections, keepalive_expiry,
 | 
						|
        proxies, local_address, retries, max_redirects, hook_log_response  ):
 | 
						|
    limit = httpx.Limits(
 | 
						|
        max_connections=max_connections,
 | 
						|
        max_keepalive_connections=max_keepalive_connections,
 | 
						|
        keepalive_expiry=keepalive_expiry
 | 
						|
    )
 | 
						|
    # See https://www.python-httpx.org/advanced/#routing
 | 
						|
    mounts = {}
 | 
						|
    for pattern, proxy_url in proxies.items():
 | 
						|
        if not enable_http and pattern.startswith('http://'):
 | 
						|
            continue
 | 
						|
        if (proxy_url.startswith('socks4://')
 | 
						|
           or proxy_url.startswith('socks5://')
 | 
						|
            or proxy_url.startswith('socks5h://')
 | 
						|
        ):
 | 
						|
            mounts[pattern] = get_transport_for_socks_proxy(
 | 
						|
                verify, enable_http2, local_address, proxy_url, limit, retries
 | 
						|
            )
 | 
						|
        else:
 | 
						|
            mounts[pattern] = get_transport(
 | 
						|
                verify, enable_http2, local_address, proxy_url, limit, retries
 | 
						|
            )
 | 
						|
 | 
						|
    if not enable_http:
 | 
						|
        mounts['http://'] = AsyncHTTPTransportNoHttp()
 | 
						|
 | 
						|
    transport = get_transport(verify, enable_http2, local_address, None, limit, retries)
 | 
						|
 | 
						|
    event_hooks = None
 | 
						|
    if hook_log_response:
 | 
						|
        event_hooks = {'response': [ hook_log_response ]}
 | 
						|
 | 
						|
    return httpx.AsyncClient(
 | 
						|
        transport=transport,
 | 
						|
        mounts=mounts,
 | 
						|
        max_redirects=max_redirects,
 | 
						|
        event_hooks=event_hooks,
 | 
						|
    )
 | 
						|
 | 
						|
 | 
						|
def get_loop():
 | 
						|
    return LOOP
 | 
						|
 | 
						|
 | 
						|
def init():
 | 
						|
    # log
 | 
						|
    for logger_name in ('hpack.hpack', 'hpack.table', 'httpx._client'):
 | 
						|
        logging.getLogger(logger_name).setLevel(logging.WARNING)
 | 
						|
 | 
						|
    # loop
 | 
						|
    def loop_thread():
 | 
						|
        global LOOP
 | 
						|
        LOOP = asyncio.new_event_loop()
 | 
						|
        LOOP.run_forever()
 | 
						|
 | 
						|
    thread = threading.Thread(
 | 
						|
        target=loop_thread,
 | 
						|
        name='asyncio_loop',
 | 
						|
        daemon=True,
 | 
						|
    )
 | 
						|
    thread.start()
 | 
						|
 | 
						|
 | 
						|
init()
 |