mirror of
				https://github.com/searxng/searxng.git
				synced 2025-11-03 19:17:07 -05:00 
			
		
		
		
	[mod] refactoring: processors
Report to the user suspended engines. searx.search.processor.abstract: * manages suspend time (per network). * reports suspended time to the ResultContainer (method extend_container_if_suspended) * adds the results to the ResultContainer (method extend_container) * handles exceptions (method handle_exception)
This commit is contained in:
		
							parent
							
								
									ae5954f2da
								
							
						
					
					
						commit
						aae7830d14
					
				@ -51,8 +51,6 @@ engine_default_args = {'paging': False,
 | 
			
		||||
                       'shortcut': '-',
 | 
			
		||||
                       'disabled': False,
 | 
			
		||||
                       'enable_http': False,
 | 
			
		||||
                       'suspend_end_time': 0,
 | 
			
		||||
                       'continuous_errors': 0,
 | 
			
		||||
                       'time_range_support': False,
 | 
			
		||||
                       'engine_type': 'online',
 | 
			
		||||
                       'display_error_messages': True,
 | 
			
		||||
 | 
			
		||||
@ -199,7 +199,7 @@ class Network:
 | 
			
		||||
 | 
			
		||||
def get_network(name=None):
 | 
			
		||||
    global NETWORKS
 | 
			
		||||
    return NETWORKS[name or DEFAULT_NAME]
 | 
			
		||||
    return NETWORKS.get(name or DEFAULT_NAME)
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
def initialize(settings_engines=None, settings_outgoing=None):
 | 
			
		||||
 | 
			
		||||
@ -369,9 +369,9 @@ class ResultContainer:
 | 
			
		||||
            return 0
 | 
			
		||||
        return resultnum_sum / len(self._number_of_results)
 | 
			
		||||
 | 
			
		||||
    def add_unresponsive_engine(self, engine_name, error_type, error_message=None):
 | 
			
		||||
    def add_unresponsive_engine(self, engine_name, error_type, error_message=None, suspended=False):
 | 
			
		||||
        if engines[engine_name].display_error_messages:
 | 
			
		||||
            self.unresponsive_engines.add((engine_name, error_type, error_message))
 | 
			
		||||
            self.unresponsive_engines.add((engine_name, error_type, error_message, suspended))
 | 
			
		||||
 | 
			
		||||
    def add_timing(self, engine_name, engine_time, page_load_time):
 | 
			
		||||
        self.timings.append({
 | 
			
		||||
 | 
			
		||||
@ -106,12 +106,16 @@ class Search:
 | 
			
		||||
        for engineref in self.search_query.engineref_list:
 | 
			
		||||
            processor = processors[engineref.name]
 | 
			
		||||
 | 
			
		||||
            # stop the request now if the engine is suspend
 | 
			
		||||
            if processor.extend_container_if_suspended(self.result_container):
 | 
			
		||||
                continue
 | 
			
		||||
 | 
			
		||||
            # set default request parameters
 | 
			
		||||
            request_params = processor.get_params(self.search_query, engineref.category)
 | 
			
		||||
            if request_params is None:
 | 
			
		||||
                continue
 | 
			
		||||
 | 
			
		||||
            with threading.RLock():
 | 
			
		||||
            with processor.lock:
 | 
			
		||||
                processor.engine.stats['sent_search_count'] += 1
 | 
			
		||||
 | 
			
		||||
            # append request to list
 | 
			
		||||
 | 
			
		||||
@ -4,7 +4,6 @@ import typing
 | 
			
		||||
import types
 | 
			
		||||
import functools
 | 
			
		||||
import itertools
 | 
			
		||||
import threading
 | 
			
		||||
from time import time
 | 
			
		||||
from urllib.parse import urlparse
 | 
			
		||||
 | 
			
		||||
@ -385,7 +384,7 @@ class Checker:
 | 
			
		||||
        engineref_category = search_query.engineref_list[0].category
 | 
			
		||||
        params = self.processor.get_params(search_query, engineref_category)
 | 
			
		||||
        if params is not None:
 | 
			
		||||
            with threading.RLock():
 | 
			
		||||
            with self.processor.lock:
 | 
			
		||||
                self.processor.engine.stats['sent_search_count'] += 1
 | 
			
		||||
            self.processor.search(search_query.query, params, result_container, time(), 5)
 | 
			
		||||
        return result_container
 | 
			
		||||
 | 
			
		||||
@ -1,17 +1,115 @@
 | 
			
		||||
# SPDX-License-Identifier: AGPL-3.0-or-later
 | 
			
		||||
 | 
			
		||||
import threading
 | 
			
		||||
from abc import abstractmethod, ABC
 | 
			
		||||
from time import time
 | 
			
		||||
 | 
			
		||||
from searx import logger
 | 
			
		||||
from searx.engines import settings
 | 
			
		||||
from searx.network import get_time_for_thread, get_network
 | 
			
		||||
from searx.metrology.error_recorder import record_exception, record_error
 | 
			
		||||
from searx.exceptions import SearxEngineAccessDeniedException
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
logger = logger.getChild('searx.search.processor')
 | 
			
		||||
SUSPENDED_STATUS = {}
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
class SuspendedStatus:
 | 
			
		||||
 | 
			
		||||
    __slots__ = 'suspend_end_time', 'suspend_reason', 'continuous_errors', 'lock'
 | 
			
		||||
 | 
			
		||||
    def __init__(self):
 | 
			
		||||
        self.lock = threading.Lock()
 | 
			
		||||
        self.continuous_errors = 0
 | 
			
		||||
        self.suspend_end_time = 0
 | 
			
		||||
        self.suspend_reason = None
 | 
			
		||||
 | 
			
		||||
    @property
 | 
			
		||||
    def is_suspended(self):
 | 
			
		||||
        return self.suspend_end_time >= time()
 | 
			
		||||
 | 
			
		||||
    def suspend(self, suspended_time, suspend_reason):
 | 
			
		||||
        with self.lock:
 | 
			
		||||
            # update continuous_errors / suspend_end_time
 | 
			
		||||
            self.continuous_errors += 1
 | 
			
		||||
            if suspended_time is None:
 | 
			
		||||
                suspended_time = min(settings['search']['max_ban_time_on_fail'],
 | 
			
		||||
                                     self.continuous_errors * settings['search']['ban_time_on_fail'])
 | 
			
		||||
            self.suspend_end_time = time() + suspended_time
 | 
			
		||||
            self.suspend_reason = suspend_reason
 | 
			
		||||
        logger.debug('Suspend engine for %i seconds', suspended_time)
 | 
			
		||||
 | 
			
		||||
    def resume(self):
 | 
			
		||||
        with self.lock:
 | 
			
		||||
            # reset the suspend variables
 | 
			
		||||
            self.continuous_errors = 0
 | 
			
		||||
            self.suspend_end_time = 0
 | 
			
		||||
            self.suspend_reason = None
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
class EngineProcessor(ABC):
 | 
			
		||||
 | 
			
		||||
    __slots__ = 'engine', 'engine_name', 'lock', 'suspended_status'
 | 
			
		||||
 | 
			
		||||
    def __init__(self, engine, engine_name):
 | 
			
		||||
        self.engine = engine
 | 
			
		||||
        self.engine_name = engine_name
 | 
			
		||||
        self.lock = threading.Lock()
 | 
			
		||||
        key = get_network(self.engine_name)
 | 
			
		||||
        key = id(key) if key else self.engine_name
 | 
			
		||||
        self.suspended_status = SUSPENDED_STATUS.setdefault(key, SuspendedStatus())
 | 
			
		||||
 | 
			
		||||
    def handle_exception(self, result_container, reason, exception, suspend=False, display_exception=True):
 | 
			
		||||
        # update result_container
 | 
			
		||||
        error_message = str(exception) if display_exception and exception else None
 | 
			
		||||
        result_container.add_unresponsive_engine(self.engine_name, reason, error_message)
 | 
			
		||||
        # metrics
 | 
			
		||||
        with self.lock:
 | 
			
		||||
            self.engine.stats['errors'] += 1
 | 
			
		||||
        if exception:
 | 
			
		||||
            record_exception(self.engine_name, exception)
 | 
			
		||||
        else:
 | 
			
		||||
            record_error(self.engine_name, reason)
 | 
			
		||||
        # suspend the engine ?
 | 
			
		||||
        if suspend:
 | 
			
		||||
            suspended_time = None
 | 
			
		||||
            if isinstance(exception, SearxEngineAccessDeniedException):
 | 
			
		||||
                suspended_time = exception.suspended_time
 | 
			
		||||
            self.suspended_status.suspend(suspended_time, reason)  # pylint: disable=no-member
 | 
			
		||||
 | 
			
		||||
    def _extend_container_basic(self, result_container, start_time, search_results):
 | 
			
		||||
        # update result_container
 | 
			
		||||
        result_container.extend(self.engine_name, search_results)
 | 
			
		||||
        engine_time = time() - start_time
 | 
			
		||||
        page_load_time = get_time_for_thread()
 | 
			
		||||
        result_container.add_timing(self.engine_name, engine_time, page_load_time)
 | 
			
		||||
        # metrics
 | 
			
		||||
        with self.lock:
 | 
			
		||||
            self.engine.stats['engine_time'] += engine_time
 | 
			
		||||
            self.engine.stats['engine_time_count'] += 1
 | 
			
		||||
            # update stats with the total HTTP time
 | 
			
		||||
            if page_load_time is not None and 'page_load_time' in self.engine.stats:
 | 
			
		||||
                self.engine.stats['page_load_time'] += page_load_time
 | 
			
		||||
                self.engine.stats['page_load_count'] += 1
 | 
			
		||||
 | 
			
		||||
    def extend_container(self, result_container, start_time, search_results):
 | 
			
		||||
        if getattr(threading.current_thread(), '_timeout', False):
 | 
			
		||||
            # the main thread is not waiting anymore
 | 
			
		||||
            self.handle_exception(result_container, 'Timeout', None)
 | 
			
		||||
        else:
 | 
			
		||||
            # check if the engine accepted the request
 | 
			
		||||
            if search_results is not None:
 | 
			
		||||
                self._extend_container_basic(result_container, start_time, search_results)
 | 
			
		||||
            self.suspended_status.resume()
 | 
			
		||||
 | 
			
		||||
    def extend_container_if_suspended(self, result_container):
 | 
			
		||||
        if self.suspended_status.is_suspended:
 | 
			
		||||
            result_container.add_unresponsive_engine(self.engine_name,
 | 
			
		||||
                                                     self.suspended_status.suspend_reason,
 | 
			
		||||
                                                     suspended=True)
 | 
			
		||||
            return True
 | 
			
		||||
        return False
 | 
			
		||||
 | 
			
		||||
    def get_params(self, search_query, engine_category):
 | 
			
		||||
        # if paging is not supported, skip
 | 
			
		||||
 | 
			
		||||
@ -1,51 +1,26 @@
 | 
			
		||||
# SPDX-License-Identifier: AGPL-3.0-or-later
 | 
			
		||||
 | 
			
		||||
import threading
 | 
			
		||||
from time import time
 | 
			
		||||
from searx import logger
 | 
			
		||||
from searx.metrology.error_recorder import record_exception, record_error
 | 
			
		||||
from searx.search.processors.abstract import EngineProcessor
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
logger = logger.getChild('search.processor.offline')
 | 
			
		||||
logger = logger.getChild('searx.search.processor.offline')
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
class OfflineProcessor(EngineProcessor):
 | 
			
		||||
 | 
			
		||||
    engine_type = 'offline'
 | 
			
		||||
 | 
			
		||||
    def _record_stats_on_error(self, result_container, start_time):
 | 
			
		||||
        engine_time = time() - start_time
 | 
			
		||||
        result_container.add_timing(self.engine_name, engine_time, engine_time)
 | 
			
		||||
 | 
			
		||||
        with threading.RLock():
 | 
			
		||||
            self.engine.stats['errors'] += 1
 | 
			
		||||
 | 
			
		||||
    def _search_basic(self, query, params):
 | 
			
		||||
        return self.engine.search(query, params)
 | 
			
		||||
 | 
			
		||||
    def search(self, query, params, result_container, start_time, timeout_limit):
 | 
			
		||||
        try:
 | 
			
		||||
            search_results = self._search_basic(query, params)
 | 
			
		||||
 | 
			
		||||
            if search_results:
 | 
			
		||||
                result_container.extend(self.engine_name, search_results)
 | 
			
		||||
 | 
			
		||||
                engine_time = time() - start_time
 | 
			
		||||
                result_container.add_timing(self.engine_name, engine_time, engine_time)
 | 
			
		||||
                with threading.RLock():
 | 
			
		||||
                    self.engine.stats['engine_time'] += engine_time
 | 
			
		||||
                    self.engine.stats['engine_time_count'] += 1
 | 
			
		||||
 | 
			
		||||
            self.extend_container(result_container, start_time, search_results)
 | 
			
		||||
        except ValueError as e:
 | 
			
		||||
            record_exception(self.engine_name, e)
 | 
			
		||||
            self._record_stats_on_error(result_container, start_time)
 | 
			
		||||
            # do not record the error
 | 
			
		||||
            logger.exception('engine {0} : invalid input : {1}'.format(self.engine_name, e))
 | 
			
		||||
        except Exception as e:
 | 
			
		||||
            record_exception(self.engine_name, e)
 | 
			
		||||
            self._record_stats_on_error(result_container, start_time)
 | 
			
		||||
            result_container.add_unresponsive_engine(self.engine_name, 'unexpected crash', str(e))
 | 
			
		||||
            self.handle_exception(result_container, 'unexpected crash', e)
 | 
			
		||||
            logger.exception('engine {0} : exception : {1}'.format(self.engine_name, e))
 | 
			
		||||
        else:
 | 
			
		||||
            if getattr(threading.current_thread(), '_timeout', False):
 | 
			
		||||
                record_error(self.engine_name, 'Timeout')
 | 
			
		||||
 | 
			
		||||
@ -1,23 +1,21 @@
 | 
			
		||||
# SPDX-License-Identifier: AGPL-3.0-or-later
 | 
			
		||||
 | 
			
		||||
from time import time
 | 
			
		||||
import threading
 | 
			
		||||
import asyncio
 | 
			
		||||
 | 
			
		||||
import httpx
 | 
			
		||||
 | 
			
		||||
import searx.network
 | 
			
		||||
from searx.engines import settings
 | 
			
		||||
from searx import logger
 | 
			
		||||
from searx.utils import gen_useragent
 | 
			
		||||
from searx.exceptions import (SearxEngineAccessDeniedException, SearxEngineCaptchaException,
 | 
			
		||||
                              SearxEngineTooManyRequestsException,)
 | 
			
		||||
from searx.metrology.error_recorder import record_exception, record_error
 | 
			
		||||
from searx.metrology.error_recorder import record_error
 | 
			
		||||
 | 
			
		||||
from searx.search.processors.abstract import EngineProcessor
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
logger = logger.getChild('search.processor.online')
 | 
			
		||||
logger = logger.getChild('searx.search.processor.online')
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
def default_request_params():
 | 
			
		||||
@ -41,11 +39,6 @@ class OnlineProcessor(EngineProcessor):
 | 
			
		||||
        if params is None:
 | 
			
		||||
            return None
 | 
			
		||||
 | 
			
		||||
        # skip suspended engines
 | 
			
		||||
        if self.engine.suspend_end_time >= time():
 | 
			
		||||
            logger.debug('Engine currently suspended: %s', self.engine_name)
 | 
			
		||||
            return None
 | 
			
		||||
 | 
			
		||||
        # add default params
 | 
			
		||||
        params.update(default_request_params())
 | 
			
		||||
 | 
			
		||||
@ -130,89 +123,38 @@ class OnlineProcessor(EngineProcessor):
 | 
			
		||||
        # set the network
 | 
			
		||||
        searx.network.set_context_network_name(self.engine_name)
 | 
			
		||||
 | 
			
		||||
        # suppose everything will be alright
 | 
			
		||||
        http_exception = False
 | 
			
		||||
        suspended_time = None
 | 
			
		||||
 | 
			
		||||
        try:
 | 
			
		||||
            # send requests and parse the results
 | 
			
		||||
            search_results = self._search_basic(query, params)
 | 
			
		||||
 | 
			
		||||
            # check if the engine accepted the request
 | 
			
		||||
            if search_results is not None:
 | 
			
		||||
                # yes, so add results
 | 
			
		||||
                result_container.extend(self.engine_name, search_results)
 | 
			
		||||
 | 
			
		||||
                # update engine time when there is no exception
 | 
			
		||||
                engine_time = time() - start_time
 | 
			
		||||
                page_load_time = searx.network.get_time_for_thread()
 | 
			
		||||
                result_container.add_timing(self.engine_name, engine_time, page_load_time)
 | 
			
		||||
                with threading.RLock():
 | 
			
		||||
                    self.engine.stats['engine_time'] += engine_time
 | 
			
		||||
                    self.engine.stats['engine_time_count'] += 1
 | 
			
		||||
                    # update stats with the total HTTP time
 | 
			
		||||
                    self.engine.stats['page_load_time'] += page_load_time
 | 
			
		||||
                    self.engine.stats['page_load_count'] += 1
 | 
			
		||||
        except Exception as e:
 | 
			
		||||
            record_exception(self.engine_name, e)
 | 
			
		||||
 | 
			
		||||
            # Timing
 | 
			
		||||
            engine_time = time() - start_time
 | 
			
		||||
            page_load_time = searx.network.get_time_for_thread()
 | 
			
		||||
            result_container.add_timing(self.engine_name, engine_time, page_load_time)
 | 
			
		||||
 | 
			
		||||
            # Record the errors
 | 
			
		||||
            with threading.RLock():
 | 
			
		||||
                self.engine.stats['errors'] += 1
 | 
			
		||||
 | 
			
		||||
            if (issubclass(e.__class__, (httpx.TimeoutException, asyncio.TimeoutError))):
 | 
			
		||||
                result_container.add_unresponsive_engine(self.engine_name, 'HTTP timeout')
 | 
			
		||||
            self.extend_container(result_container, start_time, search_results)
 | 
			
		||||
        except (httpx.TimeoutException, asyncio.TimeoutError) as e:
 | 
			
		||||
            # requests timeout (connect or read)
 | 
			
		||||
            self.handle_exception(result_container, 'HTTP timeout', e, suspend=True, display_exception=False)
 | 
			
		||||
            logger.error("engine {0} : HTTP requests timeout"
 | 
			
		||||
                         "(search duration : {1} s, timeout: {2} s) : {3}"
 | 
			
		||||
                             .format(self.engine_name, engine_time, timeout_limit, e.__class__.__name__))
 | 
			
		||||
                http_exception = True
 | 
			
		||||
            elif (issubclass(e.__class__, (httpx.HTTPError, httpx.StreamError))):
 | 
			
		||||
                result_container.add_unresponsive_engine(self.engine_name, 'HTTP error')
 | 
			
		||||
                         .format(self.engine_name, time() - start_time,
 | 
			
		||||
                                 timeout_limit,
 | 
			
		||||
                                 e.__class__.__name__))
 | 
			
		||||
        except (httpx.HTTPError, httpx.StreamError) as e:
 | 
			
		||||
            # other requests exception
 | 
			
		||||
            self.handle_exception(result_container, 'HTTP error', e, suspend=True, display_exception=False)
 | 
			
		||||
            logger.exception("engine {0} : requests exception"
 | 
			
		||||
                             "(search duration : {1} s, timeout: {2} s) : {3}"
 | 
			
		||||
                                 .format(self.engine_name, engine_time, timeout_limit, e))
 | 
			
		||||
                http_exception = True
 | 
			
		||||
            elif (issubclass(e.__class__, SearxEngineCaptchaException)):
 | 
			
		||||
                result_container.add_unresponsive_engine(self.engine_name, 'CAPTCHA required')
 | 
			
		||||
                             .format(self.engine_name, time() - start_time,
 | 
			
		||||
                                     timeout_limit,
 | 
			
		||||
                                     e))
 | 
			
		||||
        except SearxEngineCaptchaException as e:
 | 
			
		||||
            self.handle_exception(result_container, 'CAPTCHA required', e, suspend=True, display_exception=False)
 | 
			
		||||
            logger.exception('engine {0} : CAPTCHA'.format(self.engine_name))
 | 
			
		||||
                suspended_time = e.suspended_time  # pylint: disable=no-member
 | 
			
		||||
            elif (issubclass(e.__class__, SearxEngineTooManyRequestsException)):
 | 
			
		||||
                result_container.add_unresponsive_engine(self.engine_name, 'too many requests')
 | 
			
		||||
        except SearxEngineTooManyRequestsException as e:
 | 
			
		||||
            self.handle_exception(result_container, 'too many requests', e, suspend=True, display_exception=False)
 | 
			
		||||
            logger.exception('engine {0} : Too many requests'.format(self.engine_name))
 | 
			
		||||
                suspended_time = e.suspended_time  # pylint: disable=no-member
 | 
			
		||||
            elif (issubclass(e.__class__, SearxEngineAccessDeniedException)):
 | 
			
		||||
                result_container.add_unresponsive_engine(self.engine_name, 'blocked')
 | 
			
		||||
        except SearxEngineAccessDeniedException as e:
 | 
			
		||||
            self.handle_exception(result_container, 'blocked', e, suspend=True, display_exception=False)
 | 
			
		||||
            logger.exception('engine {0} : Searx is blocked'.format(self.engine_name))
 | 
			
		||||
                suspended_time = e.suspended_time  # pylint: disable=no-member
 | 
			
		||||
            else:
 | 
			
		||||
                result_container.add_unresponsive_engine(self.engine_name, 'unexpected crash')
 | 
			
		||||
                # others errors
 | 
			
		||||
        except Exception as e:
 | 
			
		||||
            self.handle_exception(result_container, 'unexpected crash', e, display_exception=False)
 | 
			
		||||
            logger.exception('engine {0} : exception : {1}'.format(self.engine_name, e))
 | 
			
		||||
        else:
 | 
			
		||||
            if getattr(threading.current_thread(), '_timeout', False):
 | 
			
		||||
                record_error(self.engine_name, 'Timeout')
 | 
			
		||||
 | 
			
		||||
        # suspend the engine if there is an HTTP error
 | 
			
		||||
        # or suspended_time is defined
 | 
			
		||||
        with threading.RLock():
 | 
			
		||||
            if http_exception or suspended_time:
 | 
			
		||||
                # update continuous_errors / suspend_end_time
 | 
			
		||||
                self.engine.continuous_errors += 1
 | 
			
		||||
                if suspended_time is None:
 | 
			
		||||
                    suspended_time = min(settings['search']['max_ban_time_on_fail'],
 | 
			
		||||
                                         self.engine.continuous_errors * settings['search']['ban_time_on_fail'])
 | 
			
		||||
                self.engine.suspend_end_time = time() + suspended_time
 | 
			
		||||
            else:
 | 
			
		||||
                # reset the suspend variables
 | 
			
		||||
                self.engine.continuous_errors = 0
 | 
			
		||||
                self.engine.suspend_end_time = 0
 | 
			
		||||
 | 
			
		||||
    def get_default_tests(self):
 | 
			
		||||
        tests = {}
 | 
			
		||||
 | 
			
		||||
@ -764,6 +764,8 @@ def __get_translated_errors(unresponsive_engines):
 | 
			
		||||
        error_msg = gettext(unresponsive_engine[1])
 | 
			
		||||
        if unresponsive_engine[2]:
 | 
			
		||||
            error_msg = "{} {}".format(error_msg, unresponsive_engine[2])
 | 
			
		||||
        if unresponsive_engine[3]:
 | 
			
		||||
            error_msg = gettext('Suspended') + ': ' + error_msg
 | 
			
		||||
        translated_errors.add((unresponsive_engine[0], error_msg))
 | 
			
		||||
    return translated_errors
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
		Loading…
	
	
			
			x
			
			
		
	
		Reference in New Issue
	
	Block a user