summaryrefslogtreecommitdiff
path: root/searx/search/processors/abstract.py
diff options
context:
space:
mode:
authorAlexandre Flament <alex@al-f.net>2021-04-13 15:21:53 +0200
committerAlexandre Flament <alex@al-f.net>2021-04-21 16:24:46 +0200
commitaae7830d14242ac1f98232f428654c5d2c9c5eb2 (patch)
tree83df1950b7b5889fb17eda740a095022a816642e /searx/search/processors/abstract.py
parentae5954f2dad0386868709a59a1b29d005ebb2b2d (diff)
[mod] refactoring: processors
Report to the user suspended engines. searx.search.processor.abstract: * manages suspend time (per network). * reports suspended time to the ResultContainer (method extend_container_if_suspended) * adds the results to the ResultContainer (method extend_container) * handles exceptions (method handle_exception)
Diffstat (limited to 'searx/search/processors/abstract.py')
-rw-r--r--searx/search/processors/abstract.py98
1 files changed, 98 insertions, 0 deletions
diff --git a/searx/search/processors/abstract.py b/searx/search/processors/abstract.py
index 26dab069f..e32d8f067 100644
--- a/searx/search/processors/abstract.py
+++ b/searx/search/processors/abstract.py
@@ -1,17 +1,115 @@
# SPDX-License-Identifier: AGPL-3.0-or-later
+import threading
from abc import abstractmethod, ABC
+from time import time
+
from searx import logger
+from searx.engines import settings
+from searx.network import get_time_for_thread, get_network
+from searx.metrology.error_recorder import record_exception, record_error
+from searx.exceptions import SearxEngineAccessDeniedException
logger = logger.getChild('searx.search.processor')
+SUSPENDED_STATUS = {}
+
+
+class SuspendedStatus:
+
+ __slots__ = 'suspend_end_time', 'suspend_reason', 'continuous_errors', 'lock'
+
+ def __init__(self):
+ self.lock = threading.Lock()
+ self.continuous_errors = 0
+ self.suspend_end_time = 0
+ self.suspend_reason = None
+
+ @property
+ def is_suspended(self):
+ return self.suspend_end_time >= time()
+
+ def suspend(self, suspended_time, suspend_reason):
+ with self.lock:
+ # update continuous_errors / suspend_end_time
+ self.continuous_errors += 1
+ if suspended_time is None:
+ suspended_time = min(settings['search']['max_ban_time_on_fail'],
+ self.continuous_errors * settings['search']['ban_time_on_fail'])
+ self.suspend_end_time = time() + suspended_time
+ self.suspend_reason = suspend_reason
+ logger.debug('Suspend engine for %i seconds', suspended_time)
+
+ def resume(self):
+ with self.lock:
+ # reset the suspend variables
+ self.continuous_errors = 0
+ self.suspend_end_time = 0
+ self.suspend_reason = None
class EngineProcessor(ABC):
+ __slots__ = 'engine', 'engine_name', 'lock', 'suspended_status'
+
def __init__(self, engine, engine_name):
self.engine = engine
self.engine_name = engine_name
+ self.lock = threading.Lock()
+ key = get_network(self.engine_name)
+ key = id(key) if key else self.engine_name
+ self.suspended_status = SUSPENDED_STATUS.setdefault(key, SuspendedStatus())
+
+ def handle_exception(self, result_container, reason, exception, suspend=False, display_exception=True):
+ # update result_container
+ error_message = str(exception) if display_exception and exception else None
+ result_container.add_unresponsive_engine(self.engine_name, reason, error_message)
+ # metrics
+ with self.lock:
+ self.engine.stats['errors'] += 1
+ if exception:
+ record_exception(self.engine_name, exception)
+ else:
+ record_error(self.engine_name, reason)
+ # suspend the engine ?
+ if suspend:
+ suspended_time = None
+ if isinstance(exception, SearxEngineAccessDeniedException):
+ suspended_time = exception.suspended_time
+ self.suspended_status.suspend(suspended_time, reason) # pylint: disable=no-member
+
+ def _extend_container_basic(self, result_container, start_time, search_results):
+ # update result_container
+ result_container.extend(self.engine_name, search_results)
+ engine_time = time() - start_time
+ page_load_time = get_time_for_thread()
+ result_container.add_timing(self.engine_name, engine_time, page_load_time)
+ # metrics
+ with self.lock:
+ self.engine.stats['engine_time'] += engine_time
+ self.engine.stats['engine_time_count'] += 1
+ # update stats with the total HTTP time
+ if page_load_time is not None and 'page_load_time' in self.engine.stats:
+ self.engine.stats['page_load_time'] += page_load_time
+ self.engine.stats['page_load_count'] += 1
+
+ def extend_container(self, result_container, start_time, search_results):
+ if getattr(threading.current_thread(), '_timeout', False):
+ # the main thread is not waiting anymore
+ self.handle_exception(result_container, 'Timeout', None)
+ else:
+ # check if the engine accepted the request
+ if search_results is not None:
+ self._extend_container_basic(result_container, start_time, search_results)
+ self.suspended_status.resume()
+
+ def extend_container_if_suspended(self, result_container):
+ if self.suspended_status.is_suspended:
+ result_container.add_unresponsive_engine(self.engine_name,
+ self.suspended_status.suspend_reason,
+ suspended=True)
+ return True
+ return False
def get_params(self, search_query, engine_category):
# if paging is not supported, skip