summaryrefslogtreecommitdiff
path: root/searx/plugins
diff options
context:
space:
mode:
Diffstat (limited to 'searx/plugins')
-rw-r--r--searx/plugins/__init__.py2
-rw-r--r--searx/plugins/ahmia_filter.py8
-rw-r--r--searx/plugins/hostname_replace.py32
-rw-r--r--searx/plugins/oa_doi_rewrite.py15
4 files changed, 41 insertions, 16 deletions
diff --git a/searx/plugins/__init__.py b/searx/plugins/__init__.py
index 3a35f7025..1153c9ed1 100644
--- a/searx/plugins/__init__.py
+++ b/searx/plugins/__init__.py
@@ -31,6 +31,7 @@ from searx.plugins import (oa_doi_rewrite,
hash_plugin,
infinite_scroll,
self_info,
+ hostname_replace,
search_on_category_select,
tracker_url_remover,
vim_hotkeys)
@@ -182,6 +183,7 @@ plugins.register(oa_doi_rewrite)
plugins.register(hash_plugin)
plugins.register(infinite_scroll)
plugins.register(self_info)
+plugins.register(hostname_replace)
plugins.register(search_on_category_select)
plugins.register(tracker_url_remover)
plugins.register(vim_hotkeys)
diff --git a/searx/plugins/ahmia_filter.py b/searx/plugins/ahmia_filter.py
index 83b05e4d2..70f216ee1 100644
--- a/searx/plugins/ahmia_filter.py
+++ b/searx/plugins/ahmia_filter.py
@@ -20,14 +20,8 @@ def get_ahmia_blacklist():
return ahmia_blacklist
-def not_blacklisted(result):
+def on_result(request, search, result):
if not result.get('is_onion') or not result.get('parsed_url'):
return True
result_hash = md5(result['parsed_url'].hostname.encode()).hexdigest()
return result_hash not in get_ahmia_blacklist()
-
-
-def post_search(request, search):
- filtered_results = list(filter(not_blacklisted, search.result_container._merged_results))
- search.result_container._merged_results = filtered_results
- return True
diff --git a/searx/plugins/hostname_replace.py b/searx/plugins/hostname_replace.py
new file mode 100644
index 000000000..778b84615
--- /dev/null
+++ b/searx/plugins/hostname_replace.py
@@ -0,0 +1,32 @@
+# SPDX-License-Identifier: AGPL-3.0-or-later
+
+import re
+from urllib.parse import urlunparse
+from searx import settings
+from searx.plugins import logger
+from flask_babel import gettext
+
+name = gettext('Hostname replace')
+description = gettext('Rewrite result hostnames or remove results based on the hostname')
+default_on = False
+preference_section = 'general'
+
+plugin_id = 'hostname_replace'
+
+replacements = {re.compile(p): r for (p, r) in settings[plugin_id].items()} if plugin_id in settings else {}
+
+logger = logger.getChild(plugin_id)
+parsed = 'parsed_url'
+
+
+def on_result(request, search, result):
+ if parsed not in result:
+ return True
+ for (pattern, replacement) in replacements.items():
+ if pattern.search(result[parsed].netloc):
+ if not replacement:
+ return False
+ result[parsed] = result[parsed]._replace(netloc=pattern.sub(replacement, result[parsed].netloc))
+ result['url'] = urlunparse(result[parsed])
+
+ return True
diff --git a/searx/plugins/oa_doi_rewrite.py b/searx/plugins/oa_doi_rewrite.py
index 02a712942..2dcc01e05 100644
--- a/searx/plugins/oa_doi_rewrite.py
+++ b/searx/plugins/oa_doi_rewrite.py
@@ -11,8 +11,6 @@ description = gettext('Avoid paywalls by redirecting to open-access versions of
default_on = False
preference_section = 'general'
-doi_resolvers = settings['doi_resolvers']
-
def extract_doi(url):
match = regex.search(url.path)
@@ -25,13 +23,12 @@ def extract_doi(url):
return None
-def get_doi_resolver(args, preference_doi_resolver):
+def get_doi_resolver(preferences):
doi_resolvers = settings['doi_resolvers']
- doi_resolver = args.get('doi_resolver', preference_doi_resolver)[0]
- if doi_resolver not in doi_resolvers:
- doi_resolver = settings['default_doi_resolver']
- doi_resolver_url = doi_resolvers[doi_resolver]
- return doi_resolver_url
+ selected_resolver = preferences.get_value('doi_resolver')[0]
+ if selected_resolver not in doi_resolvers:
+ selected_resolver = settings['default_doi_resolver']
+ return doi_resolvers[selected_resolver]
def on_result(request, search, result):
@@ -43,6 +40,6 @@ def on_result(request, search, result):
for suffix in ('/', '.pdf', '.xml', '/full', '/meta', '/abstract'):
if doi.endswith(suffix):
doi = doi[:-len(suffix)]
- result['url'] = get_doi_resolver(request.args, request.preferences.get_value('doi_resolver')) + doi
+ result['url'] = get_doi_resolver(request.preferences) + doi
result['parsed_url'] = urlparse(result['url'])
return True