diff options
| author | Adam Tauber <asciimoo@gmail.com> | 2016-11-04 19:38:21 +0100 |
|---|---|---|
| committer | GitHub <noreply@github.com> | 2016-11-04 19:38:21 +0100 |
| commit | 51eafdd471fe989accc060f3ebd8efba0d9b6a71 (patch) | |
| tree | c39ce3ba1bdd8d7f02543b11f5f20ea1ade2d0f5 /searx/plugins | |
| parent | 142cd870950e25fe4bb8f4b995b861162af9777c (diff) | |
| parent | 58a6c045c81a84abfb8a29368e7d5453efc5f2fa (diff) | |
Merge pull request #733 from dalf/searchpy
Search architecture
Diffstat (limited to 'searx/plugins')
| -rw-r--r-- | searx/plugins/doai_rewrite.py | 8 | ||||
| -rw-r--r-- | searx/plugins/https_rewrite.py | 3 | ||||
| -rw-r--r-- | searx/plugins/self_info.py | 16 | ||||
| -rw-r--r-- | searx/plugins/tracker_url_remover.py | 10 |
4 files changed, 18 insertions, 19 deletions
diff --git a/searx/plugins/doai_rewrite.py b/searx/plugins/doai_rewrite.py index fc5998b14..0142af672 100644 --- a/searx/plugins/doai_rewrite.py +++ b/searx/plugins/doai_rewrite.py @@ -20,12 +20,12 @@ def extract_doi(url): return None -def on_result(request, ctx): - doi = extract_doi(ctx['result']['parsed_url']) +def on_result(request, search, result): + doi = extract_doi(result['parsed_url']) if doi and len(doi) < 50: for suffix in ('/', '.pdf', '/full', '/meta', '/abstract'): if doi.endswith(suffix): doi = doi[:-len(suffix)] - ctx['result']['url'] = 'http://doai.io/' + doi - ctx['result']['parsed_url'] = urlparse(ctx['result']['url']) + result['url'] = 'http://doai.io/' + doi + result['parsed_url'] = urlparse(ctx['result']['url']) return True diff --git a/searx/plugins/https_rewrite.py b/searx/plugins/https_rewrite.py index 8a9fcd4ad..8b4c9784e 100644 --- a/searx/plugins/https_rewrite.py +++ b/searx/plugins/https_rewrite.py @@ -220,8 +220,7 @@ def https_url_rewrite(result): return result -def on_result(request, ctx): - result = ctx['result'] +def on_result(request, search, result): if result['parsed_url'].scheme == 'http': https_url_rewrite(result) return True diff --git a/searx/plugins/self_info.py b/searx/plugins/self_info.py index 438274c41..a2aeda98e 100644 --- a/searx/plugins/self_info.py +++ b/searx/plugins/self_info.py @@ -28,19 +28,19 @@ p = re.compile('.*user[ -]agent.*', re.IGNORECASE) # attach callback to the post search hook # request: flask request object # ctx: the whole local context of the pre search hook -def post_search(request, ctx): - if ctx['search'].pageno > 1: +def post_search(request, search): + if search.search_query.pageno > 1: return True - if ctx['search'].query == 'ip': + if search.search_query.query == 'ip': x_forwarded_for = request.headers.getlist("X-Forwarded-For") if x_forwarded_for: ip = x_forwarded_for[0] else: ip = request.remote_addr - ctx['search'].result_container.answers.clear() - ctx['search'].result_container.answers.add(ip) - elif p.match(ctx['search'].query): + search.result_container.answers.clear() + search.result_container.answers.add(ip) + elif p.match(search.search_query.query): ua = request.user_agent - ctx['search'].result_container.answers.clear() - ctx['search'].result_container.answers.add(ua) + search.result_container.answers.clear() + search.result_container.answers.add(ua) return True diff --git a/searx/plugins/tracker_url_remover.py b/searx/plugins/tracker_url_remover.py index b909e3fae..68a004e33 100644 --- a/searx/plugins/tracker_url_remover.py +++ b/searx/plugins/tracker_url_remover.py @@ -28,8 +28,8 @@ description = gettext('Remove trackers arguments from the returned URL') default_on = True -def on_result(request, ctx): - query = ctx['result']['parsed_url'].query +def on_result(request, search, result): + query = result['parsed_url'].query if query == "": return True @@ -37,8 +37,8 @@ def on_result(request, ctx): for reg in regexes: query = reg.sub('', query) - if query != ctx['result']['parsed_url'].query: - ctx['result']['parsed_url'] = ctx['result']['parsed_url']._replace(query=query) - ctx['result']['url'] = urlunparse(ctx['result']['parsed_url']) + if query != result['parsed_url'].query: + result['parsed_url'] = result['parsed_url']._replace(query=query) + result['url'] = urlunparse(result['parsed_url']) return True |