From cf09b500f35fd1bca3fc9cc853bd7ea932220e4e Mon Sep 17 00:00:00 2001 From: Guilhem Bonnefille Date: Sun, 3 Apr 2016 22:03:41 +0200 Subject: Add support for dokuwiki engine --- searx/engines/doku.py | 83 +++++++++++++++++++++++++++++++++++++++++++++++++++ 1 file changed, 83 insertions(+) create mode 100644 searx/engines/doku.py (limited to 'searx/engines') diff --git a/searx/engines/doku.py b/searx/engines/doku.py new file mode 100644 index 000000000..18abe75e5 --- /dev/null +++ b/searx/engines/doku.py @@ -0,0 +1,83 @@ +# Doku Wiki +# +# @website https://www.dokuwiki.org/ +# @provide-api yes +# (https://www.dokuwiki.org/devel:xmlrpc) +# +# @using-api no +# @results HTML +# @stable yes +# @parse (general) url, title, content + +from urllib import urlencode +from lxml.html import fromstring +from searx.engines.xpath import extract_text + +# engine dependent config +categories = ['general'] # TODO , 'images', 'music', 'videos', 'files' +paging = False +language_support = False +number_of_results = 5 + +# search-url +# Doku is OpenSearch compatible +base_url = 'http://localhost:8090' +search_url = '/?do=search'\ + '&id={query}' +# TODO '&startRecord={offset}'\ +# TODO '&maximumRecords={limit}'\ + +# do search-request +def request(query, params): + + params['url'] = base_url +\ + search_url.format(query=urlencode({'query': query})) + + return params + + +# get response from search-request +def response(resp): + results = [] + + doc = fromstring(resp.text) + + # parse results + # Quickhits + for r in doc.xpath('//div[@class="search_quickresult"]/ul/li'): + try: + res_url = r.xpath('.//a[@class="wikilink1"]/@href')[-1] + except: + continue + + if not res_url: + continue + + title = extract_text(r.xpath('.//a[@class="wikilink1"]/@title')) + + # append result + results.append({'title': title, + 'content': "", + 'url': base_url + res_url}) + + # Search results + for r in doc.xpath('//dl[@class="search_results"]/*'): + try: + if r.tag == "dt": + res_url = r.xpath('.//a[@class="wikilink1"]/@href')[-1] + title = extract_text(r.xpath('.//a[@class="wikilink1"]/@title')) + elif r.tag == "dd": + content = extract_text(r.xpath('.')) + + # append result + results.append({'title': title, + 'content': content, + 'url': base_url + res_url}) + except: + continue + + if not res_url: + continue + + # return results + return results -- cgit v1.2.3 From f26f0dab2e4e6a6f77ea9f04f36fe2eb2d6893df Mon Sep 17 00:00:00 2001 From: Guilhem Bonnefille Date: Mon, 4 Apr 2016 13:38:22 +0200 Subject: Fix pep8 E302 Cf. http://legacy.python.org/dev/peps/pep-0008/#blank-lines --- searx/engines/doku.py | 1 + 1 file changed, 1 insertion(+) (limited to 'searx/engines') diff --git a/searx/engines/doku.py b/searx/engines/doku.py index 18abe75e5..233fd2233 100644 --- a/searx/engines/doku.py +++ b/searx/engines/doku.py @@ -27,6 +27,7 @@ search_url = '/?do=search'\ # TODO '&startRecord={offset}'\ # TODO '&maximumRecords={limit}'\ + # do search-request def request(query, params): -- cgit v1.2.3 From 2733a92383f7f8127cdf4871c8091b0489ba7356 Mon Sep 17 00:00:00 2001 From: Guilhem Bonnefille Date: Tue, 5 Apr 2016 13:31:49 +0200 Subject: Fix query encoding --- searx/engines/doku.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) (limited to 'searx/engines') diff --git a/searx/engines/doku.py b/searx/engines/doku.py index 233fd2233..93867fd0d 100644 --- a/searx/engines/doku.py +++ b/searx/engines/doku.py @@ -23,7 +23,7 @@ number_of_results = 5 # Doku is OpenSearch compatible base_url = 'http://localhost:8090' search_url = '/?do=search'\ - '&id={query}' + '&{query}' # TODO '&startRecord={offset}'\ # TODO '&maximumRecords={limit}'\ @@ -32,7 +32,7 @@ search_url = '/?do=search'\ def request(query, params): params['url'] = base_url +\ - search_url.format(query=urlencode({'query': query})) + search_url.format(query=urlencode({'id': query})) return params -- cgit v1.2.3