summaryrefslogtreecommitdiff
path: root/searx/engines/mediawiki.py
diff options
context:
space:
mode:
authorThomas Pointhuber <thomas.pointhuber@gmx.at>2014-09-03 11:40:29 +0200
committerThomas Pointhuber <thomas.pointhuber@gmx.at>2014-09-03 11:40:29 +0200
commit52ad49ccba389346e6216dc708891cbea2b4941d (patch)
tree84c5a729d5aa554e2c943a50fe931f4dfe25d267 /searx/engines/mediawiki.py
parent629a05e149eaaab05a724dd3915ed363c364c796 (diff)
using general mediawiki-engine
* writing general mediawiki-engine * using this engine for wikipedia * using this engine for uncyclopedia
Diffstat (limited to 'searx/engines/mediawiki.py')
-rw-r--r--searx/engines/mediawiki.py63
1 files changed, 54 insertions, 9 deletions
diff --git a/searx/engines/mediawiki.py b/searx/engines/mediawiki.py
index f8cfb9afa..3c7fd4dcb 100644
--- a/searx/engines/mediawiki.py
+++ b/searx/engines/mediawiki.py
@@ -1,22 +1,67 @@
+## Wikipedia (Web)
+#
+# @website http://www.wikipedia.org
+# @provide-api yes (http://www.mediawiki.org/wiki/API:Search)
+#
+# @using-api yes
+# @results JSON
+# @stable yes
+# @parse url, title
+#
+# @todo content
+
from json import loads
from urllib import urlencode, quote
-url = 'https://en.wikipedia.org/'
+# engine dependent config
+categories = ['general']
+language_support = True
+paging = True
+number_of_results = 1
+
+# search-url
+base_url = 'https://{language}.wikipedia.org/'
+search_url = base_url + 'w/api.php?action=query&list=search&{query}&srprop=timestamp&format=json&sroffset={offset}&srlimit={limit}' # noqa
-search_url = url + 'w/api.php?action=query&list=search&{query}&srprop=timestamp&format=json&sroffset={offset}' # noqa
-number_of_results = 10
+# do search-request
+def request(query, params):
+ offset = (params['pageno'] - 1) * number_of_results
+ if params['language'] == 'all':
+ language = 'en'
+ else:
+ language = params['language'].split('_')[0]
+
+ # write search-language back to params, required in response
+ params['language'] = language
-def request(query, params):
- offset = (params['pageno'] - 1) * 10
params['url'] = search_url.format(query=urlencode({'srsearch': query}),
- offset=offset)
+ offset=offset,
+ limit=number_of_results,
+ language=language)
+
return params
+# get response from search-request
def response(resp):
+ results = []
+
search_results = loads(resp.text)
- res = search_results.get('query', {}).get('search', [])
- return [{'url': url + 'wiki/' + quote(result['title'].replace(' ', '_').encode('utf-8')), # noqa
- 'title': result['title']} for result in res[:int(number_of_results)]]
+
+ # return empty array if there are no results
+ if not search_results.get('query', {}).get('search'):
+ return []
+
+ # parse results
+ for result in search_results['query']['search']:
+ url = base_url.format(language=resp.search_params['language']) + 'wiki/' + quote(result['title'].replace(' ', '_').encode('utf-8'))
+
+ # append result
+ results.append({'url': url,
+ 'title': result['title'],
+ 'content': ''})
+
+ # return results
+ return results