diff options
| author | Alexandre Flament <alex@al-f.net> | 2019-04-19 09:59:04 +0200 |
|---|---|---|
| committer | GitHub <noreply@github.com> | 2019-04-19 09:59:04 +0200 |
| commit | 9012ad498cbb13aced7555511c6de6dadc15246a (patch) | |
| tree | a5272a2974acbcdc8f961eff5bf439733ce7690d /searx | |
| parent | 4ca7e68841fccbf75fed21f0d7c0261b2f678c98 (diff) | |
| parent | 629b36d442404ec46a800bc779a5cd8454ca8d2b (diff) | |
Merge branch 'master' into patch-1
Diffstat (limited to 'searx')
| -rw-r--r-- | searx/engines/duckduckgo_images.py | 11 | ||||
| -rw-r--r-- | searx/engines/google_images.py | 38 |
2 files changed, 28 insertions, 21 deletions
diff --git a/searx/engines/duckduckgo_images.py b/searx/engines/duckduckgo_images.py index 8c46ec92d..89924b71c 100644 --- a/searx/engines/duckduckgo_images.py +++ b/searx/engines/duckduckgo_images.py @@ -35,9 +35,12 @@ site_url = 'https://duckduckgo.com/?{query}&iar=images&iax=1&ia=images' # run query in site to get vqd number needed for requesting images # TODO: find a way to get this number without an extra request (is it a hash of the query?) -def get_vqd(query): - res = get(site_url.format(query=urlencode({'q': query}))) +def get_vqd(query, headers): + query_url = site_url.format(query=urlencode({'q': query})) + res = get(query_url, headers=headers) content = res.text + if content.find('vqd=\'') == -1: + raise Exception('Request failed') vqd = content[content.find('vqd=\'') + 5:] vqd = vqd[:vqd.find('\'')] return vqd @@ -47,7 +50,7 @@ def get_vqd(query): def request(query, params): # to avoid running actual external requests when testing if 'is_test' not in params: - vqd = get_vqd(query) + vqd = get_vqd(query, params['headers']) else: vqd = '12345' @@ -74,7 +77,7 @@ def response(resp): try: res_json = loads(content) except: - return [] + raise Exception('Cannot parse results') # parse results for result in res_json['results']: diff --git a/searx/engines/google_images.py b/searx/engines/google_images.py index 504831a10..66a543e57 100644 --- a/searx/engines/google_images.py +++ b/searx/engines/google_images.py @@ -11,7 +11,6 @@ """ from datetime import date, timedelta -from json import loads from lxml import html from searx.url_utils import urlencode, urlparse, parse_qs @@ -39,7 +38,6 @@ time_range_dict = {'day': 'd', # do search-request def request(query, params): search_options = { - 'ijn': params['pageno'] - 1, 'start': (params['pageno'] - 1) * number_of_results } @@ -53,7 +51,7 @@ def request(query, params): search_options['tbs'] = time_range_custom_attr.format(start=start, end=end) if safesearch and params['safesearch']: - search_options['safe'] = 'on' + search_options['safe'] = 'active' params['url'] = search_url.format(query=urlencode({'q': query}), search_options=urlencode(search_options)) @@ -63,24 +61,30 @@ def request(query, params): # get response from search-request def response(resp): + dom = html.fromstring(resp.text) + results = [] + for element in dom.xpath('//div[@id="search"] //td'): + link = element.xpath('./a')[0] - dom = html.fromstring(resp.text) + google_url = urlparse(link.xpath('.//@href')[0]) + query = parse_qs(google_url.query) + source_url = next(iter(query.get('q', [])), None) - # parse results - for img in dom.xpath('//a'): - r = { - 'title': u' '.join(img.xpath('.//div[class="rg_ilmbg"]//text()')), + title_parts = element.xpath('./cite//following-sibling::*/text()') + title_parts.extend(element.xpath('./cite//following-sibling::text()')[:-1]) + + result = { + 'title': ''.join(title_parts), 'content': '', 'template': 'images.html', + 'url': source_url, + 'img_src': source_url, + 'thumbnail_src': next(iter(link.xpath('.//img //@src')), None) } - url = urlparse(img.xpath('.//@href')[0]) - query = parse_qs(url.query) - r['url'] = query['imgrefurl'][0] - r['img_src'] = query['imgurl'][0] - r['thumbnail_src'] = r['img_src'] - # append result - results.append(r) - - # return results + + if not source_url or not result['thumbnail_src']: + continue + + results.append(result) return results |