diff options
| author | Zhijie He <hezhijie0327@hotmail.com> | 2025-05-22 00:25:02 +0800 |
|---|---|---|
| committer | GitHub <noreply@github.com> | 2025-05-21 18:25:02 +0200 |
| commit | 156d1eb8c85c01c94723188859fa7526c7d72698 (patch) | |
| tree | 7a8f627d62580103e504aaf8ac8ec47c132bbdf0 /searx/engines | |
| parent | 365b9426f17d3372647c89307ad7175f862834e8 (diff) | |
[feat] engines: add Naver engine (#4573)
Refactor Naver engine (Web, News, Images, Videos, Autocomplete)
- ref: https://search.naver.com/
- lang: `ko`
- Wikidata: https://www.wikidata.org/wiki/Q485639
Co-authored-by: Bnyro <bnyro@tutanota.com>
Diffstat (limited to 'searx/engines')
| -rw-r--r-- | searx/engines/naver.py | 210 |
1 files changed, 210 insertions, 0 deletions
diff --git a/searx/engines/naver.py b/searx/engines/naver.py new file mode 100644 index 000000000..d92af676e --- /dev/null +++ b/searx/engines/naver.py @@ -0,0 +1,210 @@ +# SPDX-License-Identifier: AGPL-3.0-or-later +# pylint: disable=line-too-long +"""Naver for SearXNG""" + +from urllib.parse import urlencode +from lxml import html + +from searx.exceptions import SearxEngineAPIException, SearxEngineXPathException +from searx.result_types import EngineResults, MainResult +from searx.utils import ( + eval_xpath_getindex, + eval_xpath_list, + eval_xpath, + extract_text, + extr, + html_to_text, + parse_duration_string, + js_variable_to_python, +) + +# engine metadata +about = { + "website": "https://search.naver.com", + "wikidata_id": "Q485639", + "use_official_api": False, + "require_api_key": False, + "results": "HTML", + "language": "ko", +} + +categories = [] +paging = True + +time_range_support = True +time_range_dict = {"day": "1d", "week": "1w", "month": "1m", "year": "1y"} + +base_url = "https://search.naver.com" + +naver_category = "general" +"""Naver supports general, images, news, videos search. + +- ``general``: search for general +- ``images``: search for images +- ``news``: search for news +- ``videos``: search for videos +""" + +# Naver cannot set the number of results on one page, set default value for paging +naver_category_dict = { + "general": { + "start": 15, + "where": "web", + }, + "images": { + "start": 50, + "where": "image", + }, + "news": { + "start": 10, + "where": "news", + }, + "videos": { + "start": 48, + "where": "video", + }, +} + + +def init(_): + if naver_category not in ('general', 'images', 'news', 'videos'): + raise SearxEngineAPIException(f"Unsupported category: {naver_category}") + + +def request(query, params): + query_params = { + "query": query, + } + + if naver_category in naver_category_dict: + query_params["start"] = (params["pageno"] - 1) * naver_category_dict[naver_category]["start"] + 1 + query_params["where"] = naver_category_dict[naver_category]["where"] + + if params["time_range"] in time_range_dict: + query_params["nso"] = f"p:{time_range_dict[params['time_range']]}" + + params["url"] = f"{base_url}/search.naver?{urlencode(query_params)}" + return params + + +def response(resp) -> EngineResults: + parsers = {'general': parse_general, 'images': parse_images, 'news': parse_news, 'videos': parse_videos} + + return parsers[naver_category](resp.text) + + +def parse_general(data): + results = EngineResults() + + dom = html.fromstring(data) + + for item in eval_xpath_list(dom, "//ul[contains(@class, 'lst_total')]/li[contains(@class, 'bx')]"): + thumbnail = None + try: + thumbnail = eval_xpath_getindex(item, ".//div[contains(@class, 'thumb_single')]//img/@data-lazysrc", 0) + except (ValueError, TypeError, SearxEngineXPathException): + pass + + results.add( + MainResult( + title=extract_text(eval_xpath(item, ".//a[contains(@class, 'link_tit')]")), + url=eval_xpath_getindex(item, ".//a[contains(@class, 'link_tit')]/@href", 0), + content=extract_text( + eval_xpath(item, ".//div[contains(@class, 'total_dsc_wrap')]//a[contains(@class, 'api_txt_lines')]") + ), + thumbnail=thumbnail, + ) + ) + + return results + + +def parse_images(data): + results = [] + + match = extr(data, '<script>var imageSearchTabData=', '</script>') + if match: + json = js_variable_to_python(match.strip()) + items = json.get('content', {}).get('items', []) + + for item in items: + results.append( + { + "template": "images.html", + "url": item.get('link'), + "thumbnail_src": item.get('thumb'), + "img_src": item.get('originalUrl'), + "title": html_to_text(item.get('title')), + "source": item.get('source'), + "resolution": f"{item.get('orgWidth')} x {item.get('orgHeight')}", + } + ) + + return results + + +def parse_news(data): + results = EngineResults() + dom = html.fromstring(data) + + for item in eval_xpath_list( + dom, "//div[contains(@class, 'sds-comps-base-layout') and contains(@class, 'sds-comps-full-layout')]" + ): + title = extract_text(eval_xpath(item, ".//span[contains(@class, 'sds-comps-text-type-headline1')]/text()")) + + url = eval_xpath_getindex(item, ".//a[@href and @nocr='1']/@href", 0) + + content = extract_text(eval_xpath(item, ".//span[contains(@class, 'sds-comps-text-type-body1')]")) + + thumbnail = None + try: + thumbnail = eval_xpath_getindex( + item, + ".//div[contains(@class, 'sds-comps-image') and contains(@class, 'sds-rego-thumb-overlay')]//img[@src]/@src", + 0, + ) + except (ValueError, TypeError, SearxEngineXPathException): + pass + + if title and content and url: + results.add( + MainResult( + title=title, + url=url, + content=content, + thumbnail=thumbnail, + ) + ) + + return results + + +def parse_videos(data): + results = [] + + dom = html.fromstring(data) + + for item in eval_xpath_list(dom, "//li[contains(@class, 'video_item')]"): + thumbnail = None + try: + thumbnail = eval_xpath_getindex(item, ".//img[contains(@class, 'thumb')]/@src", 0) + except (ValueError, TypeError, SearxEngineXPathException): + pass + + length = None + try: + length = parse_duration_string(extract_text(eval_xpath(item, ".//span[contains(@class, 'time')]"))) + except (ValueError, TypeError): + pass + + results.append( + { + "template": "videos.html", + "title": extract_text(eval_xpath(item, ".//a[contains(@class, 'info_title')]")), + "url": eval_xpath_getindex(item, ".//a[contains(@class, 'info_title')]/@href", 0), + "thumbnail": thumbnail, + 'length': length, + } + ) + + return results |