| Age | Commit message (Collapse) | Author |
|
the layout https://meta.wikimedia.org/wiki/List_of_Wikipedias has changed
|
|
Most engines that support languages (and regions) use the Accept-Language from
the WEB browser to build a response that fits to the language (and region).
- add new engine option: send_accept_language_header
Signed-off-by: Markus Heiser <markus.heiser@darmarit.de>
|
|
This patch was generated by black [1]::
make format.python
[1] https://github.com/psf/black
Signed-off-by: Markus Heiser <markus.heiser@darmarit.de>
|
|
settings.yml:
* outgoing.networks:
* can contains network definition
* propertiers: enable_http, verify, http2, max_connections, max_keepalive_connections,
keepalive_expiry, local_addresses, support_ipv4, support_ipv6, proxies, max_redirects, retries
* retries: 0 by default, number of times searx retries to send the HTTP request (using different IP & proxy each time)
* local_addresses can be "192.168.0.1/24" (it supports IPv6)
* support_ipv4 & support_ipv6: both True by default
see https://github.com/searx/searx/pull/1034
* each engine can define a "network" section:
* either a full network description
* either reference an existing network
* all HTTP requests of engine use the same HTTP configuration (it was not the case before, see proxy configuration in master)
|
|
fr.wikipedia.org (and it seems not other wikipedia websites),
adds HTML to api_result['displayTitle'].
(Search for '!wp :fr Braid' for example)
The commit uses api_result['title']
|
|
|
|
on some queries (like an IT error message), wikipedia returns an HTTP error 400.
this commit returns an empty result instead of showing an error to the user.
|
|
|
|
move meta information from comment to the about variable
so the preferences, the documentation can show these information
|
|
check HTTP response:
* detect some comme CAPTCHA challenge (no solving). In this case the engine is suspended for long a time.
* otherwise raise HTTPError as before
the check is done in poolrequests.py (was before in search.py).
update qwant, wikipedia, wikidata to use raise_for_httperror instead of raise_for_status
|
|
few cases.
In few cases, the JSON results doesn't contains the key 'type'.
|
|
Add a new parameter "raise_for_status", set by default to True.
When True, any HTTP status code >= 300 raise an exception ( #2332 )
When False, the engine can manage the HTTP status code by itself.
|
|
|
|
|
|
|
|
They're usually IPA pronunciations which are removed
by the API.
|
|
|
|
|
|
|
|
This reverts commit 4d1770398a6af8902e75c0bd885781584d39e796.
|
|
|
|
Add match_language function in utils to match any user given
language code with a list of engine's supported languages.
Also add language_aliases dict on each engine to translate
standard language codes into the custom codes used by the engine.
|
|
|
|
|
|
that support them.
users can still query lesser supported through the :lang_code bang.
|
|
|
|
and refactor method to make it testable without making requests
|
|
utils/fetch_languages.py gets languages supported by each engine and
generates engines_languages.json with each engine's supported language.
|
|
|
|
TODO:
merge attributes
|
|
creates simple multilingual infobox using wikipedia's api
|
|
* writing general mediawiki-engine
* using this engine for wikipedia
* using this engine for uncyclopedia
|
|
* add paging support
* make number_of_results changable
* make result calculation more clear
* add comments
|
|
|
|
|
|
|
|
|
|
|