Merge pull request #1252 from MarcAbonce/search-languages
[mod] Refactor engine's search language handling
This commit is contained in:
		
						commit
						283f6c9053
					
				
										
											
												File diff suppressed because one or more lines are too long
											
										
									
								
							| @ -20,13 +20,14 @@ import sys | |||||||
| import threading | import threading | ||||||
| from os.path import realpath, dirname | from os.path import realpath, dirname | ||||||
| from io import open | from io import open | ||||||
|  | from babel.localedata import locale_identifiers | ||||||
| from flask_babel import gettext | from flask_babel import gettext | ||||||
| from operator import itemgetter | from operator import itemgetter | ||||||
| from json import loads | from json import loads | ||||||
| from requests import get | from requests import get | ||||||
| from searx import settings | from searx import settings | ||||||
| from searx import logger | from searx import logger | ||||||
| from searx.utils import load_module | from searx.utils import load_module, match_language | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
| logger = logger.getChild('engines') | logger = logger.getChild('engines') | ||||||
| @ -38,6 +39,8 @@ engines = {} | |||||||
| categories = {'general': []} | categories = {'general': []} | ||||||
| 
 | 
 | ||||||
| languages = loads(open(engine_dir + '/../data/engines_languages.json', 'r', encoding='utf-8').read()) | languages = loads(open(engine_dir + '/../data/engines_languages.json', 'r', encoding='utf-8').read()) | ||||||
|  | babel_langs = [lang_parts[0] + '-' + lang_parts[-1] if len(lang_parts) > 1 else lang_parts[0] | ||||||
|  |                for lang_parts in (lang_code.split('_') for lang_code in locale_identifiers())] | ||||||
| 
 | 
 | ||||||
| engine_shortcuts = {} | engine_shortcuts = {} | ||||||
| engine_default_args = {'paging': False, | engine_default_args = {'paging': False, | ||||||
| @ -97,6 +100,22 @@ def load_engine(engine_data): | |||||||
|     if engine_data['name'] in languages: |     if engine_data['name'] in languages: | ||||||
|         setattr(engine, 'supported_languages', languages[engine_data['name']]) |         setattr(engine, 'supported_languages', languages[engine_data['name']]) | ||||||
| 
 | 
 | ||||||
|  |     # find custom aliases for non standard language codes | ||||||
|  |     if hasattr(engine, 'supported_languages'): | ||||||
|  |         if hasattr(engine, 'language_aliases'): | ||||||
|  |             language_aliases = getattr(engine, 'language_aliases') | ||||||
|  |         else: | ||||||
|  |             language_aliases = {} | ||||||
|  | 
 | ||||||
|  |         for engine_lang in getattr(engine, 'supported_languages'): | ||||||
|  |             iso_lang = match_language(engine_lang, babel_langs, fallback=None) | ||||||
|  |             if iso_lang and iso_lang != engine_lang and not engine_lang.startswith(iso_lang) and \ | ||||||
|  |                iso_lang not in getattr(engine, 'supported_languages'): | ||||||
|  |                 language_aliases[iso_lang] = engine_lang | ||||||
|  | 
 | ||||||
|  |         if language_aliases: | ||||||
|  |             setattr(engine, 'language_aliases', language_aliases) | ||||||
|  | 
 | ||||||
|     # assign language fetching method if auxiliary method exists |     # assign language fetching method if auxiliary method exists | ||||||
|     if hasattr(engine, '_fetch_supported_languages'): |     if hasattr(engine, '_fetch_supported_languages'): | ||||||
|         setattr(engine, 'fetch_supported_languages', |         setattr(engine, 'fetch_supported_languages', | ||||||
|  | |||||||
| @ -99,13 +99,13 @@ supported_languages = dict(lang_urls, **main_langs) | |||||||
| 
 | 
 | ||||||
| # do search-request | # do search-request | ||||||
| def request(query, params): | def request(query, params): | ||||||
|     # translate the locale (e.g. 'en_US') to language code ('en') |     # translate the locale (e.g. 'en-US') to language code ('en') | ||||||
|     language = locale_to_lang_code(params['language']) |     language = locale_to_lang_code(params['language']) | ||||||
| 
 | 
 | ||||||
|     # if our language is hosted on the main site, we need to add its name |     # if our language is hosted on the main site, we need to add its name | ||||||
|     # to the query in order to narrow the results to that language |     # to the query in order to narrow the results to that language | ||||||
|     if language in main_langs: |     if language in main_langs: | ||||||
|         query += '(' + main_langs[language] + ')' |         query += b' (' + main_langs[language] + b')' | ||||||
| 
 | 
 | ||||||
|     # prepare the request parameters |     # prepare the request parameters | ||||||
|     query = urlencode({'search': query}) |     query = urlencode({'search': query}) | ||||||
|  | |||||||
| @ -16,12 +16,14 @@ | |||||||
| from lxml import html | from lxml import html | ||||||
| from searx.engines.xpath import extract_text | from searx.engines.xpath import extract_text | ||||||
| from searx.url_utils import urlencode | from searx.url_utils import urlencode | ||||||
|  | from searx.utils import match_language | ||||||
| 
 | 
 | ||||||
| # engine dependent config | # engine dependent config | ||||||
| categories = ['general'] | categories = ['general'] | ||||||
| paging = True | paging = True | ||||||
| language_support = True | language_support = True | ||||||
| supported_languages_url = 'https://www.bing.com/account/general' | supported_languages_url = 'https://www.bing.com/account/general' | ||||||
|  | language_aliases = {'zh-CN': 'zh-CHS', 'zh-TW': 'zh-CHT', 'zh-HK': 'zh-CHT'} | ||||||
| 
 | 
 | ||||||
| # search-url | # search-url | ||||||
| base_url = 'https://www.bing.com/' | base_url = 'https://www.bing.com/' | ||||||
| @ -32,9 +34,9 @@ search_string = 'search?{query}&first={offset}' | |||||||
| def request(query, params): | def request(query, params): | ||||||
|     offset = (params['pageno'] - 1) * 10 + 1 |     offset = (params['pageno'] - 1) * 10 + 1 | ||||||
| 
 | 
 | ||||||
|     lang = params['language'].split('-')[0].upper() |     lang = match_language(params['language'], supported_languages, language_aliases) | ||||||
| 
 | 
 | ||||||
|     query = u'language:{} {}'.format(lang, query.decode('utf-8')).encode('utf-8') |     query = u'language:{} {}'.format(lang.split('-')[0].upper(), query.decode('utf-8')).encode('utf-8') | ||||||
| 
 | 
 | ||||||
|     search_path = search_string.format( |     search_path = search_string.format( | ||||||
|         query=urlencode({'q': query}), |         query=urlencode({'q': query}), | ||||||
|  | |||||||
| @ -19,6 +19,7 @@ from lxml import html | |||||||
| from json import loads | from json import loads | ||||||
| import re | import re | ||||||
| from searx.url_utils import urlencode | from searx.url_utils import urlencode | ||||||
|  | from searx.utils import match_language | ||||||
| 
 | 
 | ||||||
| # engine dependent config | # engine dependent config | ||||||
| categories = ['images'] | categories = ['images'] | ||||||
| @ -46,26 +47,6 @@ safesearch_types = {2: 'STRICT', | |||||||
| _quote_keys_regex = re.compile('({|,)([a-z][a-z0-9]*):(")', re.I | re.U) | _quote_keys_regex = re.compile('({|,)([a-z][a-z0-9]*):(")', re.I | re.U) | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
| # get supported region code |  | ||||||
| def get_region_code(lang, lang_list=None): |  | ||||||
|     region = None |  | ||||||
|     if lang in (lang_list or supported_languages): |  | ||||||
|         region = lang |  | ||||||
|     elif lang.startswith('no'): |  | ||||||
|         region = 'nb-NO' |  | ||||||
|     else: |  | ||||||
|         # try to get a supported country code with language |  | ||||||
|         lang = lang.split('-')[0] |  | ||||||
|         for lc in (lang_list or supported_languages): |  | ||||||
|             if lang == lc.split('-')[0]: |  | ||||||
|                 region = lc |  | ||||||
|                 break |  | ||||||
|     if region: |  | ||||||
|         return region.lower() |  | ||||||
|     else: |  | ||||||
|         return 'en-us' |  | ||||||
| 
 |  | ||||||
| 
 |  | ||||||
| # do search-request | # do search-request | ||||||
| def request(query, params): | def request(query, params): | ||||||
|     offset = (params['pageno'] - 1) * 10 + 1 |     offset = (params['pageno'] - 1) * 10 + 1 | ||||||
| @ -74,7 +55,7 @@ def request(query, params): | |||||||
|         query=urlencode({'q': query}), |         query=urlencode({'q': query}), | ||||||
|         offset=offset) |         offset=offset) | ||||||
| 
 | 
 | ||||||
|     language = get_region_code(params['language']) |     language = match_language(params['language'], supported_languages).lower() | ||||||
| 
 | 
 | ||||||
|     params['cookies']['SRCHHPGUSR'] = \ |     params['cookies']['SRCHHPGUSR'] = \ | ||||||
|         'ADLT=' + safesearch_types.get(params['safesearch'], 'DEMOTE') |         'ADLT=' + safesearch_types.get(params['safesearch'], 'DEMOTE') | ||||||
|  | |||||||
| @ -14,8 +14,8 @@ | |||||||
| from datetime import datetime | from datetime import datetime | ||||||
| from dateutil import parser | from dateutil import parser | ||||||
| from lxml import etree | from lxml import etree | ||||||
| from searx.utils import list_get | from searx.utils import list_get, match_language | ||||||
| from searx.engines.bing import _fetch_supported_languages, supported_languages_url | from searx.engines.bing import _fetch_supported_languages, supported_languages_url, language_aliases | ||||||
| from searx.url_utils import urlencode, urlparse, parse_qsl | from searx.url_utils import urlencode, urlparse, parse_qsl | ||||||
| 
 | 
 | ||||||
| # engine dependent config | # engine dependent config | ||||||
| @ -71,7 +71,7 @@ def request(query, params): | |||||||
| 
 | 
 | ||||||
|     offset = (params['pageno'] - 1) * 10 + 1 |     offset = (params['pageno'] - 1) * 10 + 1 | ||||||
| 
 | 
 | ||||||
|     language = params['language'] |     language = match_language(params['language'], supported_languages, language_aliases) | ||||||
| 
 | 
 | ||||||
|     params['url'] = _get_url(query, language, offset, params['time_range']) |     params['url'] = _get_url(query, language, offset, params['time_range']) | ||||||
| 
 | 
 | ||||||
|  | |||||||
| @ -12,9 +12,10 @@ | |||||||
| 
 | 
 | ||||||
| from json import loads | from json import loads | ||||||
| from lxml import html | from lxml import html | ||||||
| from searx.engines.bing_images import _fetch_supported_languages, supported_languages_url, get_region_code | from searx.engines.bing_images import _fetch_supported_languages, supported_languages_url | ||||||
| from searx.engines.xpath import extract_text | from searx.engines.xpath import extract_text | ||||||
| from searx.url_utils import urlencode | from searx.url_utils import urlencode | ||||||
|  | from searx.utils import match_language | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
| categories = ['videos'] | categories = ['videos'] | ||||||
| @ -47,8 +48,8 @@ def request(query, params): | |||||||
|         'ADLT=' + safesearch_types.get(params['safesearch'], 'DEMOTE') |         'ADLT=' + safesearch_types.get(params['safesearch'], 'DEMOTE') | ||||||
| 
 | 
 | ||||||
|     # language cookie |     # language cookie | ||||||
|     region = get_region_code(params['language'], lang_list=supported_languages) |     language = match_language(params['language'], supported_languages).lower() | ||||||
|     params['cookies']['_EDGE_S'] = 'mkt=' + region + '&F=1' |     params['cookies']['_EDGE_S'] = 'mkt=' + language + '&F=1' | ||||||
| 
 | 
 | ||||||
|     # query and paging |     # query and paging | ||||||
|     params['url'] = search_url.format(query=urlencode({'q': query}), |     params['url'] = search_url.format(query=urlencode({'q': query}), | ||||||
|  | |||||||
| @ -15,6 +15,7 @@ | |||||||
| from json import loads | from json import loads | ||||||
| from datetime import datetime | from datetime import datetime | ||||||
| from searx.url_utils import urlencode | from searx.url_utils import urlencode | ||||||
|  | from searx.utils import match_language | ||||||
| 
 | 
 | ||||||
| # engine dependent config | # engine dependent config | ||||||
| categories = ['videos'] | categories = ['videos'] | ||||||
| @ -32,7 +33,7 @@ supported_languages_url = 'https://api.dailymotion.com/languages' | |||||||
| 
 | 
 | ||||||
| # do search-request | # do search-request | ||||||
| def request(query, params): | def request(query, params): | ||||||
|     locale = params['language'] |     locale = match_language(params['language'], supported_languages) | ||||||
| 
 | 
 | ||||||
|     params['url'] = search_url.format( |     params['url'] = search_url.format( | ||||||
|         query=urlencode({'search': query, 'localization': locale}), |         query=urlencode({'search': query, 'localization': locale}), | ||||||
|  | |||||||
| @ -18,14 +18,25 @@ from json import loads | |||||||
| from searx.engines.xpath import extract_text | from searx.engines.xpath import extract_text | ||||||
| from searx.poolrequests import get | from searx.poolrequests import get | ||||||
| from searx.url_utils import urlencode | from searx.url_utils import urlencode | ||||||
|  | from searx.utils import match_language | ||||||
| 
 | 
 | ||||||
| # engine dependent config | # engine dependent config | ||||||
| categories = ['general'] | categories = ['general'] | ||||||
| paging = True | paging = True | ||||||
| language_support = True | language_support = True | ||||||
| supported_languages_url = 'https://duckduckgo.com/d2030.js' | supported_languages_url = 'https://duckduckgo.com/util/u172.js' | ||||||
| time_range_support = True | time_range_support = True | ||||||
| 
 | 
 | ||||||
|  | language_aliases = { | ||||||
|  |     'ar-SA': 'ar-XA', | ||||||
|  |     'es-419': 'es-XL', | ||||||
|  |     'ja': 'jp-JP', | ||||||
|  |     'ko': 'kr-KR', | ||||||
|  |     'sl-SI': 'sl-SL', | ||||||
|  |     'zh-TW': 'tzh-TW', | ||||||
|  |     'zh-HK': 'tzh-HK' | ||||||
|  | } | ||||||
|  | 
 | ||||||
| # search-url | # search-url | ||||||
| url = 'https://duckduckgo.com/html?{query}&s={offset}&dc={dc_param}' | url = 'https://duckduckgo.com/html?{query}&s={offset}&dc={dc_param}' | ||||||
| time_range_url = '&df={range}' | time_range_url = '&df={range}' | ||||||
| @ -42,34 +53,12 @@ content_xpath = './/a[@class="result__snippet"]' | |||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
| # match query's language to a region code that duckduckgo will accept | # match query's language to a region code that duckduckgo will accept | ||||||
| def get_region_code(lang, lang_list=None): | def get_region_code(lang, lang_list=[]): | ||||||
|     # custom fixes for languages |     lang_code = match_language(lang, lang_list, language_aliases, 'wt-WT') | ||||||
|     if lang[:2] == 'ja': |     lang_parts = lang_code.split('-') | ||||||
|         region_code = 'jp-jp' | 
 | ||||||
|     elif lang[:2] == 'sl': |     # country code goes first | ||||||
|         region_code = 'sl-sl' |     return lang_parts[1].lower() + '-' + lang_parts[0].lower() | ||||||
|     elif lang == 'zh-TW': |  | ||||||
|         region_code = 'tw-tzh' |  | ||||||
|     elif lang == 'zh-HK': |  | ||||||
|         region_code = 'hk-tzh' |  | ||||||
|     elif lang[-2:] == 'SA': |  | ||||||
|         region_code = 'xa-' + lang.split('-')[0] |  | ||||||
|     elif lang[-2:] == 'GB': |  | ||||||
|         region_code = 'uk-' + lang.split('-')[0] |  | ||||||
|     else: |  | ||||||
|         region_code = lang.split('-') |  | ||||||
|         if len(region_code) == 2: |  | ||||||
|             # country code goes first |  | ||||||
|             region_code = region_code[1].lower() + '-' + region_code[0].lower() |  | ||||||
|         else: |  | ||||||
|             # tries to get a country code from language |  | ||||||
|             region_code = region_code[0].lower() |  | ||||||
|             for lc in (lang_list or supported_languages): |  | ||||||
|                 lc = lc.split('-') |  | ||||||
|                 if region_code == lc[0]: |  | ||||||
|                     region_code = lc[1].lower() + '-' + lc[0].lower() |  | ||||||
|                     break |  | ||||||
|     return region_code |  | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
| # do search-request | # do search-request | ||||||
| @ -79,7 +68,7 @@ def request(query, params): | |||||||
| 
 | 
 | ||||||
|     offset = (params['pageno'] - 1) * 30 |     offset = (params['pageno'] - 1) * 30 | ||||||
| 
 | 
 | ||||||
|     region_code = get_region_code(params['language']) |     region_code = get_region_code(params['language'], supported_languages) | ||||||
|     params['url'] = url.format( |     params['url'] = url.format( | ||||||
|         query=urlencode({'q': query, 'kl': region_code}), offset=offset, dc_param=offset) |         query=urlencode({'q': query, 'kl': region_code}), offset=offset, dc_param=offset) | ||||||
| 
 | 
 | ||||||
|  | |||||||
| @ -2,9 +2,9 @@ import json | |||||||
| from lxml import html | from lxml import html | ||||||
| from re import compile | from re import compile | ||||||
| from searx.engines.xpath import extract_text | from searx.engines.xpath import extract_text | ||||||
| from searx.engines.duckduckgo import _fetch_supported_languages, supported_languages_url | from searx.engines.duckduckgo import _fetch_supported_languages, supported_languages_url, language_aliases | ||||||
| from searx.url_utils import urlencode | from searx.url_utils import urlencode | ||||||
| from searx.utils import html_to_text | from searx.utils import html_to_text, match_language | ||||||
| 
 | 
 | ||||||
| url = 'https://api.duckduckgo.com/'\ | url = 'https://api.duckduckgo.com/'\ | ||||||
|     + '?{query}&format=json&pretty=0&no_redirect=1&d=1' |     + '?{query}&format=json&pretty=0&no_redirect=1&d=1' | ||||||
| @ -24,7 +24,8 @@ def result_to_text(url, text, htmlResult): | |||||||
| 
 | 
 | ||||||
| def request(query, params): | def request(query, params): | ||||||
|     params['url'] = url.format(query=urlencode({'q': query})) |     params['url'] = url.format(query=urlencode({'q': query})) | ||||||
|     params['headers']['Accept-Language'] = params['language'].split('-')[0] |     language = match_language(params['language'], supported_languages, language_aliases) | ||||||
|  |     params['headers']['Accept-Language'] = language.split('-')[0] | ||||||
|     return params |     return params | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
|  | |||||||
| @ -15,7 +15,10 @@ | |||||||
| 
 | 
 | ||||||
| from json import loads | from json import loads | ||||||
| from searx.engines.xpath import extract_text | from searx.engines.xpath import extract_text | ||||||
| from searx.engines.duckduckgo import _fetch_supported_languages, supported_languages_url, get_region_code | from searx.engines.duckduckgo import ( | ||||||
|  |     _fetch_supported_languages, supported_languages_url, | ||||||
|  |     get_region_code, language_aliases | ||||||
|  | ) | ||||||
| from searx.poolrequests import get | from searx.poolrequests import get | ||||||
| from searx.url_utils import urlencode | from searx.url_utils import urlencode | ||||||
| 
 | 
 | ||||||
|  | |||||||
| @ -14,6 +14,7 @@ from lxml import html, etree | |||||||
| from searx.engines.xpath import extract_text, extract_url | from searx.engines.xpath import extract_text, extract_url | ||||||
| from searx import logger | from searx import logger | ||||||
| from searx.url_utils import urlencode, urlparse, parse_qsl | from searx.url_utils import urlencode, urlparse, parse_qsl | ||||||
|  | from searx.utils import match_language | ||||||
| 
 | 
 | ||||||
| logger = logger.getChild('google engine') | logger = logger.getChild('google engine') | ||||||
| 
 | 
 | ||||||
| @ -72,7 +73,7 @@ country_to_hostname = { | |||||||
|     'RO': 'www.google.ro',  # Romania |     'RO': 'www.google.ro',  # Romania | ||||||
|     'RU': 'www.google.ru',  # Russia |     'RU': 'www.google.ru',  # Russia | ||||||
|     'SK': 'www.google.sk',  # Slovakia |     'SK': 'www.google.sk',  # Slovakia | ||||||
|     'SL': 'www.google.si',  # Slovenia (SL -> si) |     'SI': 'www.google.si',  # Slovenia | ||||||
|     'SE': 'www.google.se',  # Sweden |     'SE': 'www.google.se',  # Sweden | ||||||
|     'TH': 'www.google.co.th',  # Thailand |     'TH': 'www.google.co.th',  # Thailand | ||||||
|     'TR': 'www.google.com.tr',  # Turkey |     'TR': 'www.google.com.tr',  # Turkey | ||||||
| @ -165,22 +166,20 @@ def extract_text_from_dom(result, xpath): | |||||||
| def request(query, params): | def request(query, params): | ||||||
|     offset = (params['pageno'] - 1) * 10 |     offset = (params['pageno'] - 1) * 10 | ||||||
| 
 | 
 | ||||||
|     # temporary fix until a way of supporting en-US is found |     language = match_language(params['language'], supported_languages) | ||||||
|     if params['language'] == 'en-US': |     language_array = language.split('-') | ||||||
|         params['language'] = 'en-GB' |     if params['language'].find('-') > 0: | ||||||
| 
 |         country = params['language'].split('-')[1] | ||||||
|     if params['language'][:2] == 'jv': |     elif len(language_array) == 2: | ||||||
|         language = 'jw' |         country = language_array[1] | ||||||
|         country = 'ID' |  | ||||||
|         url_lang = 'lang_jw' |  | ||||||
|     else: |     else: | ||||||
|         language_array = params['language'].lower().split('-') |         country = 'US' | ||||||
|         if len(language_array) == 2: | 
 | ||||||
|             country = language_array[1] |     # temporary fix until a way of supporting en-US is found | ||||||
|         else: |     if language == 'en-US': | ||||||
|             country = 'US' |         country = 'GB' | ||||||
|         language = language_array[0] + ',' + language_array[0] + '-' + country | 
 | ||||||
|         url_lang = 'lang_' + language_array[0] |     url_lang = 'lang_' + language | ||||||
| 
 | 
 | ||||||
|     if use_locale_domain: |     if use_locale_domain: | ||||||
|         google_hostname = country_to_hostname.get(country.upper(), default_hostname) |         google_hostname = country_to_hostname.get(country.upper(), default_hostname) | ||||||
| @ -196,7 +195,7 @@ def request(query, params): | |||||||
|     if params['time_range'] in time_range_dict: |     if params['time_range'] in time_range_dict: | ||||||
|         params['url'] += time_range_search.format(range=time_range_dict[params['time_range']]) |         params['url'] += time_range_search.format(range=time_range_dict[params['time_range']]) | ||||||
| 
 | 
 | ||||||
|     params['headers']['Accept-Language'] = language |     params['headers']['Accept-Language'] = language + ',' + language + '-' + country | ||||||
|     params['headers']['Accept'] = 'text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8' |     params['headers']['Accept'] = 'text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8' | ||||||
| 
 | 
 | ||||||
|     params['google_hostname'] = google_hostname |     params['google_hostname'] = google_hostname | ||||||
|  | |||||||
| @ -13,6 +13,7 @@ | |||||||
| from lxml import html | from lxml import html | ||||||
| from searx.engines.google import _fetch_supported_languages, supported_languages_url | from searx.engines.google import _fetch_supported_languages, supported_languages_url | ||||||
| from searx.url_utils import urlencode | from searx.url_utils import urlencode | ||||||
|  | from searx.utils import match_language | ||||||
| 
 | 
 | ||||||
| # search-url | # search-url | ||||||
| categories = ['news'] | categories = ['news'] | ||||||
| @ -50,8 +51,9 @@ def request(query, params): | |||||||
|     params['url'] = search_url.format(query=urlencode({'q': query}), |     params['url'] = search_url.format(query=urlencode({'q': query}), | ||||||
|                                       search_options=urlencode(search_options)) |                                       search_options=urlencode(search_options)) | ||||||
| 
 | 
 | ||||||
|     language_array = params['language'].lower().split('-') |     language = match_language(params['language'], supported_languages).split('-')[0] | ||||||
|     params['url'] += '&lr=lang_' + language_array[0] |     if language: | ||||||
|  |         params['url'] += '&lr=lang_' + language | ||||||
| 
 | 
 | ||||||
|     return params |     return params | ||||||
| 
 | 
 | ||||||
|  | |||||||
| @ -14,6 +14,7 @@ from datetime import datetime | |||||||
| from json import loads | from json import loads | ||||||
| from searx.utils import html_to_text | from searx.utils import html_to_text | ||||||
| from searx.url_utils import urlencode | from searx.url_utils import urlencode | ||||||
|  | from searx.utils import match_language | ||||||
| 
 | 
 | ||||||
| # engine dependent config | # engine dependent config | ||||||
| categories = None | categories = None | ||||||
| @ -45,16 +46,8 @@ def request(query, params): | |||||||
|                                    offset=offset) |                                    offset=offset) | ||||||
| 
 | 
 | ||||||
|     # add language tag |     # add language tag | ||||||
|     if params['language'] == 'no' or params['language'].startswith('no-'): |     language = match_language(params['language'], supported_languages) | ||||||
|         params['language'] = params['language'].replace('no', 'nb', 1) |     params['url'] += '&locale=' + language.replace('-', '_').lower() | ||||||
|     if params['language'].find('-') < 0: |  | ||||||
|         # tries to get a country code from language |  | ||||||
|         for lang in supported_languages: |  | ||||||
|             lc = lang.split('-') |  | ||||||
|             if params['language'] == lc[0]: |  | ||||||
|                 params['language'] = lang |  | ||||||
|                 break |  | ||||||
|     params['url'] += '&locale=' + params['language'].replace('-', '_').lower() |  | ||||||
| 
 | 
 | ||||||
|     return params |     return params | ||||||
| 
 | 
 | ||||||
|  | |||||||
| @ -14,6 +14,7 @@ from json import loads | |||||||
| import re | import re | ||||||
| from lxml.html import fromstring | from lxml.html import fromstring | ||||||
| from searx.url_utils import unquote, urlencode | from searx.url_utils import unquote, urlencode | ||||||
|  | from searx.utils import match_language | ||||||
| 
 | 
 | ||||||
| # engine dependent config | # engine dependent config | ||||||
| categories = ['general', 'images'] | categories = ['general', 'images'] | ||||||
| @ -35,11 +36,8 @@ regex_img_url_remove_start = re.compile(b'^https?://i\.swisscows\.ch/\?link=') | |||||||
| 
 | 
 | ||||||
| # do search-request | # do search-request | ||||||
| def request(query, params): | def request(query, params): | ||||||
|     if params['language'].split('-')[0] == 'no': |     region = match_language(params['language'], supported_languages) | ||||||
|         region = 'nb-NO' |     ui_language = region.split('-')[0] | ||||||
|     else: |  | ||||||
|         region = params['language'] |  | ||||||
|         ui_language = params['language'].split('-')[0] |  | ||||||
| 
 | 
 | ||||||
|     search_path = search_string.format( |     search_path = search_string.format( | ||||||
|         query=urlencode({'query': query, 'uiLanguage': ui_language, 'region': region}), |         query=urlencode({'query': query, 'uiLanguage': ui_language, 'region': region}), | ||||||
|  | |||||||
| @ -16,6 +16,7 @@ from searx.poolrequests import get | |||||||
| from searx.engines.xpath import extract_text | from searx.engines.xpath import extract_text | ||||||
| from searx.engines.wikipedia import _fetch_supported_languages, supported_languages_url | from searx.engines.wikipedia import _fetch_supported_languages, supported_languages_url | ||||||
| from searx.url_utils import urlencode | from searx.url_utils import urlencode | ||||||
|  | from searx.utils import match_language | ||||||
| 
 | 
 | ||||||
| from json import loads | from json import loads | ||||||
| from lxml.html import fromstring | from lxml.html import fromstring | ||||||
| @ -56,7 +57,7 @@ calendar_name_xpath = './/sup[contains(@class,"wb-calendar-name")]' | |||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
| def request(query, params): | def request(query, params): | ||||||
|     language = params['language'].split('-')[0] |     language = match_language(params['language'], supported_languages).split('-')[0] | ||||||
| 
 | 
 | ||||||
|     params['url'] = url_search.format( |     params['url'] = url_search.format( | ||||||
|         query=urlencode({'label': query, 'language': language})) |         query=urlencode({'label': query, 'language': language})) | ||||||
| @ -68,7 +69,7 @@ def response(resp): | |||||||
|     html = fromstring(resp.text) |     html = fromstring(resp.text) | ||||||
|     wikidata_ids = html.xpath(wikidata_ids_xpath) |     wikidata_ids = html.xpath(wikidata_ids_xpath) | ||||||
| 
 | 
 | ||||||
|     language = resp.search_params['language'].split('-')[0] |     language = match_language(resp.search_params['language'], supported_languages).split('-')[0] | ||||||
| 
 | 
 | ||||||
|     # TODO: make requests asynchronous to avoid timeout when result_count > 1 |     # TODO: make requests asynchronous to avoid timeout when result_count > 1 | ||||||
|     for wikidata_id in wikidata_ids[:result_count]: |     for wikidata_id in wikidata_ids[:result_count]: | ||||||
|  | |||||||
| @ -13,6 +13,7 @@ | |||||||
| from json import loads | from json import loads | ||||||
| from lxml.html import fromstring | from lxml.html import fromstring | ||||||
| from searx.url_utils import quote, urlencode | from searx.url_utils import quote, urlencode | ||||||
|  | from searx.utils import match_language | ||||||
| 
 | 
 | ||||||
| # search-url | # search-url | ||||||
| base_url = u'https://{language}.wikipedia.org/' | base_url = u'https://{language}.wikipedia.org/' | ||||||
| @ -30,13 +31,7 @@ supported_languages_url = 'https://meta.wikimedia.org/wiki/List_of_Wikipedias' | |||||||
| 
 | 
 | ||||||
| # set language in base_url | # set language in base_url | ||||||
| def url_lang(lang): | def url_lang(lang): | ||||||
|     lang = lang.split('-')[0] |     return match_language(lang, supported_languages).split('-')[0] | ||||||
|     if lang not in supported_languages: |  | ||||||
|         language = 'en' |  | ||||||
|     else: |  | ||||||
|         language = lang |  | ||||||
| 
 |  | ||||||
|     return language |  | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
| # do search-request | # do search-request | ||||||
|  | |||||||
| @ -14,6 +14,7 @@ | |||||||
| from lxml import html | from lxml import html | ||||||
| from searx.engines.xpath import extract_text, extract_url | from searx.engines.xpath import extract_text, extract_url | ||||||
| from searx.url_utils import unquote, urlencode | from searx.url_utils import unquote, urlencode | ||||||
|  | from searx.utils import match_language | ||||||
| 
 | 
 | ||||||
| # engine dependent config | # engine dependent config | ||||||
| categories = ['general'] | categories = ['general'] | ||||||
| @ -39,6 +40,8 @@ time_range_dict = {'day': ['1d', 'd'], | |||||||
|                    'week': ['1w', 'w'], |                    'week': ['1w', 'w'], | ||||||
|                    'month': ['1m', 'm']} |                    'month': ['1m', 'm']} | ||||||
| 
 | 
 | ||||||
|  | language_aliases = {'zh-CN': 'zh-CHS', 'zh-TW': 'zh-CHT', 'zh-HK': 'zh-CHT'} | ||||||
|  | 
 | ||||||
| 
 | 
 | ||||||
| # remove yahoo-specific tracking-url | # remove yahoo-specific tracking-url | ||||||
| def parse_url(url_string): | def parse_url(url_string): | ||||||
| @ -70,23 +73,16 @@ def _get_url(query, offset, language, time_range): | |||||||
|                                         lang=language) |                                         lang=language) | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
| def _get_language(params): |  | ||||||
|     if params['language'][:2] == 'zh': |  | ||||||
|         if params['language'] == 'zh' or params['language'] == 'zh-CH': |  | ||||||
|             return 'szh' |  | ||||||
|         else: |  | ||||||
|             return 'tzh' |  | ||||||
|     else: |  | ||||||
|         return params['language'].split('-')[0] |  | ||||||
| 
 |  | ||||||
| 
 |  | ||||||
| # do search-request | # do search-request | ||||||
| def request(query, params): | def request(query, params): | ||||||
|     if params['time_range'] and params['time_range'] not in time_range_dict: |     if params['time_range'] and params['time_range'] not in time_range_dict: | ||||||
|         return params |         return params | ||||||
| 
 | 
 | ||||||
|     offset = (params['pageno'] - 1) * 10 + 1 |     offset = (params['pageno'] - 1) * 10 + 1 | ||||||
|     language = _get_language(params) |     language = match_language(params['language'], supported_languages, language_aliases) | ||||||
|  |     if language not in language_aliases.values(): | ||||||
|  |         language = language.split('-')[0] | ||||||
|  |     language = language.replace('-', '_').lower() | ||||||
| 
 | 
 | ||||||
|     params['url'] = _get_url(query, offset, language, params['time_range']) |     params['url'] = _get_url(query, offset, language, params['time_range']) | ||||||
| 
 | 
 | ||||||
| @ -145,7 +141,11 @@ def _fetch_supported_languages(resp): | |||||||
|     dom = html.fromstring(resp.text) |     dom = html.fromstring(resp.text) | ||||||
|     options = dom.xpath('//div[@id="yschlang"]/span/label/input') |     options = dom.xpath('//div[@id="yschlang"]/span/label/input') | ||||||
|     for option in options: |     for option in options: | ||||||
|         code = option.xpath('./@value')[0][5:].replace('_', '-') |         code_parts = option.xpath('./@value')[0][5:].split('_') | ||||||
|  |         if len(code_parts) == 2: | ||||||
|  |             code = code_parts[0] + '-' + code_parts[1].upper() | ||||||
|  |         else: | ||||||
|  |             code = code_parts[0] | ||||||
|         supported_languages.append(code) |         supported_languages.append(code) | ||||||
| 
 | 
 | ||||||
|     return supported_languages |     return supported_languages | ||||||
|  | |||||||
| @ -13,9 +13,12 @@ import re | |||||||
| from datetime import datetime, timedelta | from datetime import datetime, timedelta | ||||||
| from lxml import html | from lxml import html | ||||||
| from searx.engines.xpath import extract_text, extract_url | from searx.engines.xpath import extract_text, extract_url | ||||||
| from searx.engines.yahoo import parse_url, _fetch_supported_languages, supported_languages_url | from searx.engines.yahoo import ( | ||||||
|  |     parse_url, _fetch_supported_languages, supported_languages_url, language_aliases | ||||||
|  | ) | ||||||
| from dateutil import parser | from dateutil import parser | ||||||
| from searx.url_utils import urlencode | from searx.url_utils import urlencode | ||||||
|  | from searx.utils import match_language | ||||||
| 
 | 
 | ||||||
| # engine dependent config | # engine dependent config | ||||||
| categories = ['news'] | categories = ['news'] | ||||||
| @ -38,7 +41,7 @@ suggestion_xpath = '//div[contains(@class,"VerALSOTRY")]//a' | |||||||
| def request(query, params): | def request(query, params): | ||||||
|     offset = (params['pageno'] - 1) * 10 + 1 |     offset = (params['pageno'] - 1) * 10 + 1 | ||||||
| 
 | 
 | ||||||
|     language = params['language'].split('-')[0] |     language = match_language(params['language'], supported_languages, language_aliases).split('-')[0] | ||||||
| 
 | 
 | ||||||
|     params['url'] = search_url.format(offset=offset, |     params['url'] = search_url.format(offset=offset, | ||||||
|                                       query=urlencode({'p': query}), |                                       query=urlencode({'p': query}), | ||||||
|  | |||||||
| @ -5,11 +5,7 @@ | |||||||
| language_codes = ( | language_codes = ( | ||||||
|     (u"ar-SA", u"العربية", u"", u"Arabic"), |     (u"ar-SA", u"العربية", u"", u"Arabic"), | ||||||
|     (u"bg-BG", u"Български", u"", u"Bulgarian"), |     (u"bg-BG", u"Български", u"", u"Bulgarian"), | ||||||
|     (u"ca", u"Català", u"", u"Catalan"), |     (u"ca-ES", u"Català", u"", u"Catalan"), | ||||||
|     (u"ca-AD", u"Català", u"Andorra", u"Catalan"), |  | ||||||
|     (u"ca-CT", u"Català", u"", u"Catalan"), |  | ||||||
|     (u"ca-ES", u"Català", u"Espanya", u"Catalan"), |  | ||||||
|     (u"ca-FR", u"Català", u"França", u"Catalan"), |  | ||||||
|     (u"cs-CZ", u"Čeština", u"", u"Czech"), |     (u"cs-CZ", u"Čeština", u"", u"Czech"), | ||||||
|     (u"da-DK", u"Dansk", u"", u"Danish"), |     (u"da-DK", u"Dansk", u"", u"Danish"), | ||||||
|     (u"de", u"Deutsch", u"", u"German"), |     (u"de", u"Deutsch", u"", u"German"), | ||||||
| @ -21,55 +17,51 @@ language_codes = ( | |||||||
|     (u"en-AU", u"English", u"Australia", u"English"), |     (u"en-AU", u"English", u"Australia", u"English"), | ||||||
|     (u"en-CA", u"English", u"Canada", u"English"), |     (u"en-CA", u"English", u"Canada", u"English"), | ||||||
|     (u"en-GB", u"English", u"United Kingdom", u"English"), |     (u"en-GB", u"English", u"United Kingdom", u"English"), | ||||||
|     (u"en-ID", u"English", u"Indonesia", u"English"), |  | ||||||
|     (u"en-IE", u"English", u"Ireland", u"English"), |  | ||||||
|     (u"en-IN", u"English", u"India", u"English"), |     (u"en-IN", u"English", u"India", u"English"), | ||||||
|     (u"en-MY", u"English", u"Malaysia", u"English"), |     (u"en-MY", u"English", u"Malaysia", u"English"), | ||||||
|     (u"en-NZ", u"English", u"New Zealand", u"English"), |  | ||||||
|     (u"en-PH", u"English", u"Philippines", u"English"), |  | ||||||
|     (u"en-SG", u"English", u"Singapore", u"English"), |  | ||||||
|     (u"en-US", u"English", u"United States", u"English"), |     (u"en-US", u"English", u"United States", u"English"), | ||||||
|     (u"en-ZA", u"English", u"South Africa", u"English"), |  | ||||||
|     (u"es", u"Español", u"", u"Spanish"), |     (u"es", u"Español", u"", u"Spanish"), | ||||||
|     (u"es-AD", u"Español", u"Andorra", u"Spanish"), |  | ||||||
|     (u"es-AR", u"Español", u"Argentina", u"Spanish"), |     (u"es-AR", u"Español", u"Argentina", u"Spanish"), | ||||||
|     (u"es-CL", u"Español", u"Chile", u"Spanish"), |  | ||||||
|     (u"es-CO", u"Español", u"Colombia", u"Spanish"), |  | ||||||
|     (u"es-ES", u"Español", u"España", u"Spanish"), |     (u"es-ES", u"Español", u"España", u"Spanish"), | ||||||
|     (u"es-MX", u"Español", u"México", u"Spanish"), |     (u"es-MX", u"Español", u"México", u"Spanish"), | ||||||
|     (u"es-PE", u"Español", u"Perú", u"Spanish"), |  | ||||||
|     (u"es-US", u"Español", u"Estados Unidos", u"Spanish"), |  | ||||||
|     (u"et-EE", u"Eesti", u"", u"Estonian"), |     (u"et-EE", u"Eesti", u"", u"Estonian"), | ||||||
|  |     (u"fa-IR", u"فارسی", u"", u"Persian"), | ||||||
|     (u"fi-FI", u"Suomi", u"", u"Finnish"), |     (u"fi-FI", u"Suomi", u"", u"Finnish"), | ||||||
|     (u"fr", u"Français", u"", u"French"), |     (u"fr", u"Français", u"", u"French"), | ||||||
|     (u"fr-AD", u"Français", u"Andorre", u"French"), |  | ||||||
|     (u"fr-BE", u"Français", u"Belgique", u"French"), |     (u"fr-BE", u"Français", u"Belgique", u"French"), | ||||||
|     (u"fr-CA", u"Français", u"Canada", u"French"), |     (u"fr-CA", u"Français", u"Canada", u"French"), | ||||||
|     (u"fr-CH", u"Français", u"Suisse", u"French"), |     (u"fr-CH", u"Français", u"Suisse", u"French"), | ||||||
|     (u"fr-FR", u"Français", u"France", u"French"), |     (u"fr-FR", u"Français", u"France", u"French"), | ||||||
|     (u"he-IL", u"עברית", u"", u"Hebrew"), |     (u"he-IL", u"עברית", u"", u"Hebrew"), | ||||||
|  |     (u"hr-HR", u"Hrvatski", u"", u"Croatian"), | ||||||
|     (u"hu-HU", u"Magyar", u"", u"Hungarian"), |     (u"hu-HU", u"Magyar", u"", u"Hungarian"), | ||||||
|     (u"it", u"Italiano", u"", u"Italian"), |     (u"id-ID", u"Indonesia", u"", u"Indonesian"), | ||||||
|     (u"it-CH", u"Italiano", u"Svizzera", u"Italian"), |     (u"is-IS", u"Íslenska", u"", u"Icelandic"), | ||||||
|     (u"it-IT", u"Italiano", u"Italia", u"Italian"), |     (u"it-IT", u"Italiano", u"", u"Italian"), | ||||||
|     (u"ja-JP", u"日本語", u"", u"Japanese"), |     (u"ja-JP", u"日本語", u"", u"Japanese"), | ||||||
|     (u"ko-KR", u"한국어", u"", u"Korean"), |     (u"ko-KR", u"한국어", u"", u"Korean"), | ||||||
|  |     (u"lt-LT", u"Lietuvių", u"", u"Lithuanian"), | ||||||
|  |     (u"lv-LV", u"Latviešu", u"", u"Latvian"), | ||||||
|  |     (u"ms-MY", u"Bahasa Melayu", u"", u"Malay"), | ||||||
|  |     (u"nb-NO", u"Norsk Bokmål", u"", u"Norwegian Bokmål"), | ||||||
|     (u"nl", u"Nederlands", u"", u"Dutch"), |     (u"nl", u"Nederlands", u"", u"Dutch"), | ||||||
|     (u"nl-BE", u"Nederlands", u"België", u"Dutch"), |     (u"nl-BE", u"Nederlands", u"België", u"Dutch"), | ||||||
|     (u"nl-NL", u"Nederlands", u"Nederland", u"Dutch"), |     (u"nl-NL", u"Nederlands", u"Nederland", u"Dutch"), | ||||||
|     (u"no-NO", u"Norsk", u"", u"Norwegian"), |  | ||||||
|     (u"pl-PL", u"Polski", u"", u"Polish"), |     (u"pl-PL", u"Polski", u"", u"Polish"), | ||||||
|     (u"pt", u"Português", u"", u"Portuguese"), |     (u"pt", u"Português", u"", u"Portuguese"), | ||||||
|     (u"pt-AD", u"Português", u"Andorra", u"Portuguese"), |  | ||||||
|     (u"pt-BR", u"Português", u"Brasil", u"Portuguese"), |     (u"pt-BR", u"Português", u"Brasil", u"Portuguese"), | ||||||
|     (u"pt-PT", u"Português", u"Portugal", u"Portuguese"), |     (u"pt-PT", u"Português", u"Portugal", u"Portuguese"), | ||||||
|     (u"ro-RO", u"Română", u"", u"Romanian"), |     (u"ro-RO", u"Română", u"", u"Romanian"), | ||||||
|     (u"ru-RU", u"Русский", u"", u"Russian"), |     (u"ru-RU", u"Русский", u"", u"Russian"), | ||||||
|  |     (u"sk-SK", u"Slovenčina", u"", u"Slovak"), | ||||||
|  |     (u"sl-SI", u"Slovenščina", u"", u"Slovenian"), | ||||||
|  |     (u"sr-RS", u"Српски", u"", u"Serbian"), | ||||||
|     (u"sv-SE", u"Svenska", u"", u"Swedish"), |     (u"sv-SE", u"Svenska", u"", u"Swedish"), | ||||||
|     (u"th-TH", u"ไทย", u"", u"Thai"), |     (u"th-TH", u"ไทย", u"", u"Thai"), | ||||||
|     (u"tr-TR", u"Türkçe", u"", u"Turkish"), |     (u"tr-TR", u"Türkçe", u"", u"Turkish"), | ||||||
|  |     (u"uk-UA", u"Українська", u"", u"Ukrainian"), | ||||||
|  |     (u"vi-VN", u"Tiếng Việt", u"", u"Vietnamese"), | ||||||
|     (u"zh", u"中文", u"", u"Chinese"), |     (u"zh", u"中文", u"", u"Chinese"), | ||||||
|     (u"zh-CN", u"中文", u"中国", u"Chinese"), |     (u"zh-CN", u"中文", u"中国", u"Chinese"), | ||||||
|     (u"zh-HK", u"中文", u"香港", u"Chinese"), |     (u"zh-TW", u"中文", u"台灣", u"Chinese") | ||||||
|     (u"zh-TW", u"中文", u"台湾", u"Chinese") |  | ||||||
| ) | ) | ||||||
|  | |||||||
| @ -115,10 +115,6 @@ class SearchLanguageSetting(EnumStringSetting): | |||||||
|                 pass |                 pass | ||||||
|             elif lang in self.choices: |             elif lang in self.choices: | ||||||
|                 data = lang |                 data = lang | ||||||
|             elif data == 'nb-NO': |  | ||||||
|                 data = 'no-NO' |  | ||||||
|             elif data == 'ar-XA': |  | ||||||
|                 data = 'ar-SA' |  | ||||||
|             else: |             else: | ||||||
|                 data = self.value |                 data = self.value | ||||||
|         self.value = data |         self.value = data | ||||||
|  | |||||||
| @ -96,9 +96,13 @@ class RawTextQuery(object): | |||||||
|                                 break |                                 break | ||||||
| 
 | 
 | ||||||
|                 # user may set a valid, yet not selectable language |                 # user may set a valid, yet not selectable language | ||||||
|                 if not self.languages and VALID_LANGUAGE_CODE.match(lang): |                 if VALID_LANGUAGE_CODE.match(lang): | ||||||
|                     self.languages.append(lang) |                     lang_parts = lang.split('-') | ||||||
|                     parse_next = True |                     if len(lang_parts) > 1: | ||||||
|  |                         lang = lang_parts[0].lower() + '-' + lang_parts[1].upper() | ||||||
|  |                     if lang not in self.languages: | ||||||
|  |                         self.languages.append(lang) | ||||||
|  |                         parse_next = True | ||||||
| 
 | 
 | ||||||
|             # this force a engine or category |             # this force a engine or category | ||||||
|             if query_part[0] == '!' or query_part[0] == '?': |             if query_part[0] == '!' or query_part[0] == '?': | ||||||
|  | |||||||
| @ -187,7 +187,7 @@ | |||||||
|                                     </td> |                                     </td> | ||||||
|                                     <th>{{ search_engine.name }}</th> |                                     <th>{{ search_engine.name }}</th> | ||||||
| 				    <td class="name">{{ shortcuts[search_engine.name] }}</td> | 				    <td class="name">{{ shortcuts[search_engine.name] }}</td> | ||||||
| 					<td>{{ support_toggle(current_language in search_engine.supported_languages or current_language.split('-')[0] in search_engine.supported_languages) }}</td> | 					<td>{{ support_toggle(stats[search_engine.name].supports_selected_language) }}</td> | ||||||
| 					<td>{{ support_toggle(search_engine.safesearch==True) }}</td> | 					<td>{{ support_toggle(search_engine.safesearch==True) }}</td> | ||||||
| 					<td>{{ support_toggle(search_engine.time_range_support==True) }}</td> | 					<td>{{ support_toggle(search_engine.time_range_support==True) }}</td> | ||||||
| 					<td class="{{ 'danger' if stats[search_engine.name]['warn_time'] else '' }}">{{ 'N/A' if stats[search_engine.name].time==None else stats[search_engine.name].time }}</td> | 					<td class="{{ 'danger' if stats[search_engine.name]['warn_time'] else '' }}">{{ 'N/A' if stats[search_engine.name].time==None else stats[search_engine.name].time }}</td> | ||||||
| @ -197,7 +197,7 @@ | |||||||
| 					<td class="{{ 'danger' if stats[search_engine.name]['warn_time'] else '' }}">{{ 'N/A' if stats[search_engine.name].time==None else stats[search_engine.name].time }}</td> | 					<td class="{{ 'danger' if stats[search_engine.name]['warn_time'] else '' }}">{{ 'N/A' if stats[search_engine.name].time==None else stats[search_engine.name].time }}</td> | ||||||
| 					<td>{{ support_toggle(search_engine.time_range_support==True) }}</td> | 					<td>{{ support_toggle(search_engine.time_range_support==True) }}</td> | ||||||
| 					<td>{{ support_toggle(search_engine.safesearch==True) }}</td> | 					<td>{{ support_toggle(search_engine.safesearch==True) }}</td> | ||||||
| 					<td>{{ support_toggle(current_language in search_engine.supported_languages or current_language.split('-')[0] in search_engine.supported_languages) }}</td> | 					<td>{{ support_toggle(stats[search_engine.name].supports_selected_language) }}</td> | ||||||
| 					<td>{{ shortcuts[search_engine.name] }}</td> | 					<td>{{ shortcuts[search_engine.name] }}</td> | ||||||
|                                     <th>{{ search_engine.name }}</th> |                                     <th>{{ search_engine.name }}</th> | ||||||
|                                     <td class="onoff-checkbox"> |                                     <td class="onoff-checkbox"> | ||||||
|  | |||||||
| @ -4,6 +4,7 @@ import hmac | |||||||
| import os | import os | ||||||
| import re | import re | ||||||
| 
 | 
 | ||||||
|  | from babel.core import get_global | ||||||
| from babel.dates import format_date | from babel.dates import format_date | ||||||
| from codecs import getincrementalencoder | from codecs import getincrementalencoder | ||||||
| from imp import load_source | from imp import load_source | ||||||
| @ -12,6 +13,7 @@ from os.path import splitext, join | |||||||
| from random import choice | from random import choice | ||||||
| import sys | import sys | ||||||
| 
 | 
 | ||||||
|  | from searx import settings | ||||||
| from searx.version import VERSION_STRING | from searx.version import VERSION_STRING | ||||||
| from searx.languages import language_codes | from searx.languages import language_codes | ||||||
| from searx import settings | from searx import settings | ||||||
| @ -322,6 +324,65 @@ def is_valid_lang(lang): | |||||||
|         return False |         return False | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
|  | # auxiliary function to match lang_code in lang_list | ||||||
|  | def _match_language(lang_code, lang_list=[], custom_aliases={}): | ||||||
|  |     # replace language code with a custom alias if necessary | ||||||
|  |     if lang_code in custom_aliases: | ||||||
|  |         lang_code = custom_aliases[lang_code] | ||||||
|  | 
 | ||||||
|  |     if lang_code in lang_list: | ||||||
|  |         return lang_code | ||||||
|  | 
 | ||||||
|  |     # try to get the most likely country for this language | ||||||
|  |     subtags = get_global('likely_subtags').get(lang_code) | ||||||
|  |     if subtags: | ||||||
|  |         subtag_parts = subtags.split('_') | ||||||
|  |         new_code = subtag_parts[0] + '-' + subtag_parts[-1] | ||||||
|  |         if new_code in custom_aliases: | ||||||
|  |             new_code = custom_aliases[new_code] | ||||||
|  |         if new_code in lang_list: | ||||||
|  |             return new_code | ||||||
|  | 
 | ||||||
|  |     # try to get the any supported country for this language | ||||||
|  |     for lc in lang_list: | ||||||
|  |         if lang_code == lc.split('-')[0]: | ||||||
|  |             return lc | ||||||
|  | 
 | ||||||
|  |     return None | ||||||
|  | 
 | ||||||
|  | 
 | ||||||
|  | # get the language code from lang_list that best matches locale_code | ||||||
|  | def match_language(locale_code, lang_list=[], custom_aliases={}, fallback='en-US'): | ||||||
|  |     # try to get language from given locale_code | ||||||
|  |     language = _match_language(locale_code, lang_list, custom_aliases) | ||||||
|  |     if language: | ||||||
|  |         return language | ||||||
|  | 
 | ||||||
|  |     locale_parts = locale_code.split('-') | ||||||
|  |     lang_code = locale_parts[0] | ||||||
|  | 
 | ||||||
|  |     # try to get language using an equivalent country code | ||||||
|  |     if len(locale_parts) > 1: | ||||||
|  |         country_alias = get_global('territory_aliases').get(locale_parts[-1]) | ||||||
|  |         if country_alias: | ||||||
|  |             language = _match_language(lang_code + '-' + country_alias[0], lang_list, custom_aliases) | ||||||
|  |             if language: | ||||||
|  |                 return language | ||||||
|  | 
 | ||||||
|  |     # try to get language using an equivalent language code | ||||||
|  |     alias = get_global('language_aliases').get(lang_code) | ||||||
|  |     if alias: | ||||||
|  |         language = _match_language(alias, lang_list, custom_aliases) | ||||||
|  |         if language: | ||||||
|  |             return language | ||||||
|  | 
 | ||||||
|  |     if lang_code != locale_code: | ||||||
|  |         # try to get language from given language without giving the country | ||||||
|  |         language = _match_language(lang_code, lang_list, custom_aliases) | ||||||
|  | 
 | ||||||
|  |     return language or fallback | ||||||
|  | 
 | ||||||
|  | 
 | ||||||
| def load_module(filename, module_dir): | def load_module(filename, module_dir): | ||||||
|     modname = splitext(filename)[0] |     modname = splitext(filename)[0] | ||||||
|     if modname in sys.modules: |     if modname in sys.modules: | ||||||
|  | |||||||
| @ -58,16 +58,16 @@ from searx.engines import ( | |||||||
| from searx.utils import ( | from searx.utils import ( | ||||||
|     UnicodeWriter, highlight_content, html_to_text, get_resources_directory, |     UnicodeWriter, highlight_content, html_to_text, get_resources_directory, | ||||||
|     get_static_files, get_result_templates, get_themes, gen_useragent, |     get_static_files, get_result_templates, get_themes, gen_useragent, | ||||||
|     dict_subset, prettify_url |     dict_subset, prettify_url, match_language | ||||||
| ) | ) | ||||||
| from searx.version import VERSION_STRING | from searx.version import VERSION_STRING | ||||||
| from searx.languages import language_codes | from searx.languages import language_codes as languages | ||||||
| from searx.search import SearchWithPlugins, get_search_query_from_webapp | from searx.search import SearchWithPlugins, get_search_query_from_webapp | ||||||
| from searx.query import RawTextQuery | from searx.query import RawTextQuery | ||||||
| from searx.autocomplete import searx_bang, backends as autocomplete_backends | from searx.autocomplete import searx_bang, backends as autocomplete_backends | ||||||
| from searx.plugins import plugins | from searx.plugins import plugins | ||||||
| from searx.plugins.oa_doi_rewrite import get_doi_resolver | from searx.plugins.oa_doi_rewrite import get_doi_resolver | ||||||
| from searx.preferences import Preferences, ValidationException | from searx.preferences import Preferences, ValidationException, LANGUAGE_CODES | ||||||
| from searx.answerers import answerers | from searx.answerers import answerers | ||||||
| from searx.url_utils import urlencode, urlparse, urljoin | from searx.url_utils import urlencode, urlparse, urljoin | ||||||
| from searx.utils import new_hmac | from searx.utils import new_hmac | ||||||
| @ -133,7 +133,7 @@ if not searx_debug \ | |||||||
| babel = Babel(app) | babel = Babel(app) | ||||||
| 
 | 
 | ||||||
| rtl_locales = ['ar', 'arc', 'bcc', 'bqi', 'ckb', 'dv', 'fa', 'glk', 'he', | rtl_locales = ['ar', 'arc', 'bcc', 'bqi', 'ckb', 'dv', 'fa', 'glk', 'he', | ||||||
|                'ku', 'mzn', 'pnb'', ''ps', 'sd', 'ug', 'ur', 'yi'] |                'ku', 'mzn', 'pnb', 'ps', 'sd', 'ug', 'ur', 'yi'] | ||||||
| 
 | 
 | ||||||
| # used when translating category names | # used when translating category names | ||||||
| _category_names = (gettext('files'), | _category_names = (gettext('files'), | ||||||
| @ -352,9 +352,11 @@ def render(template_name, override_theme=None, **kwargs): | |||||||
| 
 | 
 | ||||||
|     kwargs['safesearch'] = str(request.preferences.get_value('safesearch')) |     kwargs['safesearch'] = str(request.preferences.get_value('safesearch')) | ||||||
| 
 | 
 | ||||||
|     kwargs['language_codes'] = language_codes |     kwargs['language_codes'] = languages | ||||||
|     if 'current_language' not in kwargs: |     if 'current_language' not in kwargs: | ||||||
|         kwargs['current_language'] = request.preferences.get_value('language') |         kwargs['current_language'] = match_language(request.preferences.get_value('language'), | ||||||
|  |                                                     LANGUAGE_CODES, | ||||||
|  |                                                     fallback=settings['search']['language']) | ||||||
| 
 | 
 | ||||||
|     # override url_for function in templates |     # override url_for function in templates | ||||||
|     kwargs['url_for'] = url_for_theme |     kwargs['url_for'] = url_for_theme | ||||||
| @ -590,7 +592,9 @@ def index(): | |||||||
|         infoboxes=result_container.infoboxes, |         infoboxes=result_container.infoboxes, | ||||||
|         paging=result_container.paging, |         paging=result_container.paging, | ||||||
|         unresponsive_engines=result_container.unresponsive_engines, |         unresponsive_engines=result_container.unresponsive_engines, | ||||||
|         current_language=search_query.lang, |         current_language=match_language(search_query.lang, | ||||||
|  |                                         LANGUAGE_CODES, | ||||||
|  |                                         fallback=settings['search']['language']), | ||||||
|         base_url=get_base_url(), |         base_url=get_base_url(), | ||||||
|         theme=get_current_theme_name(), |         theme=get_current_theme_name(), | ||||||
|         favicons=global_favicons[themes.index(get_current_theme_name())] |         favicons=global_favicons[themes.index(get_current_theme_name())] | ||||||
| @ -687,6 +691,10 @@ def preferences(): | |||||||
|                              'warn_time': False} |                              'warn_time': False} | ||||||
|             if e.timeout > settings['outgoing']['request_timeout']: |             if e.timeout > settings['outgoing']['request_timeout']: | ||||||
|                 stats[e.name]['warn_timeout'] = True |                 stats[e.name]['warn_timeout'] = True | ||||||
|  |             if match_language(request.preferences.get_value('language'), | ||||||
|  |                               getattr(e, 'supported_languages', []), | ||||||
|  |                               getattr(e, 'language_aliases', {}), None): | ||||||
|  |                 stats[e.name]['supports_selected_language'] = True | ||||||
| 
 | 
 | ||||||
|     # get first element [0], the engine time, |     # get first element [0], the engine time, | ||||||
|     # and then the second element [1] : the time (the first one is the label) |     # and then the second element [1] : the time (the first one is the label) | ||||||
|  | |||||||
| @ -19,12 +19,17 @@ class TestArchLinuxEngine(SearxTestCase): | |||||||
|         query = 'test_query' |         query = 'test_query' | ||||||
|         dic = defaultdict(dict) |         dic = defaultdict(dict) | ||||||
|         dic['pageno'] = 1 |         dic['pageno'] = 1 | ||||||
|         dic['language'] = 'en_US' |         dic['language'] = 'en-US' | ||||||
|         params = archlinux.request(query, dic) |         params = archlinux.request(query, dic) | ||||||
|         self.assertTrue('url' in params) |         self.assertTrue('url' in params) | ||||||
|         self.assertTrue(query in params['url']) |         self.assertTrue(query in params['url']) | ||||||
|         self.assertTrue('wiki.archlinux.org' in params['url']) |         self.assertTrue('wiki.archlinux.org' in params['url']) | ||||||
| 
 | 
 | ||||||
|  |         for lang, name in archlinux.main_langs: | ||||||
|  |             dic['language'] = lang | ||||||
|  |             params = archlinux.request(query, dic) | ||||||
|  |             self.assertTrue(name in params['url']) | ||||||
|  | 
 | ||||||
|         for lang, domain in domains.items(): |         for lang, domain in domains.items(): | ||||||
|             dic['language'] = lang |             dic['language'] = lang | ||||||
|             params = archlinux.request(query, dic) |             params = archlinux.request(query, dic) | ||||||
|  | |||||||
| @ -7,6 +7,7 @@ from searx.testing import SearxTestCase | |||||||
| class TestBingEngine(SearxTestCase): | class TestBingEngine(SearxTestCase): | ||||||
| 
 | 
 | ||||||
|     def test_request(self): |     def test_request(self): | ||||||
|  |         bing.supported_languages = ['en', 'fr', 'zh-CHS', 'zh-CHT', 'pt-PT', 'pt-BR'] | ||||||
|         query = u'test_query' |         query = u'test_query' | ||||||
|         dicto = defaultdict(dict) |         dicto = defaultdict(dict) | ||||||
|         dicto['pageno'] = 0 |         dicto['pageno'] = 0 | ||||||
|  | |||||||
| @ -9,7 +9,6 @@ class TestBingImagesEngine(SearxTestCase): | |||||||
| 
 | 
 | ||||||
|     def test_request(self): |     def test_request(self): | ||||||
|         bing_images.supported_languages = ['fr-FR', 'en-US'] |         bing_images.supported_languages = ['fr-FR', 'en-US'] | ||||||
| 
 |  | ||||||
|         query = 'test_query' |         query = 'test_query' | ||||||
|         dicto = defaultdict(dict) |         dicto = defaultdict(dict) | ||||||
|         dicto['pageno'] = 1 |         dicto['pageno'] = 1 | ||||||
|  | |||||||
| @ -8,10 +8,11 @@ import lxml | |||||||
| class TestBingNewsEngine(SearxTestCase): | class TestBingNewsEngine(SearxTestCase): | ||||||
| 
 | 
 | ||||||
|     def test_request(self): |     def test_request(self): | ||||||
|  |         bing_news.supported_languages = ['en', 'fr'] | ||||||
|         query = 'test_query' |         query = 'test_query' | ||||||
|         dicto = defaultdict(dict) |         dicto = defaultdict(dict) | ||||||
|         dicto['pageno'] = 1 |         dicto['pageno'] = 1 | ||||||
|         dicto['language'] = 'fr_FR' |         dicto['language'] = 'fr-FR' | ||||||
|         dicto['time_range'] = '' |         dicto['time_range'] = '' | ||||||
|         params = bing_news.request(query, dicto) |         params = bing_news.request(query, dicto) | ||||||
|         self.assertIn('url', params) |         self.assertIn('url', params) | ||||||
|  | |||||||
| @ -9,7 +9,6 @@ class TestBingVideosEngine(SearxTestCase): | |||||||
| 
 | 
 | ||||||
|     def test_request(self): |     def test_request(self): | ||||||
|         bing_videos.supported_languages = ['fr-FR', 'en-US'] |         bing_videos.supported_languages = ['fr-FR', 'en-US'] | ||||||
| 
 |  | ||||||
|         query = 'test_query' |         query = 'test_query' | ||||||
|         dicto = defaultdict(dict) |         dicto = defaultdict(dict) | ||||||
|         dicto['pageno'] = 1 |         dicto['pageno'] = 1 | ||||||
|  | |||||||
| @ -8,10 +8,11 @@ from searx.testing import SearxTestCase | |||||||
| class TestDailymotionEngine(SearxTestCase): | class TestDailymotionEngine(SearxTestCase): | ||||||
| 
 | 
 | ||||||
|     def test_request(self): |     def test_request(self): | ||||||
|  |         dailymotion.supported_languages = ['en', 'fr'] | ||||||
|         query = 'test_query' |         query = 'test_query' | ||||||
|         dicto = defaultdict(dict) |         dicto = defaultdict(dict) | ||||||
|         dicto['pageno'] = 0 |         dicto['pageno'] = 0 | ||||||
|         dicto['language'] = 'fr_FR' |         dicto['language'] = 'fr-FR' | ||||||
|         params = dailymotion.request(query, dicto) |         params = dailymotion.request(query, dicto) | ||||||
|         self.assertTrue('url' in params) |         self.assertTrue('url' in params) | ||||||
|         self.assertTrue(query in params['url']) |         self.assertTrue(query in params['url']) | ||||||
|  | |||||||
| @ -1,18 +1,21 @@ | |||||||
| # -*- coding: utf-8 -*- | # -*- coding: utf-8 -*- | ||||||
| from collections import defaultdict | from collections import defaultdict | ||||||
| import mock | import mock | ||||||
| from searx.engines import duckduckgo | from searx.engines import load_engine, duckduckgo | ||||||
| from searx.testing import SearxTestCase | from searx.testing import SearxTestCase | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
| class TestDuckduckgoEngine(SearxTestCase): | class TestDuckduckgoEngine(SearxTestCase): | ||||||
| 
 | 
 | ||||||
|     def test_request(self): |     def test_request(self): | ||||||
|  |         duckduckgo = load_engine({'engine': 'duckduckgo', 'name': 'duckduckgo'}) | ||||||
|  | 
 | ||||||
|         query = 'test_query' |         query = 'test_query' | ||||||
|         dicto = defaultdict(dict) |         dicto = defaultdict(dict) | ||||||
|         dicto['pageno'] = 1 |         dicto['pageno'] = 1 | ||||||
|         dicto['language'] = 'de-CH' |  | ||||||
|         dicto['time_range'] = '' |         dicto['time_range'] = '' | ||||||
|  | 
 | ||||||
|  |         dicto['language'] = 'de-CH' | ||||||
|         params = duckduckgo.request(query, dicto) |         params = duckduckgo.request(query, dicto) | ||||||
|         self.assertIn('url', params) |         self.assertIn('url', params) | ||||||
|         self.assertIn(query, params['url']) |         self.assertIn(query, params['url']) | ||||||
| @ -20,16 +23,19 @@ class TestDuckduckgoEngine(SearxTestCase): | |||||||
|         self.assertIn('ch-de', params['url']) |         self.assertIn('ch-de', params['url']) | ||||||
|         self.assertIn('s=0', params['url']) |         self.assertIn('s=0', params['url']) | ||||||
| 
 | 
 | ||||||
|         # when ddg uses non standard code |         # when ddg uses non standard codes | ||||||
|  |         dicto['language'] = 'zh-HK' | ||||||
|  |         params = duckduckgo.request(query, dicto) | ||||||
|  |         self.assertIn('hk-tzh', params['url']) | ||||||
|  | 
 | ||||||
|         dicto['language'] = 'en-GB' |         dicto['language'] = 'en-GB' | ||||||
|         params = duckduckgo.request(query, dicto) |         params = duckduckgo.request(query, dicto) | ||||||
|         self.assertIn('uk-en', params['url']) |         self.assertIn('uk-en', params['url']) | ||||||
| 
 | 
 | ||||||
|         # no country given |         # no country given | ||||||
|         duckduckgo.supported_languages = ['de-CH', 'en-US'] |         dicto['language'] = 'en' | ||||||
|         dicto['language'] = 'de' |  | ||||||
|         params = duckduckgo.request(query, dicto) |         params = duckduckgo.request(query, dicto) | ||||||
|         self.assertIn('ch-de', params['url']) |         self.assertIn('us-en', params['url']) | ||||||
| 
 | 
 | ||||||
|     def test_no_url_in_request_year_time_range(self): |     def test_no_url_in_request_year_time_range(self): | ||||||
|         dicto = defaultdict(dict) |         dicto = defaultdict(dict) | ||||||
|  | |||||||
| @ -18,6 +18,7 @@ class TestDDGDefinitionsEngine(SearxTestCase): | |||||||
|         self.assertEqual(result, 'Text in link') |         self.assertEqual(result, 'Text in link') | ||||||
| 
 | 
 | ||||||
|     def test_request(self): |     def test_request(self): | ||||||
|  |         duckduckgo_definitions.supported_languages = ['en-US', 'es-ES'] | ||||||
|         query = 'test_query' |         query = 'test_query' | ||||||
|         dicto = defaultdict(dict) |         dicto = defaultdict(dict) | ||||||
|         dicto['pageno'] = 1 |         dicto['pageno'] = 1 | ||||||
|  | |||||||
| @ -9,7 +9,6 @@ class TestDuckduckgoImagesEngine(SearxTestCase): | |||||||
| 
 | 
 | ||||||
|     def test_request(self): |     def test_request(self): | ||||||
|         duckduckgo_images.supported_languages = ['de-CH', 'en-US'] |         duckduckgo_images.supported_languages = ['de-CH', 'en-US'] | ||||||
| 
 |  | ||||||
|         query = 'test_query' |         query = 'test_query' | ||||||
|         dicto = defaultdict(dict) |         dicto = defaultdict(dict) | ||||||
|         dicto['is_test'] = True |         dicto['is_test'] = True | ||||||
|  | |||||||
| @ -15,6 +15,8 @@ class TestGoogleEngine(SearxTestCase): | |||||||
|         return response |         return response | ||||||
| 
 | 
 | ||||||
|     def test_request(self): |     def test_request(self): | ||||||
|  |         google.supported_languages = ['en', 'fr', 'zh-CN'] | ||||||
|  | 
 | ||||||
|         query = 'test_query' |         query = 'test_query' | ||||||
|         dicto = defaultdict(dict) |         dicto = defaultdict(dict) | ||||||
|         dicto['pageno'] = 1 |         dicto['pageno'] = 1 | ||||||
| @ -31,6 +33,11 @@ class TestGoogleEngine(SearxTestCase): | |||||||
|         self.assertIn('google.co', params['url']) |         self.assertIn('google.co', params['url']) | ||||||
|         self.assertIn('en', params['headers']['Accept-Language']) |         self.assertIn('en', params['headers']['Accept-Language']) | ||||||
| 
 | 
 | ||||||
|  |         dicto['language'] = 'zh' | ||||||
|  |         params = google.request(query, dicto) | ||||||
|  |         self.assertIn('google.com', params['url']) | ||||||
|  |         self.assertIn('zh-CN', params['headers']['Accept-Language']) | ||||||
|  | 
 | ||||||
|     def test_response(self): |     def test_response(self): | ||||||
|         self.assertRaises(AttributeError, google.response, None) |         self.assertRaises(AttributeError, google.response, None) | ||||||
|         self.assertRaises(AttributeError, google.response, []) |         self.assertRaises(AttributeError, google.response, []) | ||||||
|  | |||||||
| @ -9,6 +9,7 @@ from searx.testing import SearxTestCase | |||||||
| class TestGoogleNewsEngine(SearxTestCase): | class TestGoogleNewsEngine(SearxTestCase): | ||||||
| 
 | 
 | ||||||
|     def test_request(self): |     def test_request(self): | ||||||
|  |         google_news.supported_languages = ['en-US', 'fr-FR'] | ||||||
|         query = 'test_query' |         query = 'test_query' | ||||||
|         dicto = defaultdict(dict) |         dicto = defaultdict(dict) | ||||||
|         dicto['pageno'] = 1 |         dicto['pageno'] = 1 | ||||||
|  | |||||||
| @ -7,6 +7,7 @@ from searx.testing import SearxTestCase | |||||||
| class TestQwantEngine(SearxTestCase): | class TestQwantEngine(SearxTestCase): | ||||||
| 
 | 
 | ||||||
|     def test_request(self): |     def test_request(self): | ||||||
|  |         qwant.supported_languages = ['en-US', 'fr-CA', 'fr-FR'] | ||||||
|         query = 'test_query' |         query = 'test_query' | ||||||
|         dicto = defaultdict(dict) |         dicto = defaultdict(dict) | ||||||
|         dicto['pageno'] = 0 |         dicto['pageno'] = 0 | ||||||
| @ -26,7 +27,6 @@ class TestQwantEngine(SearxTestCase): | |||||||
|         self.assertIn('en_us', params['url']) |         self.assertIn('en_us', params['url']) | ||||||
|         self.assertIn('news', params['url']) |         self.assertIn('news', params['url']) | ||||||
| 
 | 
 | ||||||
|         qwant.supported_languages = ['en', 'fr-FR', 'fr-CA'] |  | ||||||
|         dicto['language'] = 'fr' |         dicto['language'] = 'fr' | ||||||
|         params = qwant.request(query, dicto) |         params = qwant.request(query, dicto) | ||||||
|         self.assertIn('fr_fr', params['url']) |         self.assertIn('fr_fr', params['url']) | ||||||
|  | |||||||
| @ -7,6 +7,7 @@ from searx.testing import SearxTestCase | |||||||
| class TestSwisscowsEngine(SearxTestCase): | class TestSwisscowsEngine(SearxTestCase): | ||||||
| 
 | 
 | ||||||
|     def test_request(self): |     def test_request(self): | ||||||
|  |         swisscows.supported_languages = ['de-AT', 'de-DE'] | ||||||
|         query = 'test_query' |         query = 'test_query' | ||||||
|         dicto = defaultdict(dict) |         dicto = defaultdict(dict) | ||||||
|         dicto['pageno'] = 1 |         dicto['pageno'] = 1 | ||||||
|  | |||||||
| @ -9,6 +9,7 @@ from searx.testing import SearxTestCase | |||||||
| class TestWikidataEngine(SearxTestCase): | class TestWikidataEngine(SearxTestCase): | ||||||
| 
 | 
 | ||||||
|     def test_request(self): |     def test_request(self): | ||||||
|  |         wikidata.supported_languages = ['en', 'es'] | ||||||
|         query = 'test_query' |         query = 'test_query' | ||||||
|         dicto = defaultdict(dict) |         dicto = defaultdict(dict) | ||||||
|         dicto['language'] = 'en-US' |         dicto['language'] = 'en-US' | ||||||
|  | |||||||
| @ -25,11 +25,12 @@ class TestYahooEngine(SearxTestCase): | |||||||
|         self.assertEqual('https://this.is.the.url/', url) |         self.assertEqual('https://this.is.the.url/', url) | ||||||
| 
 | 
 | ||||||
|     def test_request(self): |     def test_request(self): | ||||||
|  |         yahoo.supported_languages = ['en', 'fr', 'zh-CHT', 'zh-CHS'] | ||||||
|         query = 'test_query' |         query = 'test_query' | ||||||
|         dicto = defaultdict(dict) |         dicto = defaultdict(dict) | ||||||
|         dicto['pageno'] = 1 |         dicto['pageno'] = 1 | ||||||
|         dicto['time_range'] = '' |         dicto['time_range'] = '' | ||||||
|         dicto['language'] = 'fr_FR' |         dicto['language'] = 'fr-FR' | ||||||
|         params = yahoo.request(query, dicto) |         params = yahoo.request(query, dicto) | ||||||
|         self.assertIn('url', params) |         self.assertIn('url', params) | ||||||
|         self.assertIn(query, params['url']) |         self.assertIn(query, params['url']) | ||||||
| @ -39,6 +40,16 @@ class TestYahooEngine(SearxTestCase): | |||||||
|         self.assertIn('sB', params['cookies']) |         self.assertIn('sB', params['cookies']) | ||||||
|         self.assertIn('fr', params['cookies']['sB']) |         self.assertIn('fr', params['cookies']['sB']) | ||||||
| 
 | 
 | ||||||
|  |         dicto['language'] = 'zh' | ||||||
|  |         params = yahoo.request(query, dicto) | ||||||
|  |         self.assertIn('zh_chs', params['url']) | ||||||
|  |         self.assertIn('zh_chs', params['cookies']['sB']) | ||||||
|  | 
 | ||||||
|  |         dicto['language'] = 'zh-TW' | ||||||
|  |         params = yahoo.request(query, dicto) | ||||||
|  |         self.assertIn('zh_cht', params['url']) | ||||||
|  |         self.assertIn('zh_cht', params['cookies']['sB']) | ||||||
|  | 
 | ||||||
|     def test_no_url_in_request_year_time_range(self): |     def test_no_url_in_request_year_time_range(self): | ||||||
|         dicto = defaultdict(dict) |         dicto = defaultdict(dict) | ||||||
|         query = 'test_query' |         query = 'test_query' | ||||||
| @ -168,5 +179,5 @@ class TestYahooEngine(SearxTestCase): | |||||||
|         self.assertEqual(type(languages), list) |         self.assertEqual(type(languages), list) | ||||||
|         self.assertEqual(len(languages), 3) |         self.assertEqual(len(languages), 3) | ||||||
|         self.assertIn('ar', languages) |         self.assertIn('ar', languages) | ||||||
|         self.assertIn('zh-chs', languages) |         self.assertIn('zh-CHS', languages) | ||||||
|         self.assertIn('zh-cht', languages) |         self.assertIn('zh-CHT', languages) | ||||||
|  | |||||||
| @ -9,10 +9,11 @@ from searx.testing import SearxTestCase | |||||||
| class TestYahooNewsEngine(SearxTestCase): | class TestYahooNewsEngine(SearxTestCase): | ||||||
| 
 | 
 | ||||||
|     def test_request(self): |     def test_request(self): | ||||||
|  |         yahoo_news.supported_languages = ['en', 'fr'] | ||||||
|         query = 'test_query' |         query = 'test_query' | ||||||
|         dicto = defaultdict(dict) |         dicto = defaultdict(dict) | ||||||
|         dicto['pageno'] = 1 |         dicto['pageno'] = 1 | ||||||
|         dicto['language'] = 'fr_FR' |         dicto['language'] = 'fr-FR' | ||||||
|         params = yahoo_news.request(query, dicto) |         params = yahoo_news.request(query, dicto) | ||||||
|         self.assertIn('url', params) |         self.assertIn('url', params) | ||||||
|         self.assertIn(query, params['url']) |         self.assertIn(query, params['url']) | ||||||
|  | |||||||
| @ -65,6 +65,31 @@ class TestUtils(SearxTestCase): | |||||||
|         for test_url, expected in data: |         for test_url, expected in data: | ||||||
|             self.assertEqual(utils.prettify_url(test_url, max_length=32), expected) |             self.assertEqual(utils.prettify_url(test_url, max_length=32), expected) | ||||||
| 
 | 
 | ||||||
|  |     def test_match_language(self): | ||||||
|  |         self.assertEqual(utils.match_language('es', ['es']), 'es') | ||||||
|  |         self.assertEqual(utils.match_language('es', [], fallback='fallback'), 'fallback') | ||||||
|  |         self.assertEqual(utils.match_language('ja', ['jp'], {'ja': 'jp'}), 'jp') | ||||||
|  | 
 | ||||||
|  |         aliases = {'en-GB': 'en-UK', 'he': 'iw'} | ||||||
|  | 
 | ||||||
|  |         # guess country | ||||||
|  |         self.assertEqual(utils.match_language('de-DE', ['de']), 'de') | ||||||
|  |         self.assertEqual(utils.match_language('de', ['de-DE']), 'de-DE') | ||||||
|  |         self.assertEqual(utils.match_language('es-CO', ['es-AR', 'es-ES', 'es-MX']), 'es-ES') | ||||||
|  |         self.assertEqual(utils.match_language('es-CO', ['es-MX']), 'es-MX') | ||||||
|  |         self.assertEqual(utils.match_language('en-UK', ['en-AU', 'en-GB', 'en-US']), 'en-GB') | ||||||
|  |         self.assertEqual(utils.match_language('en-GB', ['en-AU', 'en-UK', 'en-US'], aliases), 'en-UK') | ||||||
|  | 
 | ||||||
|  |         # language aliases | ||||||
|  |         self.assertEqual(utils.match_language('iw', ['he']), 'he') | ||||||
|  |         self.assertEqual(utils.match_language('he', ['iw'], aliases), 'iw') | ||||||
|  |         self.assertEqual(utils.match_language('iw-IL', ['he']), 'he') | ||||||
|  |         self.assertEqual(utils.match_language('he-IL', ['iw'], aliases), 'iw') | ||||||
|  |         self.assertEqual(utils.match_language('iw', ['he-IL']), 'he-IL') | ||||||
|  |         self.assertEqual(utils.match_language('he', ['iw-IL'], aliases), 'iw-IL') | ||||||
|  |         self.assertEqual(utils.match_language('iw-IL', ['he-IL']), 'he-IL') | ||||||
|  |         self.assertEqual(utils.match_language('he-IL', ['iw-IL'], aliases), 'iw-IL') | ||||||
|  | 
 | ||||||
| 
 | 
 | ||||||
| class TestHTMLTextExtractor(SearxTestCase): | class TestHTMLTextExtractor(SearxTestCase): | ||||||
| 
 | 
 | ||||||
|  | |||||||
| @ -2,83 +2,27 @@ | |||||||
| 
 | 
 | ||||||
| # This script generates languages.py from intersecting each engine's supported languages. | # This script generates languages.py from intersecting each engine's supported languages. | ||||||
| # | # | ||||||
| # The country names are obtained from http://api.geonames.org which requires registering as a user. |  | ||||||
| # |  | ||||||
| # Output files (engines_languages.json and languages.py) | # Output files (engines_languages.json and languages.py) | ||||||
| # are written in current directory to avoid overwriting in case something goes wrong. | # are written in current directory to avoid overwriting in case something goes wrong. | ||||||
| 
 | 
 | ||||||
| from requests import get | from json import dump | ||||||
| from lxml.html import fromstring |  | ||||||
| from json import loads, dump |  | ||||||
| import io | import io | ||||||
| from sys import path | from sys import path | ||||||
|  | from babel import Locale, UnknownLocaleError | ||||||
|  | from babel.languages import get_global | ||||||
|  | 
 | ||||||
| path.append('../searx')  # noqa | path.append('../searx')  # noqa | ||||||
| from searx import settings | from searx import settings | ||||||
| from searx.url_utils import urlencode |  | ||||||
| from searx.engines import initialize_engines, engines | from searx.engines import initialize_engines, engines | ||||||
| 
 | 
 | ||||||
| # Geonames API for country names. |  | ||||||
| geonames_user = ''  # ADD USER NAME HERE |  | ||||||
| country_names_url = 'http://api.geonames.org/countryInfoJSON?{parameters}' |  | ||||||
| 
 |  | ||||||
| # Output files. | # Output files. | ||||||
| engines_languages_file = 'engines_languages.json' | engines_languages_file = 'engines_languages.json' | ||||||
| languages_file = 'languages.py' | languages_file = 'languages.py' | ||||||
| 
 | 
 | ||||||
| engines_languages = {} |  | ||||||
| 
 |  | ||||||
| 
 |  | ||||||
| # To filter out invalid codes and dialects. |  | ||||||
| def valid_code(lang_code): |  | ||||||
|     # filter invalid codes |  | ||||||
|     # sl-SL is technically not invalid, but still a mistake |  | ||||||
|     invalid_codes = ['sl-SL', 'wt-WT', 'jw'] |  | ||||||
|     invalid_countries = ['UK', 'XA', 'XL'] |  | ||||||
|     if lang_code[:2] == 'xx'\ |  | ||||||
|        or lang_code in invalid_codes\ |  | ||||||
|        or lang_code[-2:] in invalid_countries\ |  | ||||||
|        or is_dialect(lang_code): |  | ||||||
|         return False |  | ||||||
| 
 |  | ||||||
|     return True |  | ||||||
| 
 |  | ||||||
| 
 |  | ||||||
| # Language codes with any additional tags other than language and country. |  | ||||||
| def is_dialect(lang_code): |  | ||||||
|     lang_code = lang_code.split('-') |  | ||||||
|     if len(lang_code) > 2 or len(lang_code[0]) > 3: |  | ||||||
|         return True |  | ||||||
|     if len(lang_code) == 2 and len(lang_code[1]) > 2: |  | ||||||
|         return True |  | ||||||
| 
 |  | ||||||
|     return False |  | ||||||
| 
 |  | ||||||
| 
 |  | ||||||
| # Get country name in specified language. |  | ||||||
| def get_country_name(locale): |  | ||||||
|     if geonames_user is '': |  | ||||||
|         return '' |  | ||||||
| 
 |  | ||||||
|     locale = locale.split('-') |  | ||||||
|     if len(locale) != 2: |  | ||||||
|         return '' |  | ||||||
| 
 |  | ||||||
|     url = country_names_url.format(parameters=urlencode({'lang': locale[0], |  | ||||||
|                                                          'country': locale[1], |  | ||||||
|                                                          'username': geonames_user})) |  | ||||||
|     response = get(url) |  | ||||||
|     json = loads(response.text) |  | ||||||
|     content = json.get('geonames', None) |  | ||||||
|     if content is None or len(content) != 1: |  | ||||||
|         print("No country name found for " + locale[0] + "-" + locale[1]) |  | ||||||
|         return '' |  | ||||||
| 
 |  | ||||||
|     return content[0].get('countryName', '') |  | ||||||
| 
 |  | ||||||
| 
 | 
 | ||||||
| # Fetchs supported languages for each engine and writes json file with those. | # Fetchs supported languages for each engine and writes json file with those. | ||||||
| def fetch_supported_languages(): | def fetch_supported_languages(): | ||||||
|     initialize_engines(settings['engines']) |     engines_languages = {} | ||||||
|     for engine_name in engines: |     for engine_name in engines: | ||||||
|         if hasattr(engines[engine_name], 'fetch_supported_languages'): |         if hasattr(engines[engine_name], 'fetch_supported_languages'): | ||||||
|             try: |             try: | ||||||
| @ -90,81 +34,135 @@ def fetch_supported_languages(): | |||||||
|     with io.open(engines_languages_file, "w", encoding="utf-8") as f: |     with io.open(engines_languages_file, "w", encoding="utf-8") as f: | ||||||
|         dump(engines_languages, f, ensure_ascii=False) |         dump(engines_languages, f, ensure_ascii=False) | ||||||
| 
 | 
 | ||||||
|  |     return engines_languages | ||||||
|  | 
 | ||||||
|  | 
 | ||||||
|  | # Get babel Locale object from lang_code if possible. | ||||||
|  | def get_locale(lang_code): | ||||||
|  |     try: | ||||||
|  |         locale = Locale.parse(lang_code, sep='-') | ||||||
|  |         return locale | ||||||
|  |     except (UnknownLocaleError, ValueError): | ||||||
|  |         return None | ||||||
|  | 
 | ||||||
|  | 
 | ||||||
|  | # Append engine_name to list of engines that support locale. | ||||||
|  | def add_engine_counter(lang_code, engine_name, languages): | ||||||
|  |     if lang_code in languages: | ||||||
|  |         if 'counter' not in languages[lang_code]: | ||||||
|  |             languages[lang_code]['counter'] = [engine_name] | ||||||
|  |         elif engine_name not in languages[lang_code]['counter']: | ||||||
|  |             languages[lang_code]['counter'].append(engine_name) | ||||||
|  | 
 | ||||||
| 
 | 
 | ||||||
| # Join all language lists. | # Join all language lists. | ||||||
| # Iterate all languages supported by each engine. | # TODO: Add language names from engine's language list if name not known by babel. | ||||||
| def join_language_lists(): | def join_language_lists(engines_languages): | ||||||
|     global languages |     language_list = {} | ||||||
|     # include wikipedia first for more accurate language names |  | ||||||
|     languages = {code: lang for code, lang |  | ||||||
|                  in engines_languages['wikipedia'].items() |  | ||||||
|                  if valid_code(code)} |  | ||||||
| 
 |  | ||||||
|     for engine_name in engines_languages: |     for engine_name in engines_languages: | ||||||
|         for locale in engines_languages[engine_name]: |         for lang_code in engines_languages[engine_name]: | ||||||
|             if valid_code(locale): |  | ||||||
|                 # if language is not on list or if it has no name yet |  | ||||||
|                 if locale not in languages or not languages[locale].get('name'): |  | ||||||
|                     if isinstance(engines_languages[engine_name], dict): |  | ||||||
|                         languages[locale] = engines_languages[engine_name][locale] |  | ||||||
|                     else: |  | ||||||
|                         languages[locale] = {} |  | ||||||
| 
 | 
 | ||||||
|             # add to counter of engines that support given language |             # apply custom fixes if necessary | ||||||
|             lang = locale.split('-')[0] |             if lang_code in getattr(engines[engine_name], 'language_aliases', {}).values(): | ||||||
|             if lang in languages: |                 lang_code = next(lc for lc, alias in engines[engine_name].language_aliases.items() | ||||||
|                 if 'counter' not in languages[lang]: |                                  if lang_code == alias) | ||||||
|                     languages[lang]['counter'] = [engine_name] |  | ||||||
|                 elif engine_name not in languages[lang]['counter']: |  | ||||||
|                     languages[lang]['counter'].append(engine_name) |  | ||||||
| 
 | 
 | ||||||
|     # filter list to include only languages supported by most engines |             locale = get_locale(lang_code) | ||||||
|     min_supported_engines = int(0.70 * len(engines_languages)) |  | ||||||
|     languages = {code: lang for code, lang |  | ||||||
|                  in languages.items() |  | ||||||
|                  if len(lang.get('counter', [])) >= min_supported_engines or |  | ||||||
|                  len(languages.get(code.split('-')[0], {}).get('counter', [])) >= min_supported_engines} |  | ||||||
| 
 | 
 | ||||||
|     # get locales that have no name or country yet |             # ensure that lang_code uses standard language and country codes | ||||||
|     for locale in languages.keys(): |             if locale and locale.territory: | ||||||
|         # try to get language names |                 lang_code = locale.language + '-' + locale.territory | ||||||
|         if not languages[locale].get('name'): |  | ||||||
|             name = languages.get(locale.split('-')[0], {}).get('name', None) |  | ||||||
|             if name: |  | ||||||
|                 languages[locale]['name'] = name |  | ||||||
|             else: |  | ||||||
|                 # filter out locales with no name |  | ||||||
|                 del languages[locale] |  | ||||||
|                 continue |  | ||||||
| 
 | 
 | ||||||
|         # try to get language name in english |             # add locale if it's not in list | ||||||
|         if not languages[locale].get('english_name'): |             if lang_code not in language_list: | ||||||
|             languages[locale]['english_name'] = languages.get(locale.split('-')[0], {}).get('english_name', '') |                 if locale: | ||||||
|  |                     language_list[lang_code] = {'name': locale.get_language_name().title(), | ||||||
|  |                                                 'english_name': locale.english_name, | ||||||
|  |                                                 'country': locale.get_territory_name() or ''} | ||||||
| 
 | 
 | ||||||
|         # try to get country name |                     # also add language without country | ||||||
|         if locale.find('-') > 0 and not languages[locale].get('country'): |                     if locale.language not in language_list: | ||||||
|             languages[locale]['country'] = get_country_name(locale) or '' |                         language_list[locale.language] = {'name': locale.get_language_name().title(), | ||||||
|  |                                                           'english_name': locale.english_name} | ||||||
|  |                 else: | ||||||
|  |                     language_list[lang_code] = {} | ||||||
|  | 
 | ||||||
|  |             # count engine for both language_country combination and language alone | ||||||
|  |             add_engine_counter(lang_code, engine_name, language_list) | ||||||
|  |             add_engine_counter(lang_code.split('-')[0], engine_name, language_list) | ||||||
|  | 
 | ||||||
|  |     return language_list | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
| # Remove countryless language if language is featured in only one country. | # Filter language list so it only includes the most supported languages and countries. | ||||||
| def filter_single_country_languages(): | def filter_language_list(all_languages): | ||||||
|     prev_lang = None |     min_supported_engines = 10 | ||||||
|     prev_code = None |     main_engines = [engine_name for engine_name in engines.keys() | ||||||
|     for code in sorted(languages): |                     if 'general' in engines[engine_name].categories and | ||||||
|         lang = code.split('-')[0] |                        engines[engine_name].supported_languages and | ||||||
|         if lang == prev_lang: |                        not engines[engine_name].disabled] | ||||||
|  | 
 | ||||||
|  |     # filter list to include only languages supported by most engines or all default general engines | ||||||
|  |     filtered_languages = {code: lang for code, lang | ||||||
|  |                           in all_languages.items() | ||||||
|  |                           if (len(lang.get('counter', [])) >= min_supported_engines or | ||||||
|  |                               all(main_engine in lang.get('counter', []) | ||||||
|  |                                   for main_engine in main_engines))} | ||||||
|  | 
 | ||||||
|  |     return filtered_languages | ||||||
|  | 
 | ||||||
|  | 
 | ||||||
|  | # Add country codes to languages without one and filter out language codes. | ||||||
|  | def assign_country_codes(filtered_languages, all_languages): | ||||||
|  |     sorted_languages = sorted(all_languages, | ||||||
|  |                               key=lambda lang: len(all_languages[lang].get('counter', [])), | ||||||
|  |                               reverse=True) | ||||||
|  |     previous_lang = None | ||||||
|  |     previous_code = None | ||||||
|  |     countries = 0 | ||||||
|  |     for current_code in sorted(filtered_languages): | ||||||
|  |         current_lang = current_code.split('-')[0] | ||||||
|  | 
 | ||||||
|  |         # count country codes per language | ||||||
|  |         if current_lang == previous_lang: | ||||||
|             countries += 1 |             countries += 1 | ||||||
|  | 
 | ||||||
|         else: |         else: | ||||||
|             if prev_lang is not None and countries == 1: |             if previous_lang is not None: | ||||||
|                 del languages[prev_lang] |                 # if language has no single country code | ||||||
|                 languages[prev_code]['country'] = '' |                 if countries == 0: | ||||||
|  |                     # try to get country code with most supported engines | ||||||
|  |                     for l in sorted_languages: | ||||||
|  |                         l_parts = l.split('-') | ||||||
|  |                         if len(l_parts) == 2 and l_parts[0] == previous_lang: | ||||||
|  |                             filtered_languages[l] = all_languages[l] | ||||||
|  |                             filtered_languages[l]['country'] = '' | ||||||
|  |                             countries = 1 | ||||||
|  |                             break | ||||||
|  | 
 | ||||||
|  |                     if countries == 0: | ||||||
|  |                         # get most likely country code from babel | ||||||
|  |                         subtags = get_global('likely_subtags').get(previous_lang) | ||||||
|  |                         if subtags: | ||||||
|  |                             subtag_parts = subtags.split('_') | ||||||
|  |                             new_code = subtag_parts[0] + '-' + subtag_parts[-1] | ||||||
|  |                             filtered_languages[new_code] = all_languages[previous_lang] | ||||||
|  |                             countries = 1 | ||||||
|  | 
 | ||||||
|  |                 if countries == 1: | ||||||
|  |                     # remove countryless version of language if there's only one country | ||||||
|  |                     del filtered_languages[previous_lang] | ||||||
|  |                     if previous_code in filtered_languages: | ||||||
|  |                         filtered_languages[previous_code]['country'] = '' | ||||||
|  | 
 | ||||||
|             countries = 0 |             countries = 0 | ||||||
|             prev_lang = lang |             previous_lang = current_lang | ||||||
|         prev_code = code | 
 | ||||||
|  |         previous_code = current_code | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
| # Write languages.py. | # Write languages.py. | ||||||
| def write_languages_file(): | def write_languages_file(languages): | ||||||
|     new_file = open(languages_file, 'wb') |     new_file = open(languages_file, 'wb') | ||||||
|     file_content = '# -*- coding: utf-8 -*-\n'\ |     file_content = '# -*- coding: utf-8 -*-\n'\ | ||||||
|                    + '# list of language codes\n'\ |                    + '# list of language codes\n'\ | ||||||
| @ -183,7 +181,9 @@ def write_languages_file(): | |||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
| if __name__ == "__main__": | if __name__ == "__main__": | ||||||
|     fetch_supported_languages() |     initialize_engines(settings['engines']) | ||||||
|     join_language_lists() |     engines_languages = fetch_supported_languages() | ||||||
|     filter_single_country_languages() |     all_languages = join_language_lists(engines_languages) | ||||||
|     write_languages_file() |     filtered_languages = filter_language_list(all_languages) | ||||||
|  |     assign_country_codes(filtered_languages, all_languages) | ||||||
|  |     write_languages_file(filtered_languages) | ||||||
|  | |||||||
		Loading…
	
	
			
			x
			
			
		
	
		Reference in New Issue
	
	Block a user
	 Adam Tauber
						Adam Tauber