[mod] one logger per engine - drop obsolete logger.getChild
Remove the no longer needed `logger = logger.getChild(...)` from engines. Signed-off-by: Markus Heiser <markus.heiser@darmarit.de>
This commit is contained in:
		
							parent
							
								
									7b235a1c36
								
							
						
					
					
						commit
						aecfb2300d
					
				| @ -8,15 +8,12 @@ | |||||||
| from urllib.parse import urlencode | from urllib.parse import urlencode | ||||||
| from lxml import html | from lxml import html | ||||||
| 
 | 
 | ||||||
| from searx import logger |  | ||||||
| from searx.utils import ( | from searx.utils import ( | ||||||
|     eval_xpath_list, |     eval_xpath_list, | ||||||
|     eval_xpath_getindex, |     eval_xpath_getindex, | ||||||
|     extract_text, |     extract_text, | ||||||
| ) | ) | ||||||
| 
 | 
 | ||||||
| logger = logger.getChild('APKMirror engine') |  | ||||||
| 
 |  | ||||||
| about = { | about = { | ||||||
|     "website": 'https://www.apkmirror.com', |     "website": 'https://www.apkmirror.com', | ||||||
|     "wikidata_id": None, |     "wikidata_id": None, | ||||||
|  | |||||||
| @ -13,9 +13,6 @@ Explore thousands of artworks from The Art Institute of Chicago. | |||||||
| from json import loads | from json import loads | ||||||
| from urllib.parse import urlencode | from urllib.parse import urlencode | ||||||
| 
 | 
 | ||||||
| from searx import logger |  | ||||||
| logger = logger.getChild('APKMirror engine') |  | ||||||
| 
 |  | ||||||
| about = { | about = { | ||||||
|     "website": 'https://www.artic.edu', |     "website": 'https://www.artic.edu', | ||||||
|     "wikidata_id": 'Q239303', |     "wikidata_id": 'Q239303', | ||||||
|  | |||||||
| @ -6,11 +6,8 @@ | |||||||
| import re | import re | ||||||
| from urllib.parse import urlencode | from urllib.parse import urlencode | ||||||
| from lxml import html | from lxml import html | ||||||
| from searx import logger |  | ||||||
| from searx.utils import eval_xpath, extract_text, match_language | from searx.utils import eval_xpath, extract_text, match_language | ||||||
| 
 | 
 | ||||||
| logger = logger.getChild('bing engine') |  | ||||||
| 
 |  | ||||||
| # about | # about | ||||||
| about = { | about = { | ||||||
|     "website": 'https://www.bing.com', |     "website": 'https://www.bing.com', | ||||||
|  | |||||||
| @ -9,11 +9,8 @@ from json import loads | |||||||
| from datetime import datetime | from datetime import datetime | ||||||
| from urllib.parse import urlencode | from urllib.parse import urlencode | ||||||
| 
 | 
 | ||||||
| from searx import logger |  | ||||||
| from searx.exceptions import SearxEngineAPIException | from searx.exceptions import SearxEngineAPIException | ||||||
| 
 | 
 | ||||||
| logger = logger.getChild('CORE engine') |  | ||||||
| 
 |  | ||||||
| about = { | about = { | ||||||
|     "website": 'https://core.ac.uk', |     "website": 'https://core.ac.uk', | ||||||
|     "wikidata_id": 'Q22661180', |     "wikidata_id": 'Q22661180', | ||||||
| @ -29,8 +26,6 @@ nb_per_page = 10 | |||||||
| 
 | 
 | ||||||
| api_key = 'unset' | api_key = 'unset' | ||||||
| 
 | 
 | ||||||
| logger = logger.getChild('CORE engine') |  | ||||||
| 
 |  | ||||||
| base_url = 'https://core.ac.uk:443/api-v2/search/' | base_url = 'https://core.ac.uk:443/api-v2/search/' | ||||||
| search_string = '{query}?page={page}&pageSize={nb_per_page}&apiKey={apikey}' | search_string = '{query}?page={page}&pageSize={nb_per_page}&apiKey={apikey}' | ||||||
| 
 | 
 | ||||||
|  | |||||||
| @ -9,15 +9,12 @@ import json | |||||||
| from urllib.parse import urlencode, urlparse, urljoin | from urllib.parse import urlencode, urlparse, urljoin | ||||||
| from lxml import html | from lxml import html | ||||||
| 
 | 
 | ||||||
| from searx import logger |  | ||||||
| from searx.data import WIKIDATA_UNITS | from searx.data import WIKIDATA_UNITS | ||||||
| from searx.engines.duckduckgo import language_aliases | from searx.engines.duckduckgo import language_aliases | ||||||
| from searx.engines.duckduckgo import _fetch_supported_languages, supported_languages_url  # NOQA # pylint: disable=unused-import | from searx.engines.duckduckgo import _fetch_supported_languages, supported_languages_url  # NOQA # pylint: disable=unused-import | ||||||
| from searx.utils import extract_text, html_to_text, match_language, get_string_replaces_function | from searx.utils import extract_text, html_to_text, match_language, get_string_replaces_function | ||||||
| from searx.external_urls import get_external_url, get_earth_coordinates_url, area_to_osm_zoom | from searx.external_urls import get_external_url, get_earth_coordinates_url, area_to_osm_zoom | ||||||
| 
 | 
 | ||||||
| logger = logger.getChild('duckduckgo_definitions') |  | ||||||
| 
 |  | ||||||
| # about | # about | ||||||
| about = { | about = { | ||||||
|     "website": 'https://duckduckgo.com/', |     "website": 'https://duckduckgo.com/', | ||||||
|  | |||||||
| @ -7,11 +7,8 @@ from json import loads | |||||||
| from time import time | from time import time | ||||||
| import re | import re | ||||||
| from urllib.parse import urlencode | from urllib.parse import urlencode | ||||||
| from searx.engines import logger |  | ||||||
| from searx.utils import ecma_unescape, html_to_text | from searx.utils import ecma_unescape, html_to_text | ||||||
| 
 | 
 | ||||||
| logger = logger.getChild('flickr-noapi') |  | ||||||
| 
 |  | ||||||
| # about | # about | ||||||
| about = { | about = { | ||||||
|     "website": 'https://www.flickr.com', |     "website": 'https://www.flickr.com', | ||||||
|  | |||||||
| @ -9,9 +9,6 @@ from json import loads | |||||||
| from urllib.parse import urlencode | from urllib.parse import urlencode | ||||||
| from datetime import datetime | from datetime import datetime | ||||||
| 
 | 
 | ||||||
| from searx import logger |  | ||||||
| logger = logger.getChild('genius engine') |  | ||||||
| 
 |  | ||||||
| # about | # about | ||||||
| about = { | about = { | ||||||
|     "website": 'https://genius.com/', |     "website": 'https://genius.com/', | ||||||
|  | |||||||
| @ -8,7 +8,6 @@ | |||||||
| import re | import re | ||||||
| from json import loads | from json import loads | ||||||
| from urllib.parse import urlencode | from urllib.parse import urlencode | ||||||
| # from searx import logger |  | ||||||
| from searx.network import get | from searx.network import get | ||||||
| 
 | 
 | ||||||
| # about | # about | ||||||
|  | |||||||
| @ -29,12 +29,9 @@ The google WEB engine itself has a special setup option: | |||||||
| 
 | 
 | ||||||
| from urllib.parse import urlencode | from urllib.parse import urlencode | ||||||
| from lxml import html | from lxml import html | ||||||
| from searx import logger |  | ||||||
| from searx.utils import match_language, extract_text, eval_xpath, eval_xpath_list, eval_xpath_getindex | from searx.utils import match_language, extract_text, eval_xpath, eval_xpath_list, eval_xpath_getindex | ||||||
| from searx.exceptions import SearxEngineCaptchaException | from searx.exceptions import SearxEngineCaptchaException | ||||||
| 
 | 
 | ||||||
| logger = logger.getChild('google engine') |  | ||||||
| 
 |  | ||||||
| # about | # about | ||||||
| about = { | about = { | ||||||
|     "website": 'https://www.google.com', |     "website": 'https://www.google.com', | ||||||
|  | |||||||
| @ -16,7 +16,6 @@ | |||||||
| from urllib.parse import urlencode, unquote | from urllib.parse import urlencode, unquote | ||||||
| from lxml import html | from lxml import html | ||||||
| 
 | 
 | ||||||
| from searx import logger |  | ||||||
| from searx.utils import ( | from searx.utils import ( | ||||||
|     eval_xpath, |     eval_xpath, | ||||||
|     eval_xpath_list, |     eval_xpath_list, | ||||||
| @ -37,8 +36,6 @@ from searx.engines.google import ( | |||||||
| ) | ) | ||||||
| # pylint: enable=unused-import | # pylint: enable=unused-import | ||||||
| 
 | 
 | ||||||
| logger = logger.getChild('google images') |  | ||||||
| 
 |  | ||||||
| # about | # about | ||||||
| about = { | about = { | ||||||
|     "website": 'https://images.google.com', |     "website": 'https://images.google.com', | ||||||
|  | |||||||
| @ -20,7 +20,6 @@ from urllib.parse import urlencode | |||||||
| from base64 import b64decode | from base64 import b64decode | ||||||
| from lxml import html | from lxml import html | ||||||
| 
 | 
 | ||||||
| from searx import logger |  | ||||||
| from searx.utils import ( | from searx.utils import ( | ||||||
|     eval_xpath, |     eval_xpath, | ||||||
|     eval_xpath_list, |     eval_xpath_list, | ||||||
| @ -50,8 +49,6 @@ about = { | |||||||
|     "results": 'HTML', |     "results": 'HTML', | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
| logger = logger.getChild('google news') |  | ||||||
| 
 |  | ||||||
| # compared to other google engines google-news has a different time range | # compared to other google engines google-news has a different time range | ||||||
| # support.  The time range is included in the search term. | # support.  The time range is included in the search term. | ||||||
| time_range_dict = { | time_range_dict = { | ||||||
|  | |||||||
| @ -14,7 +14,6 @@ Definitions`_. | |||||||
| from urllib.parse import urlencode | from urllib.parse import urlencode | ||||||
| from datetime import datetime | from datetime import datetime | ||||||
| from lxml import html | from lxml import html | ||||||
| from searx import logger |  | ||||||
| 
 | 
 | ||||||
| from searx.utils import ( | from searx.utils import ( | ||||||
|     eval_xpath, |     eval_xpath, | ||||||
| @ -53,8 +52,6 @@ use_locale_domain = True | |||||||
| time_range_support = True | time_range_support = True | ||||||
| safesearch = False | safesearch = False | ||||||
| 
 | 
 | ||||||
| logger = logger.getChild('google scholar') |  | ||||||
| 
 |  | ||||||
| def time_range_url(params): | def time_range_url(params): | ||||||
|     """Returns a URL query component for a google-Scholar time range based on |     """Returns a URL query component for a google-Scholar time range based on | ||||||
|     ``params['time_range']``.  Google-Scholar does only support ranges in years. |     ``params['time_range']``.  Google-Scholar does only support ranges in years. | ||||||
|  | |||||||
| @ -20,7 +20,6 @@ import re | |||||||
| from urllib.parse import urlencode | from urllib.parse import urlencode | ||||||
| from lxml import html | from lxml import html | ||||||
| 
 | 
 | ||||||
| from searx import logger |  | ||||||
| from searx.utils import ( | from searx.utils import ( | ||||||
|     eval_xpath, |     eval_xpath, | ||||||
|     eval_xpath_list, |     eval_xpath_list, | ||||||
| @ -59,8 +58,6 @@ about = { | |||||||
|     "results": 'HTML', |     "results": 'HTML', | ||||||
| } | } | ||||||
| 
 | 
 | ||||||
| logger = logger.getChild('google video') |  | ||||||
| 
 |  | ||||||
| # engine dependent config | # engine dependent config | ||||||
| 
 | 
 | ||||||
| categories = ['videos'] | categories = ['videos'] | ||||||
|  | |||||||
| @ -8,9 +8,6 @@ | |||||||
| 
 | 
 | ||||||
| from json import loads | from json import loads | ||||||
| from urllib.parse import urlencode | from urllib.parse import urlencode | ||||||
| from searx import logger |  | ||||||
| 
 |  | ||||||
| logger = logger.getChild('solidtor engine') |  | ||||||
| 
 | 
 | ||||||
| about = { | about = { | ||||||
|     "website": 'https://www.solidtorrents.net/', |     "website": 'https://www.solidtorrents.net/', | ||||||
|  | |||||||
| @ -8,7 +8,6 @@ from json import loads | |||||||
| from lxml import html | from lxml import html | ||||||
| from dateutil import parser | from dateutil import parser | ||||||
| from urllib.parse import quote_plus, urlencode | from urllib.parse import quote_plus, urlencode | ||||||
| from searx import logger |  | ||||||
| from searx.network import get as http_get | from searx.network import get as http_get | ||||||
| 
 | 
 | ||||||
| # about | # about | ||||||
|  | |||||||
| @ -10,11 +10,8 @@ from datetime import datetime | |||||||
| from json import loads | from json import loads | ||||||
| from urllib.parse import urlencode | from urllib.parse import urlencode | ||||||
| 
 | 
 | ||||||
| from searx import logger |  | ||||||
| from searx.exceptions import SearxEngineAPIException | from searx.exceptions import SearxEngineAPIException | ||||||
| 
 | 
 | ||||||
| logger = logger.getChild('Springer Nature engine') |  | ||||||
| 
 |  | ||||||
| about = { | about = { | ||||||
|     "website": 'https://www.springernature.com/', |     "website": 'https://www.springernature.com/', | ||||||
|     "wikidata_id": 'Q21096327', |     "wikidata_id": 'Q21096327', | ||||||
|  | |||||||
| @ -9,11 +9,6 @@ | |||||||
| import sqlite3 | import sqlite3 | ||||||
| import contextlib | import contextlib | ||||||
| 
 | 
 | ||||||
| from searx import logger |  | ||||||
| 
 |  | ||||||
| 
 |  | ||||||
| logger = logger.getChild('SQLite engine') |  | ||||||
| 
 |  | ||||||
| engine_type = 'offline' | engine_type = 'offline' | ||||||
| database = "" | database = "" | ||||||
| query_str = "" | query_str = "" | ||||||
|  | |||||||
| @ -8,9 +8,6 @@ | |||||||
| from urllib.parse import urlencode, urlparse, urlunparse, parse_qsl | from urllib.parse import urlencode, urlparse, urlunparse, parse_qsl | ||||||
| from json import loads | from json import loads | ||||||
| 
 | 
 | ||||||
| from searx import logger |  | ||||||
| 
 |  | ||||||
| logger = logger.getChild('unsplash engine') |  | ||||||
| # about | # about | ||||||
| about = { | about = { | ||||||
|     "website": 'https://unsplash.com', |     "website": 'https://unsplash.com', | ||||||
|  | |||||||
| @ -10,15 +10,12 @@ from json import loads | |||||||
| from dateutil.parser import isoparse | from dateutil.parser import isoparse | ||||||
| from babel.dates import format_datetime, format_date, format_time, get_datetime_format | from babel.dates import format_datetime, format_date, format_time, get_datetime_format | ||||||
| 
 | 
 | ||||||
| from searx import logger |  | ||||||
| from searx.data import WIKIDATA_UNITS | from searx.data import WIKIDATA_UNITS | ||||||
| from searx.network import post, get | from searx.network import post, get | ||||||
| from searx.utils import match_language, searx_useragent, get_string_replaces_function | from searx.utils import match_language, searx_useragent, get_string_replaces_function | ||||||
| from searx.external_urls import get_external_url, get_earth_coordinates_url, area_to_osm_zoom | from searx.external_urls import get_external_url, get_earth_coordinates_url, area_to_osm_zoom | ||||||
| from searx.engines.wikipedia import _fetch_supported_languages, supported_languages_url  # NOQA # pylint: disable=unused-import | from searx.engines.wikipedia import _fetch_supported_languages, supported_languages_url  # NOQA # pylint: disable=unused-import | ||||||
| 
 | 
 | ||||||
| logger = logger.getChild('wikidata') |  | ||||||
| 
 |  | ||||||
| # about | # about | ||||||
| about = { | about = { | ||||||
|     "website": 'https://wikidata.org/', |     "website": 'https://wikidata.org/', | ||||||
|  | |||||||
| @ -4,12 +4,9 @@ | |||||||
| """ | """ | ||||||
| 
 | 
 | ||||||
| from lxml.html import fromstring | from lxml.html import fromstring | ||||||
| from searx import logger |  | ||||||
| from searx.utils import extract_text | from searx.utils import extract_text | ||||||
| from searx.network import raise_for_httperror | from searx.network import raise_for_httperror | ||||||
| 
 | 
 | ||||||
| logger = logger.getChild('Wordnik engine') |  | ||||||
| 
 |  | ||||||
| # about | # about | ||||||
| about = { | about = { | ||||||
|     "website": 'https://www.wordnik.com', |     "website": 'https://www.wordnik.com', | ||||||
|  | |||||||
| @ -23,9 +23,6 @@ from urllib.parse import urlencode | |||||||
| 
 | 
 | ||||||
| from lxml import html | from lxml import html | ||||||
| from searx.utils import extract_text, extract_url, eval_xpath, eval_xpath_list | from searx.utils import extract_text, extract_url, eval_xpath, eval_xpath_list | ||||||
| from searx import logger |  | ||||||
| 
 |  | ||||||
| logger = logger.getChild('XPath engine') |  | ||||||
| 
 | 
 | ||||||
| search_url = None | search_url = None | ||||||
| """ | """ | ||||||
|  | |||||||
| @ -14,7 +14,6 @@ from datetime import datetime, timedelta | |||||||
| from dateutil import parser | from dateutil import parser | ||||||
| from lxml import html | from lxml import html | ||||||
| 
 | 
 | ||||||
| from searx import logger |  | ||||||
| from searx.utils import ( | from searx.utils import ( | ||||||
|     eval_xpath_list, |     eval_xpath_list, | ||||||
|     eval_xpath_getindex, |     eval_xpath_getindex, | ||||||
| @ -23,8 +22,6 @@ from searx.utils import ( | |||||||
| 
 | 
 | ||||||
| from searx.engines.yahoo import parse_url | from searx.engines.yahoo import parse_url | ||||||
| 
 | 
 | ||||||
| logger = logger.getChild('yahoo_news engine') |  | ||||||
| 
 |  | ||||||
| # about | # about | ||||||
| about = { | about = { | ||||||
|     "website": 'https://news.yahoo.com', |     "website": 'https://news.yahoo.com', | ||||||
|  | |||||||
		Loading…
	
	
			
			x
			
			
		
	
		Reference in New Issue
	
	Block a user
	 Markus Heiser
						Markus Heiser