.oO SearXNG Developer Documentation Oo.
Loading...
Searching...
No Matches
google.py
Go to the documentation of this file.
1# SPDX-License-Identifier: AGPL-3.0-or-later
2"""This is the implementation of the Google WEB engine. Some of this
3implementations (manly the :py:obj:`get_google_info`) are shared by other
4engines:
5
6- :ref:`google images engine`
7- :ref:`google news engine`
8- :ref:`google videos engine`
9- :ref:`google scholar engine`
10- :ref:`google autocomplete`
11
12"""
13from __future__ import annotations
14
15from typing import TYPE_CHECKING
16
17import re
18import random
19import string
20import time
21from urllib.parse import urlencode
22from lxml import html
23import babel
24import babel.core
25import babel.languages
26
27from searx.utils import extract_text, eval_xpath, eval_xpath_list, eval_xpath_getindex
28from searx.locales import language_tag, region_tag, get_official_locales
29from searx.network import get # see https://github.com/searxng/searxng/issues/762
30from searx.exceptions import SearxEngineCaptchaException
31from searx.enginelib.traits import EngineTraits
32from searx.result_types import EngineResults
33
34if TYPE_CHECKING:
35 import logging
36
37 logger: logging.Logger
38
39traits: EngineTraits
40
41
42# about
43about = {
44 "website": 'https://www.google.com',
45 "wikidata_id": 'Q9366',
46 "official_api_documentation": 'https://developers.google.com/custom-search/',
47 "use_official_api": False,
48 "require_api_key": False,
49 "results": 'HTML',
50}
51
52# engine dependent config
53categories = ['general', 'web']
54paging = True
55max_page = 50
56"""`Google max 50 pages`_
57
58.. _Google max 50 pages: https://github.com/searxng/searxng/issues/2982
59"""
60time_range_support = True
61safesearch = True
62
63time_range_dict = {'day': 'd', 'week': 'w', 'month': 'm', 'year': 'y'}
64
65# Filter results. 0: None, 1: Moderate, 2: Strict
66filter_mapping = {0: 'off', 1: 'medium', 2: 'high'}
67
68# specific xpath variables
69# ------------------------
70
71# Suggestions are links placed in a *card-section*, we extract only the text
72# from the links not the links itself.
73suggestion_xpath = '//div[contains(@class, "EIaa9b")]//a'
74
75
76_arcid_range = string.ascii_letters + string.digits + "_-"
77_arcid_random: tuple[str, int] | None = None
78
79
80def ui_async(start: int) -> str:
81 """Format of the response from UI's async request.
82
83 - ``arc_id:<...>,use_ac:true,_fmt:prog``
84
85 The arc_id is random generated every hour.
86 """
87 global _arcid_random # pylint: disable=global-statement
88
89 use_ac = "use_ac:true"
90 # _fmt:html returns a HTTP 500 when user search for celebrities like
91 # '!google natasha allegri' or '!google chris evans'
92 _fmt = "_fmt:prog"
93
94 # create a new random arc_id every hour
95 if not _arcid_random or (int(time.time()) - _arcid_random[1]) > 3600:
96 _arcid_random = (''.join(random.choices(_arcid_range, k=23)), int(time.time()))
97 arc_id = f"arc_id:srp_{_arcid_random[0]}_1{start:02}"
98
99 return ",".join([arc_id, use_ac, _fmt])
100
101
102def get_google_info(params, eng_traits):
103 """Composing various (language) properties for the google engines (:ref:`google
104 API`).
105
106 This function is called by the various google engines (:ref:`google web
107 engine`, :ref:`google images engine`, :ref:`google news engine` and
108 :ref:`google videos engine`).
109
110 :param dict param: Request parameters of the engine. At least
111 a ``searxng_locale`` key should be in the dictionary.
112
113 :param eng_traits: Engine's traits fetched from google preferences
114 (:py:obj:`searx.enginelib.traits.EngineTraits`)
115
116 :rtype: dict
117 :returns:
118 Py-Dictionary with the key/value pairs:
119
120 language:
121 The language code that is used by google (e.g. ``lang_en`` or
122 ``lang_zh-TW``)
123
124 country:
125 The country code that is used by google (e.g. ``US`` or ``TW``)
126
127 locale:
128 A instance of :py:obj:`babel.core.Locale` build from the
129 ``searxng_locale`` value.
130
131 subdomain:
132 Google subdomain :py:obj:`google_domains` that fits to the country
133 code.
134
135 params:
136 Py-Dictionary with additional request arguments (can be passed to
137 :py:func:`urllib.parse.urlencode`).
138
139 - ``hl`` parameter: specifies the interface language of user interface.
140 - ``lr`` parameter: restricts search results to documents written in
141 a particular language.
142 - ``cr`` parameter: restricts search results to documents
143 originating in a particular country.
144 - ``ie`` parameter: sets the character encoding scheme that should
145 be used to interpret the query string ('utf8').
146 - ``oe`` parameter: sets the character encoding scheme that should
147 be used to decode the XML result ('utf8').
148
149 headers:
150 Py-Dictionary with additional HTTP headers (can be passed to
151 request's headers)
152
153 - ``Accept: '*/*``
154
155 """
156
157 ret_val = {
158 'language': None,
159 'country': None,
160 'subdomain': None,
161 'params': {},
162 'headers': {},
163 'cookies': {},
164 'locale': None,
165 }
166
167 sxng_locale = params.get('searxng_locale', 'all')
168 try:
169 locale = babel.Locale.parse(sxng_locale, sep='-')
170 except babel.core.UnknownLocaleError:
171 locale = None
172
173 eng_lang = eng_traits.get_language(sxng_locale, 'lang_en')
174 lang_code = eng_lang.split('_')[-1] # lang_zh-TW --> zh-TW / lang_en --> en
175 country = eng_traits.get_region(sxng_locale, eng_traits.all_locale)
176
177 # Test zh_hans & zh_hant --> in the topmost links in the result list of list
178 # TW and HK you should a find wiktionary.org zh_hant link. In the result
179 # list of zh-CN should not be no hant link instead you should find
180 # zh.m.wikipedia.org/zh somewhere in the top.
181
182 # '!go 日 :zh-TW' --> https://zh.m.wiktionary.org/zh-hant/%E6%97%A5
183 # '!go 日 :zh-CN' --> https://zh.m.wikipedia.org/zh/%E6%97%A5
184
185 ret_val['language'] = eng_lang
186 ret_val['country'] = country
187 ret_val['locale'] = locale
188 ret_val['subdomain'] = eng_traits.custom['supported_domains'].get(country.upper(), 'www.google.com')
189
190 # hl parameter:
191 # The hl parameter specifies the interface language (host language) of
192 # your user interface. To improve the performance and the quality of your
193 # search results, you are strongly encouraged to set this parameter
194 # explicitly.
195 # https://developers.google.com/custom-search/docs/xml_results#hlsp
196 # The Interface Language:
197 # https://developers.google.com/custom-search/docs/xml_results_appendices#interfaceLanguages
198
199 # https://github.com/searxng/searxng/issues/2515#issuecomment-1607150817
200 ret_val['params']['hl'] = f'{lang_code}-{country}'
201
202 # lr parameter:
203 # The lr (language restrict) parameter restricts search results to
204 # documents written in a particular language.
205 # https://developers.google.com/custom-search/docs/xml_results#lrsp
206 # Language Collection Values:
207 # https://developers.google.com/custom-search/docs/xml_results_appendices#languageCollections
208 #
209 # To select 'all' languages an empty 'lr' value is used.
210 #
211 # Different to other google services, Google Scholar supports to select more
212 # than one language. The languages are separated by a pipe '|' (logical OR).
213 # By example: &lr=lang_zh-TW%7Clang_de selects articles written in
214 # traditional chinese OR german language.
215
216 ret_val['params']['lr'] = eng_lang
217 if sxng_locale == 'all':
218 ret_val['params']['lr'] = ''
219
220 # cr parameter:
221 # The cr parameter restricts search results to documents originating in a
222 # particular country.
223 # https://developers.google.com/custom-search/docs/xml_results#crsp
224
225 # specify a region (country) only if a region is given in the selected
226 # locale --> https://github.com/searxng/searxng/issues/2672
227 ret_val['params']['cr'] = ''
228 if len(sxng_locale.split('-')) > 1:
229 ret_val['params']['cr'] = 'country' + country
230
231 # gl parameter: (mandatory by Google News)
232 # The gl parameter value is a two-letter country code. For WebSearch
233 # results, the gl parameter boosts search results whose country of origin
234 # matches the parameter value. See the Country Codes section for a list of
235 # valid values.
236 # Specifying a gl parameter value in WebSearch requests should improve the
237 # relevance of results. This is particularly true for international
238 # customers and, even more specifically, for customers in English-speaking
239 # countries other than the United States.
240 # https://developers.google.com/custom-search/docs/xml_results#glsp
241
242 # https://github.com/searxng/searxng/issues/2515#issuecomment-1606294635
243 # ret_val['params']['gl'] = country
244
245 # ie parameter:
246 # The ie parameter sets the character encoding scheme that should be used
247 # to interpret the query string. The default ie value is latin1.
248 # https://developers.google.com/custom-search/docs/xml_results#iesp
249
250 ret_val['params']['ie'] = 'utf8'
251
252 # oe parameter:
253 # The oe parameter sets the character encoding scheme that should be used
254 # to decode the XML result. The default oe value is latin1.
255 # https://developers.google.com/custom-search/docs/xml_results#oesp
256
257 ret_val['params']['oe'] = 'utf8'
258
259 # num parameter:
260 # The num parameter identifies the number of search results to return.
261 # The default num value is 10, and the maximum value is 20. If you request
262 # more than 20 results, only 20 results will be returned.
263 # https://developers.google.com/custom-search/docs/xml_results#numsp
264
265 # HINT: seems to have no effect (tested in google WEB & Images)
266 # ret_val['params']['num'] = 20
267
268 # HTTP headers
269
270 ret_val['headers']['Accept'] = '*/*'
271
272 # Cookies
273
274 # - https://github.com/searxng/searxng/pull/1679#issuecomment-1235432746
275 # - https://github.com/searxng/searxng/issues/1555
276 ret_val['cookies']['CONSENT'] = "YES+"
277
278 return ret_val
279
280
282 if resp.url.host == 'sorry.google.com' or resp.url.path.startswith('/sorry'):
284
285
286def request(query, params):
287 """Google search request"""
288 # pylint: disable=line-too-long
289 start = (params['pageno'] - 1) * 10
290 str_async = ui_async(start)
291 google_info = get_google_info(params, traits)
292 logger.debug("ARC_ID: %s", str_async)
293
294 # https://www.google.de/search?q=corona&hl=de&lr=lang_de&start=0&tbs=qdr%3Ad&safe=medium
295 query_url = (
296 'https://'
297 + google_info['subdomain']
298 + '/search'
299 + "?"
300 + urlencode(
301 {
302 'q': query,
303 **google_info['params'],
304 'filter': '0',
305 'start': start,
306 # 'vet': '12ahUKEwik3ZbIzfn7AhXMX_EDHbUDBh0QxK8CegQIARAC..i',
307 # 'ved': '2ahUKEwik3ZbIzfn7AhXMX_EDHbUDBh0Q_skCegQIARAG',
308 # 'cs' : 1,
309 # 'sa': 'N',
310 # 'yv': 3,
311 # 'prmd': 'vin',
312 # 'ei': 'GASaY6TxOcy_xc8PtYeY6AE',
313 # 'sa': 'N',
314 # 'sstk': 'AcOHfVkD7sWCSAheZi-0tx_09XDO55gTWY0JNq3_V26cNN-c8lfD45aZYPI8s_Bqp8s57AHz5pxchDtAGCA_cikAWSjy9kw3kgg'
315 # formally known as use_mobile_ui
316 'asearch': 'arc',
317 'async': str_async,
318 }
319 )
320 )
321
322 if params['time_range'] in time_range_dict:
323 query_url += '&' + urlencode({'tbs': 'qdr:' + time_range_dict[params['time_range']]})
324 if params['safesearch']:
325 query_url += '&' + urlencode({'safe': filter_mapping[params['safesearch']]})
326 params['url'] = query_url
327
328 params['cookies'] = google_info['cookies']
329 params['headers'].update(google_info['headers'])
330 return params
331
332
333# =26;[3,"dimg_ZNMiZPCqE4apxc8P3a2tuAQ_137"]a87;data:image/jpeg;base64,/9j/4AAQSkZJRgABA
334# ...6T+9Nl4cnD+gr9OK8I56/tX3l86nWYw//2Q==26;
335RE_DATA_IMAGE = re.compile(r'"(dimg_[^"]*)"[^;]*;(data:image[^;]*;[^;]*);')
336RE_DATA_IMAGE_end = re.compile(r'"(dimg_[^"]*)"[^;]*;(data:image[^;]*;[^;]*)$')
337
338
339def parse_data_images(text: str):
340 data_image_map = {}
341
342 for img_id, data_image in RE_DATA_IMAGE.findall(text):
343 end_pos = data_image.rfind('=')
344 if end_pos > 0:
345 data_image = data_image[: end_pos + 1]
346 data_image_map[img_id] = data_image
347 last = RE_DATA_IMAGE_end.search(text)
348 if last:
349 data_image_map[last.group(1)] = last.group(2)
350 logger.debug('data:image objects --> %s', list(data_image_map.keys()))
351 return data_image_map
352
353
354def response(resp) -> EngineResults:
355 """Get response from google's search request"""
356 # pylint: disable=too-many-branches, too-many-statements
358 data_image_map = parse_data_images(resp.text)
359
360 results = EngineResults()
361
362 # convert the text to dom
363 dom = html.fromstring(resp.text)
364
365 # results --> answer
366 answer_list = eval_xpath(dom, '//div[contains(@class, "LGOjhe")]')
367 for item in answer_list:
368 for bubble in eval_xpath(item, './/div[@class="nnFGuf"]'):
369 bubble.drop_tree()
370 results.add(
371 results.types.Answer(
372 answer=extract_text(item),
373 url=(eval_xpath(item, '../..//a/@href') + [None])[0],
374 )
375 )
376
377 # parse results
378
379 for result in eval_xpath_list(dom, './/div[contains(@jscontroller, "SC7lYd")]'):
380 # pylint: disable=too-many-nested-blocks
381
382 try:
383 title_tag = eval_xpath_getindex(result, './/a/h3[1]', 0, default=None)
384 if title_tag is None:
385 # this not one of the common google results *section*
386 logger.debug('ignoring item from the result_xpath list: missing title')
387 continue
388 title = extract_text(title_tag)
389
390 url = eval_xpath_getindex(result, './/a[h3]/@href', 0, None)
391 if url is None:
392 logger.debug('ignoring item from the result_xpath list: missing url of title "%s"', title)
393 continue
394
395 content_nodes = eval_xpath(result, './/div[contains(@data-sncf, "1")]')
396 for item in content_nodes:
397 for script in item.xpath(".//script"):
398 script.getparent().remove(script)
399
400 content = extract_text(content_nodes)
401
402 if not content:
403 logger.debug('ignoring item from the result_xpath list: missing content of title "%s"', title)
404 continue
405
406 thumbnail = content_nodes[0].xpath('.//img/@src')
407 if thumbnail:
408 thumbnail = thumbnail[0]
409 if thumbnail.startswith('data:image'):
410 img_id = content_nodes[0].xpath('.//img/@id')
411 if img_id:
412 thumbnail = data_image_map.get(img_id[0])
413 else:
414 thumbnail = None
415
416 results.append({'url': url, 'title': title, 'content': content, 'thumbnail': thumbnail})
417
418 except Exception as e: # pylint: disable=broad-except
419 logger.error(e, exc_info=True)
420 continue
421
422 # parse suggestion
423 for suggestion in eval_xpath_list(dom, suggestion_xpath):
424 # append suggestion
425 results.append({'suggestion': extract_text(suggestion)})
426
427 # return results
428 return results
429
430
431# get supported languages from their site
432
433
434skip_countries = [
435 # official language of google-country not in google-languages
436 'AL', # Albanien (sq)
437 'AZ', # Aserbaidschan (az)
438 'BD', # Bangladesch (bn)
439 'BN', # Brunei Darussalam (ms)
440 'BT', # Bhutan (dz)
441 'ET', # Äthiopien (am)
442 'GE', # Georgien (ka, os)
443 'GL', # Grönland (kl)
444 'KH', # Kambodscha (km)
445 'LA', # Laos (lo)
446 'LK', # Sri Lanka (si, ta)
447 'ME', # Montenegro (sr)
448 'MK', # Nordmazedonien (mk, sq)
449 'MM', # Myanmar (my)
450 'MN', # Mongolei (mn)
451 'MV', # Malediven (dv) // dv_MV is unknown by babel
452 'MY', # Malaysia (ms)
453 'NP', # Nepal (ne)
454 'TJ', # Tadschikistan (tg)
455 'TM', # Turkmenistan (tk)
456 'UZ', # Usbekistan (uz)
457]
458
459
460def fetch_traits(engine_traits: EngineTraits, add_domains: bool = True):
461 """Fetch languages from Google."""
462 # pylint: disable=import-outside-toplevel, too-many-branches
463
464 engine_traits.custom['supported_domains'] = {}
465
466 resp = get('https://www.google.com/preferences')
467 if not resp.ok: # type: ignore
468 raise RuntimeError("Response from Google's preferences is not OK.")
469
470 dom = html.fromstring(resp.text.replace('<?xml version="1.0" encoding="UTF-8"?>', ''))
471
472 # supported language codes
473
474 lang_map = {'no': 'nb'}
475 for x in eval_xpath_list(dom, "//select[@name='hl']/option"):
476 eng_lang = x.get("value")
477 try:
478 locale = babel.Locale.parse(lang_map.get(eng_lang, eng_lang), sep='-')
479 except babel.UnknownLocaleError:
480 print("INFO: google UI language %s (%s) is unknown by babel" % (eng_lang, x.text.split("(")[0].strip()))
481 continue
482 sxng_lang = language_tag(locale)
483
484 conflict = engine_traits.languages.get(sxng_lang)
485 if conflict:
486 if conflict != eng_lang:
487 print("CONFLICT: babel %s --> %s, %s" % (sxng_lang, conflict, eng_lang))
488 continue
489 engine_traits.languages[sxng_lang] = 'lang_' + eng_lang
490
491 # alias languages
492 engine_traits.languages['zh'] = 'lang_zh-CN'
493
494 # supported region codes
495
496 for x in eval_xpath_list(dom, "//select[@name='gl']/option"):
497 eng_country = x.get("value")
498
499 if eng_country in skip_countries:
500 continue
501 if eng_country == 'ZZ':
502 engine_traits.all_locale = 'ZZ'
503 continue
504
505 sxng_locales = get_official_locales(eng_country, engine_traits.languages.keys(), regional=True)
506
507 if not sxng_locales:
508 print("ERROR: can't map from google country %s (%s) to a babel region." % (x.get('data-name'), eng_country))
509 continue
510
511 for sxng_locale in sxng_locales:
512 engine_traits.regions[region_tag(sxng_locale)] = eng_country
513
514 # alias regions
515 engine_traits.regions['zh-CN'] = 'HK'
516
517 # supported domains
518
519 if add_domains:
520 resp = get('https://www.google.com/supported_domains')
521 if not resp.ok: # type: ignore
522 raise RuntimeError("Response from https://www.google.com/supported_domains is not OK.")
523
524 for domain in resp.text.split(): # type: ignore
525 domain = domain.strip()
526 if not domain or domain in [
527 '.google.com',
528 ]:
529 continue
530 region = domain.split('.')[-1].upper()
531 engine_traits.custom['supported_domains'][region] = 'www' + domain # type: ignore
532 if region == 'HK':
533 # There is no google.cn, we use .com.hk for zh-CN
534 engine_traits.custom['supported_domains']['CN'] = 'www' + domain # type: ignore
request(query, params)
Definition google.py:286
parse_data_images(str text)
Definition google.py:339
str ui_async(int start)
Definition google.py:80
EngineResults response(resp)
Definition google.py:354
get_google_info(params, eng_traits)
Definition google.py:102
fetch_traits(EngineTraits engine_traits, bool add_domains=True)
Definition google.py:460
detect_google_sorry(resp)
Definition google.py:281