163 """Parse results from Qwant-Lite"""
166 dom = lxml.html.fromstring(resp.text)
168 for item
in eval_xpath_list(dom,
'//section/article'):
169 if eval_xpath(item,
"./span[contains(@class, 'tooltip')]"):
174 'url': extract_text(eval_xpath(item,
"./span[contains(@class, 'url partner')]")),
175 'title': extract_text(eval_xpath(item,
'./h2/a')),
176 'content': extract_text(eval_xpath(item,
'./p')),
184 """Parse results from Qwant's API"""
190 search_results = loads(resp.text)
191 data = search_results.get(
'data', {})
194 if search_results.get(
'status') !=
'success':
195 error_code = data.get(
'error_code')
198 if search_results.get(
"data", {}).get(
"error_data", {}).get(
"captchaUrl")
is not None:
200 msg =
",".join(data.get(
'message', [
'unknown']))
204 raise_for_httperror(resp)
206 if qwant_categ ==
'web':
210 mainline = data.get(
'result', {}).get(
'items', {}).get(
'mainline', {})
215 mainline = data.get(
'result', {}).get(
'items', [])
217 {
'type': qwant_categ,
'items': mainline},
225 mainline_type = row.get(
'type',
'web')
226 if mainline_type != qwant_categ:
229 if mainline_type ==
'ads':
233 mainline_items = row.get(
'items', [])
234 for item
in mainline_items:
236 title = item.get(
'title',
None)
237 res_url = item.get(
'url',
None)
239 if mainline_type ==
'web':
240 content = item[
'desc']
249 elif mainline_type ==
'news':
251 pub_date = item[
'date']
252 if pub_date
is not None:
253 pub_date = datetime.fromtimestamp(pub_date)
254 news_media = item.get(
'media', [])
257 thumbnail = news_media[0].get(
'pict', {}).get(
'url',
None)
262 'publishedDate': pub_date,
263 'thumbnail': thumbnail,
267 elif mainline_type ==
'images':
268 thumbnail = item[
'thumbnail']
269 img_src = item[
'media']
274 'template':
'images.html',
275 'thumbnail_src': thumbnail,
277 'resolution': f
"{item['width']} x {item['height']}",
278 'img_format': item.get(
'thumb_type'),
282 elif mainline_type ==
'videos':
286 d, s, c = item.get(
'desc'), item.get(
'source'), item.get(
'channel')
289 content_parts.append(d)
291 content_parts.append(
"%s: %s " % (gettext(
"Source"), s))
293 content_parts.append(
"%s: %s " % (gettext(
"Channel"), c))
294 content =
' // '.join(content_parts)
295 length = item[
'duration']
296 if length
is not None:
297 length = timedelta(milliseconds=length)
298 pub_date = item[
'date']
299 if pub_date
is not None:
300 pub_date = datetime.fromtimestamp(pub_date)
301 thumbnail = item[
'thumbnail']
304 thumbnail = thumbnail.replace(
'https://s2.qwant.com',
'https://s1.qwant.com', 1)
310 'iframe_src': get_embeded_stream_url(res_url),
311 'publishedDate': pub_date,
312 'thumbnail': thumbnail,
313 'template':
'videos.html',
324 from searx
import network
328 resp = network.get(about[
'website'])
329 json_string = extr(resp.text,
'INITIAL_PROPS = ',
'</script>')
331 q_initial_props = loads(json_string)
332 q_locales = q_initial_props.get(
'locales')
335 for country, v
in q_locales.items():
336 for lang
in v[
'langs']:
337 _locale =
"{lang}_{country}".format(lang=lang, country=country)
339 if qwant_categ ==
'news' and _locale.lower()
not in qwant_news_locales:
343 eng_tag_list.add(_locale)
345 for eng_tag
in eng_tag_list:
347 sxng_tag = region_tag(babel.Locale.parse(eng_tag, sep=
'_'))
348 except babel.UnknownLocaleError:
349 print(
"ERROR: can't determine babel locale of quant's locale %s" % eng_tag)
352 conflict = engine_traits.regions.get(sxng_tag)
354 if conflict != eng_tag:
355 print(
"CONFLICT: babel %s --> %s, %s" % (sxng_tag, conflict, eng_tag))
357 engine_traits.regions[sxng_tag] = eng_tag