183def parse_web_api(resp):
184 """Parse results from Qwant's API"""
185
186
187 results = []
188
189
190 search_results = loads(resp.text)
191 data = search_results.get('data', {})
192
193
194 if search_results.get('status') != 'success':
195 error_code = data.get('error_code')
196 if error_code == 24:
197 raise SearxEngineTooManyRequestsException()
198 if search_results.get("data", {}).get("error_data", {}).get("captchaUrl") is not None:
199 raise SearxEngineCaptchaException()
200 msg = ",".join(data.get('message', ['unknown']))
201 raise SearxEngineAPIException(f"{msg} ({error_code})")
202
203
204 raise_for_httperror(resp)
205
206 if qwant_categ == 'web':
207
208
209
210 mainline = data.get('result', {}).get('items', {}).get('mainline', {})
211 else:
212
213
214
215 mainline = data.get('result', {}).get('items', [])
216 mainline = [
217 {'type': qwant_categ, 'items': mainline},
218 ]
219
220
221 if not mainline:
222 return []
223
224 for row in mainline:
225 mainline_type = row.get('type', 'web')
226 if mainline_type != qwant_categ:
227 continue
228
229 if mainline_type == 'ads':
230
231 continue
232
233 mainline_items = row.get('items', [])
234 for item in mainline_items:
235
236 title = item.get('title', None)
237 res_url = item.get('url', None)
238
239 if mainline_type == 'web':
240 content = item['desc']
241 results.append(
242 {
243 'title': title,
244 'url': res_url,
245 'content': content,
246 }
247 )
248
249 elif mainline_type == 'news':
250
251 pub_date = item['date']
252 if pub_date is not None:
253 pub_date = datetime.fromtimestamp(pub_date)
254 news_media = item.get('media', [])
255 thumbnail = None
256 if news_media:
257 thumbnail = news_media[0].get('pict', {}).get('url', None)
258 results.append(
259 {
260 'title': title,
261 'url': res_url,
262 'publishedDate': pub_date,
263 'thumbnail': thumbnail,
264 }
265 )
266
267 elif mainline_type == 'images':
268 thumbnail = item['thumbnail']
269 img_src = item['media']
270 results.append(
271 {
272 'title': title,
273 'url': res_url,
274 'template': 'images.html',
275 'thumbnail_src': thumbnail,
276 'img_src': img_src,
277 'resolution': f"{item['width']} x {item['height']}",
278 'img_format': item.get('thumb_type'),
279 }
280 )
281
282 elif mainline_type == 'videos':
283
284
285
286 d, s, c = item.get('desc'), item.get('source'), item.get('channel')
287 content_parts = []
288 if d:
289 content_parts.append(d)
290 if s:
291 content_parts.append("%s: %s " % (gettext("Source"), s))
292 if c:
293 content_parts.append("%s: %s " % (gettext("Channel"), c))
294 content = ' // '.join(content_parts)
295 length = item['duration']
296 if length is not None:
297 length = timedelta(milliseconds=length)
298 pub_date = item['date']
299 if pub_date is not None:
300 pub_date = datetime.fromtimestamp(pub_date)
301 thumbnail = item['thumbnail']
302
303
304 thumbnail = thumbnail.replace('https://s2.qwant.com', 'https://s1.qwant.com', 1)
305 results.append(
306 {
307 'title': title,
308 'url': res_url,
309 'content': content,
310 'iframe_src': get_embeded_stream_url(res_url),
311 'publishedDate': pub_date,
312 'thumbnail': thumbnail,
313 'template': 'videos.html',
314 'length': length,
315 }
316 )
317
318 return results
319
320