3from __future__
import annotations
6from collections
import defaultdict
7from threading
import RLock
8from typing
import List, NamedTuple, Set
10from searx
import logger
as log
20 for result_engine
in result[
'engines']:
21 if hasattr(searx.engines.engines.get(result_engine),
'weight'):
22 weight *= float(searx.engines.engines[result_engine].weight)
24 weight *= len(result[
'positions'])
27 for position
in result[
'positions']:
30 if priority ==
'high':
33 score += weight / position
51 """In the result container, the results are collected, sorted and duplicates
56 main_results_map: dict[int, MainResult | LegacyResult]
57 infoboxes: list[LegacyResult]
70 self.
engine_data: dict[str, dict[str, str]] = defaultdict(dict)
80 def extend(self, engine_name: str |
None, results):
82 log.debug(
"container is closed, ignoring results: %s", results)
86 for result
in list(results):
88 if isinstance(result, Result):
89 result.engine = result.engine
or engine_name
90 result.normalize_result_fields()
92 if isinstance(result, BaseAnswer)
and self.
on_result(result):
94 elif isinstance(result, MainResult)
and self.
on_result(result):
99 raise NotImplementedError(f
"no handler implemented to process the result of type {result}")
102 result[
"engine"] = result.get(
"engine")
or engine_name
or ""
104 result.normalize_result_fields()
106 if "suggestion" in result:
111 if "answer" in result:
114 f
"answer results from engine {result.engine}"
115 " are without typification / migrate to Answer class.",
121 if "correction" in result:
126 if "infobox" in result:
131 if "number_of_results" in result:
136 if "engine_data" in result:
139 self.
engine_data[result.engine][result[
"key"]] = result[
"engine_data"]
147 if engine_name
in searx.engines.engines:
148 eng = searx.engines.engines[engine_name]
149 histogram_observe(main_count,
"engine", eng.name,
"result",
"count")
150 if not self.
paging and eng.paging:
156 new_id = getattr(new_infobox,
"id",
None)
157 if new_id
is not None:
160 if new_id == getattr(existing_infobox,
"id",
None):
167 result_hash = hash(result)
174 result.positions = [position]
180 merged.positions.append(position)
187 for eng_name
in result.engines:
188 counter_add(result.score,
'engine', eng_name,
'score')
191 """Returns a sorted list of results to be displayed in the main result
192 area (:ref:`result types`)."""
201 results = sorted(self.
main_results_map.values(), key=
lambda x: x.score, reverse=
True)
205 categoryPositions = {}
211 engine = searx.engines.engines.get(res.engine
or "")
213 res.category = engine.categories[0]
if len(engine.categories) > 0
else ""
216 category = f
"{res.category}:{res.template}:{'img_src' if (res.thumbnail or res.img_src) else ''}"
217 grp = categoryPositions.get(category)
223 if (grp
is not None)
and (grp[
"count"] > 0)
and (len(gresults) - grp[
"index"] < max_distance):
227 gresults.insert(index, res)
231 for item
in categoryPositions.values():
234 item[
"index"] = v + 1
242 categoryPositions[category] = {
"index": len(gresults),
"count": max_count}
250 """Returns the average of results number, returns zero if the average
251 result number is smaller than the actual result count."""
254 log.error(
"call to ResultContainer.number_of_results before ResultContainer.close")
270 log.error(
"call to ResultContainer.add_unresponsive_engine after ResultContainer.close")
272 if searx.engines.engines[engine_name].display_error_messages:
275 def add_timing(self, engine_name: str, engine_time: float, page_load_time: float):
278 log.error(
"call to ResultContainer.add_timing after ResultContainer.close")
280 self.
timings.append(
Timing(engine_name, total=engine_time, load=page_load_time))
285 log.error(
"call to ResultContainer.get_timings before ResultContainer.close")
291 """Merges the values from ``other`` into ``origin``."""
293 weight1 = getattr(searx.engines.engines[origin.engine],
"weight", 1)
294 weight2 = getattr(searx.engines.engines[other.engine],
"weight", 1)
296 if weight2 > weight1:
297 origin.engine = other.engine
299 origin.engines |= other.engines
302 url_items = origin.get(
"urls", [])
304 for url2
in other.urls:
306 entity_url2 = url2.get(
"entity")
308 for url1
in origin.get(
"urls", []):
309 if (entity_url2
is not None and entity_url2 == url1.get(
"entity"))
or (
310 url1.get(
"url") == url2.get(
"url")
315 url_items.append(url2)
317 origin.urls = url_items
320 if not origin.img_src:
321 origin.img_src = other.img_src
322 elif weight2 > weight1:
323 origin.img_src = other.img_src
326 if not origin.attributes:
327 origin.attributes = other.attributes
330 for attr
in origin.attributes:
331 label = attr.get(
"label")
333 attr_names_1.add(label)
335 entity = attr.get(
"entity")
337 attr_names_1.add(entity)
339 for attr
in other.attributes:
340 if attr.get(
"label")
not in attr_names_1
and attr.get(
'entity')
not in attr_names_1:
341 origin.attributes.append(attr)
344 if not origin.content:
345 origin.content = other.content
346 elif len(other.content) > len(origin.content):
347 origin.content = other.content
351 """Merges the values from ``other`` into ``origin``."""
353 if len(other.content) > len(origin.content):
355 origin.content = other.content
358 if len(other.title) > len(origin.title):
359 origin.title = other.title
362 if isinstance(other, MainResult)
and isinstance(origin, MainResult):
363 origin.defaults_from(other)
364 elif isinstance(other, LegacyResult)
and isinstance(origin, LegacyResult):
365 origin.defaults_from(other)
368 origin.engines.add(other.engine
or "")
371 if origin.parsed_url
and not origin.parsed_url.scheme.endswith(
"s"):
372 if other.parsed_url
and other.parsed_url.scheme.endswith(
"s"):
373 origin.parsed_url = origin.parsed_url._replace(scheme=other.parsed_url.scheme)
374 origin.url = origin.parsed_url.geturl()
add_unresponsive_engine(self, str engine_name, str error_type, bool suspended=False)
dict[str, dict[str, str]] engine_data
extend(self, str|None engine_name, results)
list[MainResult|LegacyResult] _main_results_sorted
Set[UnresponsiveEngine] unresponsive_engines
_merge_main_result(self, MainResult|LegacyResult result, position)
_merge_infobox(self, LegacyResult new_infobox)
int number_of_results(self)
list[MainResult|LegacyResult] get_ordered_results(self)
add_timing(self, str engine_name, float engine_time, float page_load_time)
merge_two_infoboxes(LegacyResult origin, LegacyResult other)
merge_two_main_results(MainResult|LegacyResult origin, MainResult|LegacyResult other)
float calculate_score(result, priority)