summaryrefslogtreecommitdiff
path: root/searx/engines/__init__.py
diff options
context:
space:
mode:
authorasciimoo <asciimoo@gmail.com>2013-10-27 14:21:23 +0100
committerasciimoo <asciimoo@gmail.com>2013-10-27 14:21:23 +0100
commite2c1c10657adaa248f795d5ba8059f776be04d05 (patch)
tree83ddb97edaa4a718aa8a5913b427728d198d5313 /searx/engines/__init__.py
parent13a4582a40fd91b073dd54b4d93b7df882035ba9 (diff)
[enh] engine score stats
Diffstat (limited to 'searx/engines/__init__.py')
-rw-r--r--searx/engines/__init__.py16
1 files changed, 14 insertions, 2 deletions
diff --git a/searx/engines/__init__.py b/searx/engines/__init__.py
index 366b83840..7447e8f8f 100644
--- a/searx/engines/__init__.py
+++ b/searx/engines/__init__.py
@@ -69,7 +69,7 @@ for section in engines_config.sections():
print '[E] Engine config error: Missing attribute "{0}.{1}"'.format(engine.name, engine_attr)
sys.exit(1)
engines[engine.name] = engine
- engine.stats = {'result_count': 0, 'search_count': 0, 'page_load_time': 0}
+ engine.stats = {'result_count': 0, 'search_count': 0, 'page_load_time': 0, 'score_count': 0}
if hasattr(engine, 'categories'):
for category_name in engine.categories:
categories.setdefault(category_name, []).append(engine)
@@ -136,6 +136,7 @@ def search(query, request, selected_categories):
for i,res in enumerate(flat_res):
res['parsed_url'] = urlparse(res['url'])
score = (flat_len - i)*settings.weights.get(res['engine'], 1)
+ engines[res['engine']].stats['score_count'] += score
duplicated = False
for new_res in results:
if res['parsed_url'].netloc == new_res['parsed_url'].netloc and\
@@ -163,17 +164,24 @@ def search(query, request, selected_categories):
def get_engines_stats():
pageloads = []
results = []
+ scores = []
- max_pageload = max_results = 0
+ max_pageload = max_results = max_score = 0
for engine in engines.values():
if engine.stats['search_count'] == 0:
continue
results_num = engine.stats['result_count']/float(engine.stats['search_count'])
load_times = engine.stats['page_load_time']/float(engine.stats['search_count'])
+ if results_num:
+ score = engine.stats['score_count'] / float(engine.stats['search_count']) / results_num
+ else:
+ score = 0
max_results = max(results_num, max_results)
max_pageload = max(load_times, max_pageload)
+ max_score = max(score, max_score)
pageloads.append({'avg': load_times, 'name': engine.name})
results.append({'avg': results_num, 'name': engine.name})
+ scores.append({'avg': score, 'name': engine.name})
for engine in pageloads:
engine['percentage'] = int(engine['avg']/max_pageload*100)
@@ -181,7 +189,11 @@ def get_engines_stats():
for engine in results:
engine['percentage'] = int(engine['avg']/max_results*100)
+ for engine in scores:
+ engine['percentage'] = int(engine['avg']/max_score*100)
+
return [('Page loads (sec)', sorted(pageloads, key=itemgetter('avg'), reverse=True))
,('Number of results', sorted(results, key=itemgetter('avg'), reverse=True))
+ ,('Scores', sorted(scores, key=itemgetter('avg'), reverse=True))
]