|
import glob |
|
import json |
|
import pickle |
|
import sys |
|
from typing import Dict |
|
|
|
import numpy as np |
|
from beir.retrieval.evaluation import EvaluateRetrieval |
|
|
|
|
|
def load_qrels(filename: str) -> Dict: |
|
with open(filename, "r") as f: |
|
qrels = json.load(f) |
|
return qrels |
|
|
|
|
|
def merge_retrieved_shards( |
|
suffix: str, output_file: str, top_n: int, qrels: dict, metric: str |
|
) -> None: |
|
shard_files = glob.glob(f"*{suffix}") |
|
print(f"There are {len(shard_files)} shards found") |
|
merged_results = {} |
|
print("Loading All shards") |
|
for shard_file in shard_files: |
|
print(f"Loading shard {shard_file} ") |
|
with open(shard_file, "rb") as f: |
|
shard_results = pickle.load(f) |
|
for query_id, doc_scores in shard_results.items(): |
|
if query_id not in merged_results: |
|
merged_results[query_id] = [] |
|
merged_results[query_id].extend(doc_scores.items()) |
|
print("Shards all loaded, merging results and sorting by score") |
|
run = {} |
|
per_query = [] |
|
for query_id, doc_scores in merged_results.items(): |
|
if query_id in qrels: |
|
doc_score_dict = {} |
|
for passage_id, score in doc_scores: |
|
doc_id = passage_id.split("#")[ |
|
0 |
|
] |
|
if doc_id not in doc_score_dict: |
|
doc_score_dict[doc_id] = ( |
|
-1 |
|
) |
|
if score > doc_score_dict[doc_id]: |
|
doc_score_dict[doc_id] = score |
|
top_docs = sorted(doc_score_dict.items(), key=lambda x: x[1], reverse=True)[ |
|
:top_n |
|
] |
|
run[query_id] = { |
|
doc_id: round(score * 100, 2) for doc_id, score in top_docs |
|
} |
|
scores = EvaluateRetrieval.evaluate( |
|
qrels, {query_id: run[query_id]}, k_values=[1, 3, 5, 10, 100, 1000] |
|
) |
|
scores = {k: v for d in scores for k, v in d.items()} |
|
per_query.append(scores[metric]) |
|
print("Done merging and sorting results, Evaluating and saving run") |
|
print(f"There are {len(run)} queries being evaled agaisnt qrels") |
|
print(f"There were {len(shard_files)} shards found") |
|
print( |
|
f"Per Query Score average: {np.array(per_query).mean()} for {metric}. Individual scores{per_query}" |
|
) |
|
print("Overall Score Numbers:") |
|
print(EvaluateRetrieval.evaluate(qrels, run, k_values=[1, 3, 5, 10, 100, 1000])) |
|
with open(output_file, "wb") as w: |
|
pickle.dump(run, w) |
|
|
|
|
|
if __name__ == "__main__": |
|
suffix = sys.argv[1] |
|
output_file = sys.argv[2] |
|
top_n = int(sys.argv[3]) |
|
qrel_filename = sys.argv[4] |
|
metric = sys.argv[5] |
|
merge_retrieved_shards( |
|
suffix, output_file, top_n, load_qrels(qrel_filename), metric |
|
) |
|
|