Datasets:
Tasks:
Question Answering
Modalities:
Text
Formats:
parquet
Languages:
English
Size:
10M - 100M
DOI:
| import glob | |
| import json | |
| import pickle | |
| import sys | |
| from typing import Dict | |
| import numpy as np | |
| from beir.retrieval.evaluation import EvaluateRetrieval | |
| def load_qrels(filename: str) -> Dict: | |
| with open(filename, "r") as f: | |
| qrels = json.load(f) | |
| return qrels | |
| def merge_retrieved_shards( | |
| suffix: str, output_file: str, top_n: int, qrels: dict, metric: str | |
| ) -> None: | |
| shard_files = glob.glob(f"*{suffix}") | |
| print(f"There are {len(shard_files)} shards found") | |
| merged_results = {} | |
| print("Loading All shards") | |
| for shard_file in shard_files: | |
| print(f"Loading shard {shard_file} ") | |
| with open(shard_file, "rb") as f: | |
| shard_results = pickle.load(f) | |
| for query_id, doc_scores in shard_results.items(): | |
| if query_id not in merged_results: | |
| merged_results[query_id] = [] | |
| merged_results[query_id].extend(doc_scores.items()) | |
| print("Shards all loaded, merging results and sorting by score") | |
| run = {} | |
| per_query = [] | |
| for query_id, doc_scores in merged_results.items(): | |
| if query_id in qrels: | |
| doc_score_dict = {} | |
| for passage_id, score in doc_scores: | |
| doc_id = passage_id.split("#")[ | |
| 0 | |
| ] # everything after # is the passage idenfitier withing a doc | |
| if doc_id not in doc_score_dict: | |
| doc_score_dict[doc_id] = ( | |
| -1 | |
| ) # scores are in range -1 to 1 on similairty so starting at -1 is floor | |
| if score > doc_score_dict[doc_id]: | |
| doc_score_dict[doc_id] = score | |
| top_docs = sorted(doc_score_dict.items(), key=lambda x: x[1], reverse=True)[ | |
| :top_n | |
| ] | |
| run[query_id] = { | |
| doc_id: round(score * 100, 2) for doc_id, score in top_docs | |
| } | |
| scores = EvaluateRetrieval.evaluate( | |
| qrels, {query_id: run[query_id]}, k_values=[1, 3, 5, 10, 100, 1000] | |
| ) | |
| scores = {k: v for d in scores for k, v in d.items()} | |
| per_query.append(scores[metric]) | |
| print("Done merging and sorting results, Evaluating and saving run") | |
| print(f"There are {len(run)} queries being evaled agaisnt qrels") | |
| print(f"There were {len(shard_files)} shards found") | |
| print( | |
| f"Per Query Score average: {np.array(per_query).mean()} for {metric}. Individual scores{per_query}" | |
| ) | |
| print("Overall Score Numbers:") | |
| print(EvaluateRetrieval.evaluate(qrels, run, k_values=[1, 3, 5, 10, 100, 1000])) | |
| with open(output_file, "wb") as w: | |
| pickle.dump(run, w) | |
| if __name__ == "__main__": | |
| suffix = sys.argv[1] | |
| output_file = sys.argv[2] | |
| top_n = int(sys.argv[3]) | |
| qrel_filename = sys.argv[4] | |
| metric = sys.argv[5] | |
| merge_retrieved_shards( | |
| suffix, output_file, top_n, load_qrels(qrel_filename), metric | |
| ) | |