Spaces:
Runtime error
Runtime error
# | |
# Pyserini: Reproducible IR research with sparse and dense representations | |
# | |
# Licensed under the Apache License, Version 2.0 (the "License"); | |
# you may not use this file except in compliance with the License. | |
# You may obtain a copy of the License at | |
# | |
# http://www.apache.org/licenses/LICENSE-2.0 | |
# | |
# Unless required by applicable law or agreed to in writing, software | |
# distributed under the License is distributed on an "AS IS" BASIS, | |
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |
# See the License for the specific language governing permissions and | |
# limitations under the License. | |
# | |
import argparse | |
import json | |
import os | |
import numpy as np | |
from tqdm import tqdm | |
if __name__ == '__main__': | |
parser = argparse.ArgumentParser(description='Interpolate runs') | |
parser.add_argument('--run1', required=True, help='retrieval run1') | |
parser.add_argument('--run2', required=True, help='retrieval run2') | |
parser.add_argument('--start-weight', type=float, required=True, help='start hybrid alpha') | |
parser.add_argument('--end-weight', type=float, required=True, help='end hybrid alpha') | |
parser.add_argument('--step', type=float, required=True, help='changes of alpha per step') | |
parser.add_argument('--output-dir', required=True, help='hybrid result') | |
args = parser.parse_args() | |
if not os.path.exists(args.output_dir): | |
os.makedirs(args.output_dir) | |
for alpha in np.arange(args.start_weight, args.end_weight, args.step): | |
run1_result = json.load(open(args.run1)) | |
run2_result = json.load(open(args.run2)) | |
hybrid_result = {} | |
for key in tqdm(list(run1_result.keys())): | |
question = run1_result[key]['question'] | |
answers = run1_result[key]['answers'] | |
run2_contexts = run2_result[key]['contexts'] | |
run1_contexts = run1_result[key]['contexts'] | |
run1_hits = {hit['docid']: float(hit['score']) for hit in run1_contexts} | |
run2_hits = {hit['docid']: float(hit['score']) for hit in run2_contexts} | |
hybrid_scores = {} | |
run1_scores = {} | |
run2_scores = {} | |
min_run1_score = min(run1_hits.values()) | |
min_run2_score = min(run2_hits.values()) | |
for doc in set(run1_hits.keys()) | set(run2_hits.keys()): | |
if doc not in run1_hits: | |
score = alpha * run2_hits[doc] + min_run1_score | |
run2_scores[doc] = run2_hits[doc] | |
run1_scores[doc] = -1 | |
elif doc not in run2_hits: | |
score = alpha * min_run2_score + run1_hits[doc] | |
run2_scores[doc] = -1 | |
run1_scores[doc] = run1_hits[doc] | |
else: | |
score = alpha * run2_hits[doc] + run1_hits[doc] | |
run2_scores[doc] = run2_hits[doc] | |
run1_scores[doc] = run1_hits[doc] | |
hybrid_scores[doc] = score | |
total_ids = [] | |
total_context = [] | |
for sctx, dctx in zip(run2_contexts, run1_contexts): | |
if sctx['docid'] not in total_ids: | |
total_ids.append(sctx['docid']) | |
sctx['score'] = hybrid_scores[sctx['docid']] | |
sctx['run2_score'] = run2_scores[sctx['docid']] | |
sctx['run1_score'] = run1_scores[sctx['docid']] | |
total_context.append(sctx) | |
if dctx['docid'] not in total_ids: | |
total_ids.append(dctx['docid']) | |
dctx['score'] = hybrid_scores[dctx['docid']] | |
dctx['run2_score'] = run2_scores[dctx['docid']] | |
dctx['run1_score'] = run1_scores[dctx['docid']] | |
total_context.append(dctx) | |
total_context = sorted(total_context, key=lambda x: x['score'], reverse=True) | |
hybrid_result[key] = {'question': question, 'answers': answers, 'contexts': total_context} | |
json.dump(hybrid_result, open(os.path.join(args.output_dir, f'run_fused_weight_{alpha}.json'), 'w'), indent=4) | |