KashiwaByte's picture
initial commit
7c93e9d
# -*- coding: utf-8 -*-
import utils.utils
import utils
# Key for wikipedia eval is question-id. Key for web eval is the (question_id, filename) tuple
def get_key_to_ground_truth(data):
if data['Domain'] == 'Wikipedia':
return {datum['QuestionId']: datum['Answer'] for datum in data['Data']}
else:
return get_qd_to_answer(data)
def get_question_doc_string(qid, doc_name):
return '{}--{}'.format(qid, doc_name)
def get_qd_to_answer(data):
key_to_answer = {}
for datum in data['Data']:
for page in datum.get('EntityPages', []) + datum.get('SearchResults', []):
qd_tuple = get_question_doc_string(datum['QuestionId'], page['Filename'])
key_to_answer[qd_tuple] = datum['Answer']
return key_to_answer
def read_clean_part(datum):
for key in ['EntityPages', 'SearchResults']:
new_page_list = []
for page in datum.get(key, []):
if page['DocPartOfVerifiedEval']:
new_page_list.append(page)
datum[key] = new_page_list
assert len(datum['EntityPages']) + len(datum['SearchResults']) > 0
return datum
def read_triviaqa_data(qajson):
data = utils.utils.read_json(qajson)
# read only documents and questions that are a part of clean data set
if data['VerifiedEval']:
clean_data = []
for datum in data['Data']:
if datum['QuestionPartOfVerifiedEval']:
if data['Domain'] == 'Web':
datum = read_clean_part(datum)
clean_data.append(datum)
data['Data'] = clean_data
return data
def answer_index_in_document(answer, document):
answer_list = answer['NormalizedAliases']
for answer_string_in_doc in answer_list:
index = document.lower().find(answer_string_in_doc)
if index != -1:
return answer_string_in_doc, index
return answer['NormalizedValue'], -1