mt5-base-itquad-qag / eval /metric.first.answer.paragraph.questions_answers.lmqg_qag_itquad.default.json
asahi417's picture
model update
0e76951
raw
history blame
1.17 kB
{"validation": {"Bleu_1": 0.07700925095636735, "Bleu_2": 0.0324774788907088, "Bleu_3": 0.016317013498405107, "Bleu_4": 0.009242194222026984, "METEOR": 0.14956911772409914, "ROUGE_L": 0.1391450868379757, "BERTScore": 0.6686975762295495, "MoverScore": 0.5060724864937646, "QAAlignedF1Score (BERTScore)": 0.7891348884343872, "QAAlignedRecall (BERTScore)": 0.7941148703167266, "QAAlignedPrecision (BERTScore)": 0.7845570760401714, "QAAlignedF1Score (MoverScore)": 0.5313231229326516, "QAAlignedRecall (MoverScore)": 0.5334536839757478, "QAAlignedPrecision (MoverScore)": 0.5293434889401952}, "test": {"Bleu_1": 0.21167852578781593, "Bleu_2": 0.11374335018536964, "Bleu_3": 0.05833805478144616, "Bleu_4": 0.033897654833535566, "METEOR": 0.15092571430694843, "ROUGE_L": 0.19120088449317582, "BERTScore": 0.7336431266346448, "MoverScore": 0.5163791613690608, "QAAlignedF1Score (BERTScore)": 0.7992503859263245, "QAAlignedRecall (BERTScore)": 0.7887337241402319, "QAAlignedPrecision (BERTScore)": 0.8106036093210574, "QAAlignedF1Score (MoverScore)": 0.5379928571174765, "QAAlignedRecall (MoverScore)": 0.5301687067521811, "QAAlignedPrecision (MoverScore)": 0.5463839520366399}}