mt5-base-ruquad-qag / eval /metric.first.answer.paragraph.questions_answers.lmqg_qag_ruquad.default.json
asahi417's picture
model update
248ea3d
raw
history blame
1.17 kB
{"validation": {"Bleu_1": 0.0727623029781301, "Bleu_2": 0.04124685733407366, "Bleu_3": 0.027560417651719207, "Bleu_4": 0.019835619588151256, "METEOR": 0.16482107780717625, "ROUGE_L": 0.1276036256543389, "BERTScore": 0.620910641857151, "MoverScore": 0.5054021740615375, "QAAlignedF1Score (BERTScore)": 0.7425083894545008, "QAAlignedRecall (BERTScore)": 0.7496717520922123, "QAAlignedPrecision (BERTScore)": 0.7362779845211975, "QAAlignedF1Score (MoverScore)": 0.5408001938000293, "QAAlignedRecall (MoverScore)": 0.5447353036490986, "QAAlignedPrecision (MoverScore)": 0.5375149386697426}, "test": {"Bleu_1": 0.0751331450834612, "Bleu_2": 0.043327470712133796, "Bleu_3": 0.029211298089606604, "Bleu_4": 0.02122018857138504, "METEOR": 0.16845591256816533, "ROUGE_L": 0.1312041896872967, "BERTScore": 0.6230435508206201, "MoverScore": 0.5058324046578218, "QAAlignedF1Score (BERTScore)": 0.7462666300251022, "QAAlignedRecall (BERTScore)": 0.7537636939764181, "QAAlignedPrecision (BERTScore)": 0.7397488655371148, "QAAlignedF1Score (MoverScore)": 0.5424477310159134, "QAAlignedRecall (MoverScore)": 0.5464948672511275, "QAAlignedPrecision (MoverScore)": 0.5391184382186318}}