mt5-small-koquad-qg-ae / eval /metric.first.answer.paragraph.questions_answers.lmqg_qg_koquad.default.json
asahi417's picture
add model
fc95fde
raw
history blame
1.17 kB
{"test": {"QAAlignedF1Score (BERTScore)": 0.803637745157127, "QAAlignedF1Score (MoverScore)": 0.8254996927301479, "QAAlignedRecall (BERTScore)": 0.8372135478464777, "QAAlignedPrecision (BERTScore)": 0.7734194841249756, "QAAlignedRecall (MoverScore)": 0.8668749372525869, "QAAlignedPrecision (MoverScore)": 0.7892838716044577, "Bleu_1": 0.08340644619085982, "Bleu_2": 0.043990424009451114, "Bleu_3": 0.020996488514879928, "Bleu_4": 0.010062527867853471, "METEOR": 0.2379032739551809, "ROUGE_L": 0.13215524741857645, "BERTScore": 0.6810509214552605, "MoverScore": 0.6267358673160498}, "validation": {"QAAlignedF1Score (BERTScore)": 0.8250852343448536, "QAAlignedF1Score (MoverScore)": 0.8663589339370455, "QAAlignedRecall (BERTScore)": 0.8329555714980513, "QAAlignedPrecision (BERTScore)": 0.8178138925314775, "QAAlignedRecall (MoverScore)": 0.8694045750397306, "QAAlignedPrecision (MoverScore)": 0.864580974204573, "Bleu_1": 0.26682086807810895, "Bleu_2": 0.17437393235936183, "Bleu_3": 0.10650300451687317, "Bleu_4": 0.06221188221815036, "METEOR": 0.2773174359340388, "ROUGE_L": 0.25238507797989873, "BERTScore": 0.7709162309765816, "MoverScore": 0.6946181750327773}}