Muennighoff's picture
Add
0f79083
{"GEM/web_nlg_en": {"PALM_prompt": {"bleu": 0.2334282280458586, "bleu_stderr": 0.019052307328308617, "rouge1_fmeasure": 0.08900921610644012, "rouge1_fmeasure_stderr": 0.0020657069941074835, "rouge1_precision": 0.06664206812457203, "rouge1_precision_stderr": 0.0022521713907757224, "rouge1_recall": 0.22011552548283286, "rouge1_recall_stderr": 0.00410978487191024, "rouge2_fmeasure": 0.03844075075586678, "rouge2_fmeasure_stderr": 0.0011778262164477038, "rouge2_precision": 0.02742604267640331, "rouge2_precision_stderr": 0.0012236389813676137, "rouge2_recall": 0.10166270849980795, "rouge2_recall_stderr": 0.0026960826428279448, "rougeL_fmeasure": 0.0844769580786778, "rougeL_fmeasure_stderr": 0.0018659122044865008, "rougeL_precision": 0.06259979613562147, "rougeL_precision_stderr": 0.0020549501026142056, "rougeL_recall": 0.21306622153871427, "rougeL_recall_stderr": 0.003978348602912166, "rougeLsum_fmeasure": 0.0845681547310993, "rougeLsum_fmeasure_stderr": 0.001889837918272996, "rougeLsum_precision": 0.06293334235936807, "rougeLsum_precision_stderr": 0.002091916854143351, "rougeLsum_recall": 0.21239997123355958, "rougeLsum_recall_stderr": 0.003944956006147718}}, "GEM/wiki_lingua_en": {"tldr_en": {"bleu": 1.1784483013564102, "bleu_stderr": 0.06904095646323807, "rouge1_fmeasure": 0.12041833327196463, "rouge1_fmeasure_stderr": 0.001596604081334374, "rouge1_precision": 0.13646669155030028, "rouge1_precision_stderr": 0.00220310186554878, "rouge1_recall": 0.14546845977536896, "rouge1_recall_stderr": 0.002185736272211402, "rouge2_fmeasure": 0.01499749394448183, "rouge2_fmeasure_stderr": 0.000625100665952798, "rouge2_precision": 0.01724386477933621, "rouge2_precision_stderr": 0.0008686647411341404, "rouge2_recall": 0.01912476010235633, "rouge2_recall_stderr": 0.000862139084747869, "rougeL_fmeasure": 0.09741475108454614, "rougeL_fmeasure_stderr": 0.001214366654297324, "rougeL_precision": 0.11051635705738314, "rougeL_precision_stderr": 0.001756238864615761, "rougeL_recall": 0.11966135440554557, "rougeL_recall_stderr": 0.0017926753508674628, "rougeLsum_fmeasure": 0.11380484706597256, "rougeLsum_fmeasure_stderr": 0.001488577909873337, "rougeLsum_precision": 0.12908957637877216, "rougeLsum_precision_stderr": 0.0020679024538660115, "rougeLsum_recall": 0.13765696723373635, "rougeLsum_recall_stderr": 0.0020581929100874927}}, "e2e_nlg_cleaned": {"generate_text_restaurant": {"bleu": 7.633962606953475, "bleu_stderr": 0.11159543603672754, "rouge1_fmeasure": 0.356668169483382, "rouge1_fmeasure_stderr": 0.0021607608835383096, "rouge1_precision": 0.4202208020040644, "rouge1_precision_stderr": 0.003044005246951365, "rouge1_recall": 0.3432310437592059, "rouge1_recall_stderr": 0.0025043229759053897, "rouge2_fmeasure": 0.14338933259648326, "rouge2_fmeasure_stderr": 0.0016391755775913735, "rouge2_precision": 0.17155302692760083, "rouge2_precision_stderr": 0.0021420775665914075, "rouge2_recall": 0.1381486339251807, "rouge2_recall_stderr": 0.0017426260823225216, "rougeL_fmeasure": 0.2659375155003453, "rougeL_fmeasure_stderr": 0.0017894307181340716, "rougeL_precision": 0.3143755923201533, "rougeL_precision_stderr": 0.002563069353077956, "rougeL_recall": 0.25646642804127656, "rougeL_recall_stderr": 0.0020573152371357105, "rougeLsum_fmeasure": 0.29672587998555633, "rougeLsum_fmeasure_stderr": 0.002046175837266347, "rougeLsum_precision": 0.349974374493684, "rougeLsum_precision_stderr": 0.0028149623026241734, "rougeLsum_recall": 0.2856269773413544, "rougeLsum_recall_stderr": 0.002309169188218799}}, "gem_xsum": {"article_DOC_summary": {"bleu": 0.8946101579944589, "bleu_stderr": 0.11447862558731768, "rouge1_fmeasure": 0.1474750330894204, "rouge1_fmeasure_stderr": 0.002352105574798546, "rouge1_precision": 0.12823223548348878, "rouge1_precision_stderr": 0.0025565036587658725, "rouge1_recall": 0.2143351214598643, "rouge1_recall_stderr": 0.0036128802474860813, "rouge2_fmeasure": 0.02113796673167594, "rouge2_fmeasure_stderr": 0.0011185423587926665, "rouge2_precision": 0.017523161513321936, "rouge2_precision_stderr": 0.0009922032751105664, "rouge2_recall": 0.03274318045581887, "rouge2_recall_stderr": 0.0017433402857737756, "rougeL_fmeasure": 0.12005256994367001, "rougeL_fmeasure_stderr": 0.0017848482742927902, "rougeL_precision": 0.10370089292267248, "rougeL_precision_stderr": 0.001970850041369119, "rougeL_recall": 0.17671366292027554, "rougeL_recall_stderr": 0.0028713475720421144, "rougeLsum_fmeasure": 0.11549675473891775, "rougeLsum_fmeasure_stderr": 0.0018543831002685392, "rougeLsum_precision": 0.10049270031173661, "rougeLsum_precision_stderr": 0.002030546802009858, "rougeLsum_recall": 0.16875347685809536, "rougeLsum_recall_stderr": 0.002920858053558713}}}