Muennighoff's picture
Add
0461a88
{"GEM/web_nlg_en": {"PALM_prompt": {"bleu": 0.3399554042913014, "bleu_stderr": 0.021913090586979395, "rouge1_fmeasure": 0.12623006978068804, "rouge1_fmeasure_stderr": 0.0031705923901075954, "rouge1_precision": 0.1064118398300954, "rouge1_precision_stderr": 0.0038414579273568536, "rouge1_recall": 0.273886737862104, "rouge1_recall_stderr": 0.00471386391904672, "rouge2_fmeasure": 0.06068826179342809, "rouge2_fmeasure_stderr": 0.002003189981417211, "rouge2_precision": 0.05067898489188858, "rouge2_precision_stderr": 0.0024123940717715427, "rouge2_recall": 0.13547995085592274, "rouge2_recall_stderr": 0.0032418045928431787, "rougeL_fmeasure": 0.11646280067871276, "rougeL_fmeasure_stderr": 0.0027772340366833993, "rougeL_precision": 0.09728375654596867, "rougeL_precision_stderr": 0.0034844921413732634, "rougeL_recall": 0.26054477451420094, "rougeL_recall_stderr": 0.00441271413946595, "rougeLsum_fmeasure": 0.11815946956931984, "rougeLsum_fmeasure_stderr": 0.0028493215343763714, "rougeLsum_precision": 0.09898873489552214, "rougeLsum_precision_stderr": 0.003542580332029638, "rougeLsum_recall": 0.2624167917938274, "rougeLsum_recall_stderr": 0.004457065044596146}}, "GEM/wiki_lingua_en": {"tldr_en": {"bleu": 1.7401870910957447, "bleu_stderr": 0.07175788054485463, "rouge1_fmeasure": 0.1393524447467033, "rouge1_fmeasure_stderr": 0.001855118612807439, "rouge1_precision": 0.18114650597975454, "rouge1_precision_stderr": 0.00327745607663669, "rouge1_recall": 0.1637667976395018, "rouge1_recall_stderr": 0.002532304118484624, "rouge2_fmeasure": 0.025330265068893403, "rouge2_fmeasure_stderr": 0.0009117247702535065, "rouge2_precision": 0.037672429038420364, "rouge2_precision_stderr": 0.0018618303357091047, "rouge2_recall": 0.03066817368292832, "rouge2_recall_stderr": 0.0011767896068586182, "rougeL_fmeasure": 0.11187413367216702, "rougeL_fmeasure_stderr": 0.0014072319901725367, "rougeL_precision": 0.14782023430354443, "rougeL_precision_stderr": 0.002765132022432345, "rougeL_recall": 0.13265464091360485, "rougeL_recall_stderr": 0.0020176301222716514, "rougeLsum_fmeasure": 0.13126765225635678, "rougeLsum_fmeasure_stderr": 0.0017254074201957115, "rougeLsum_precision": 0.1715016782543171, "rougeLsum_precision_stderr": 0.0031279778718448026, "rougeLsum_recall": 0.1542782441420878, "rougeLsum_recall_stderr": 0.0023682583387763466}}, "e2e_nlg_cleaned": {"generate_text_restaurant": {"bleu": 11.54760180528191, "bleu_stderr": 0.12629468298438476, "rouge1_fmeasure": 0.4421097614352985, "rouge1_fmeasure_stderr": 0.002282737379789846, "rouge1_precision": 0.5397268443122286, "rouge1_precision_stderr": 0.003393195787516768, "rouge1_recall": 0.41415860157369805, "rouge1_recall_stderr": 0.0028548390468315908, "rouge2_fmeasure": 0.20986927598013522, "rouge2_fmeasure_stderr": 0.0019440451557575739, "rouge2_precision": 0.2614812386508731, "rouge2_precision_stderr": 0.0027921317453072237, "rouge2_recall": 0.19645932238490438, "rouge2_recall_stderr": 0.0020699890947959176, "rougeL_fmeasure": 0.3267951334508187, "rougeL_fmeasure_stderr": 0.0020556182684373743, "rougeL_precision": 0.40204866842786213, "rougeL_precision_stderr": 0.0031275125023726125, "rougeL_recall": 0.3053432788698131, "rougeL_recall_stderr": 0.00236397081046382, "rougeLsum_fmeasure": 0.36552104077057357, "rougeLsum_fmeasure_stderr": 0.002246477476655884, "rougeLsum_precision": 0.4478770395828019, "rougeLsum_precision_stderr": 0.0033018789954156417, "rougeLsum_recall": 0.34170718287246327, "rougeLsum_recall_stderr": 0.002607047016378604}}, "gem_xsum": {"article_DOC_summary": {"bleu": 0.6432031193336873, "bleu_stderr": 0.07008161728396729, "rouge1_fmeasure": 0.1523418212904453, "rouge1_fmeasure_stderr": 0.0024290958510217245, "rouge1_precision": 0.15068748817746397, "rouge1_precision_stderr": 0.00288941715614469, "rouge1_recall": 0.18208291562939338, "rouge1_recall_stderr": 0.0032053536019781602, "rouge2_fmeasure": 0.015779419211779257, "rouge2_fmeasure_stderr": 0.0010453102023582498, "rouge2_precision": 0.015604618217659207, "rouge2_precision_stderr": 0.0010963410848022987, "rouge2_recall": 0.01896571125082099, "rouge2_recall_stderr": 0.0012485415757762257, "rougeL_fmeasure": 0.11757005961375068, "rougeL_fmeasure_stderr": 0.0018127903253345678, "rougeL_precision": 0.11627018885316699, "rougeL_precision_stderr": 0.00220455722450916, "rougeL_recall": 0.14096528003639522, "rougeL_recall_stderr": 0.0023945217274176397, "rougeLsum_fmeasure": 0.11944930450900756, "rougeLsum_fmeasure_stderr": 0.0018754754039972382, "rougeLsum_precision": 0.11785580405530326, "rougeLsum_precision_stderr": 0.0022307252764576448, "rougeLsum_recall": 0.1438492678290832, "rougeLsum_recall_stderr": 0.002592501693406087}}}