{"GEM/web_nlg_en": {"PALM_prompt": {"bleu": 0.38416942085325156, "bleu_stderr": 0.030367676473967677, "rouge1_fmeasure": 0.11951078699124636, "rouge1_fmeasure_stderr": 0.0030915406018018235, "rouge1_precision": 0.10036919084075267, "rouge1_precision_stderr": 0.0036423589096868324, "rouge1_recall": 0.26152155839841046, "rouge1_recall_stderr": 0.0044155091232318065, "rouge2_fmeasure": 0.05615347349900199, "rouge2_fmeasure_stderr": 0.0018782445999452166, "rouge2_precision": 0.04765565829660879, "rouge2_precision_stderr": 0.0022450697818222833, "rouge2_recall": 0.12717458010109978, "rouge2_recall_stderr": 0.0030757331493193527, "rougeL_fmeasure": 0.1093973401823277, "rougeL_fmeasure_stderr": 0.002628037936748325, "rougeL_precision": 0.09011621525155329, "rougeL_precision_stderr": 0.0031300827075074673, "rougeL_recall": 0.24790058767911233, "rougeL_recall_stderr": 0.004096430077329352, "rougeLsum_fmeasure": 0.11104611706091866, "rougeLsum_fmeasure_stderr": 0.0027315849851970936, "rougeLsum_precision": 0.09230129034589184, "rougeLsum_precision_stderr": 0.0032698936535295723, "rougeLsum_recall": 0.24883823788563983, "rougeLsum_recall_stderr": 0.004109106450249802}}, "e2e_nlg_cleaned": {"generate_text_restaurant": {"bleu": 10.981823507738188, "bleu_stderr": 0.2079037039330707, "rouge1_fmeasure": 0.4107607953917595, "rouge1_fmeasure_stderr": 0.0022344711572428997, "rouge1_precision": 0.4763105984270431, "rouge1_precision_stderr": 0.003087563680845141, "rouge1_recall": 0.39730063402648225, "rouge1_recall_stderr": 0.002745546116932068, "rouge2_fmeasure": 0.18649090753049416, "rouge2_fmeasure_stderr": 0.0018901130188229557, "rouge2_precision": 0.2188562694975304, "rouge2_precision_stderr": 0.002436310647541356, "rouge2_recall": 0.18085187025511462, "rouge2_recall_stderr": 0.0020454955600626747, "rougeL_fmeasure": 0.3130192687415252, "rougeL_fmeasure_stderr": 0.0019507309739372827, "rougeL_precision": 0.3645846077297693, "rougeL_precision_stderr": 0.0027236658829662506, "rougeL_recall": 0.30272697220116324, "rougeL_recall_stderr": 0.002324460364527574, "rougeLsum_fmeasure": 0.3502676851764353, "rougeLsum_fmeasure_stderr": 0.002198598848985128, "rougeLsum_precision": 0.4068459758429193, "rougeLsum_precision_stderr": 0.002975444916688226, "rougeLsum_recall": 0.33873681252030896, "rougeLsum_recall_stderr": 0.0025865465694752143}}}