{ "results": { "polish": { "acc,none": 0.38655822884426444, "acc_stderr,none": 0.009810522827733922, "acc_norm,none": 0.3877248155817996, "acc_norm_stderr,none": 0.009903004691484696, "f1,score-first": 0.5390263032489272, "f1_stderr,score-first": 0.09713328442043558, "accuracy,score-first": 0.5390263032489272, "accuracy_stderr,score-first": 0.09713328442043558, "alias": "polish" }, "belebele_pol_Latn": { "acc,none": 0.39555555555555555, "acc_stderr,none": 0.016308052223501762, "acc_norm,none": 0.39555555555555555, "acc_norm_stderr,none": 0.016308052223501762, "alias": " - belebele_pol_Latn" }, "polemo2_in": { "f1,score-first": 0.4695290858725762, "f1_stderr,score-first": 0.018586362627404755, "accuracy,score-first": 0.4695290858725762, "accuracy_stderr,score-first": 0.018586362627404755, "alias": " - polemo2_in" }, "polemo2_in_multiple_choice": { "acc,none": 0.16204986149584488, "acc_stderr,none": 0.013723528490778158, "acc_norm,none": 0.16204986149584488, "acc_norm_stderr,none": 0.013723528490778158, "alias": " - polemo2_in_multiple_choice" }, "polemo2_out": { "f1,score-first": 0.4473684210526316, "f1_stderr,score-first": 0.02239376166833581, "accuracy,score-first": 0.4473684210526316, "accuracy_stderr,score-first": 0.02239376166833581, "alias": " - polemo2_out" }, "polemo2_out_multiple_choice": { "acc,none": 0.0020242914979757085, "acc_stderr,none": 0.0020242914979757094, "acc_norm,none": 0.0020242914979757085, "acc_norm_stderr,none": 0.0020242914979757094, "alias": " - polemo2_out_multiple_choice" }, "polish_8tags_multiple_choice": { "acc,none": 0.5121225983531564, "acc_stderr,none": 0.0075605243767434, "acc_norm,none": 0.5146386093321135, "acc_norm_stderr,none": 0.007559505586488673, "alias": " - polish_8tags_multiple_choice" }, "polish_8tags_regex": { "f1,score-first": 0.6697163769441903, "f1_stderr,score-first": 0.007113749634252443, "accuracy,score-first": 0.6697163769441903, "accuracy_stderr,score-first": 0.007113749634252443, "alias": " - polish_8tags_regex" }, "polish_belebele_regex": { "f1,score-first": 0.44333333333333336, "f1_stderr,score-first": 0.0165684909720556, "accuracy,score-first": 0.44333333333333336, "accuracy_stderr,score-first": 0.0165684909720556, "alias": " - polish_belebele_regex" }, "polish_dyk_multiple_choice": { "acc,none": 0.2118561710398445, "acc_stderr,none": 0.012744615626342343, "acc_norm,none": 0.2118561710398445, "acc_norm_stderr,none": 0.012744615626342343, "alias": " - polish_dyk_multiple_choice" }, "polish_dyk_regex": { "f1,score-first": 0.8289601554907677, "f1_stderr,score-first": 0.011744077740056324, "accuracy,score-first": 0.8289601554907677, "accuracy_stderr,score-first": 0.011744077740056324, "alias": " - polish_dyk_regex" }, "polish_ppc_multiple_choice": { "acc,none": 0.334, "acc_stderr,none": 0.014922019523732963, "acc_norm,none": 0.334, "acc_norm_stderr,none": 0.014922019523732963, "alias": " - polish_ppc_multiple_choice" }, "polish_ppc_regex": { "f1,score-first": 0.421, "f1_stderr,score-first": 0.015620595475301318, "accuracy,score-first": 0.421, "accuracy_stderr,score-first": 0.015620595475301318, "alias": " - polish_ppc_regex" }, "polish_psc_multiple_choice": { "acc,none": 0.3070500927643785, "acc_stderr,none": 0.014055544850266423, "acc_norm,none": 0.3070500927643785, "acc_norm_stderr,none": 0.014055544850266423, "alias": " - polish_psc_multiple_choice" }, "polish_psc_regex": { "f1,score-first": 0.4322820037105751, "f1_stderr,score-first": 0.0150953107201147, "accuracy,score-first": 0.4322820037105751, "accuracy_stderr,score-first": 0.0150953107201147, "alias": " - polish_psc_regex" } }, "groups": { "polish": { "acc,none": 0.38655822884426444, "acc_stderr,none": 0.009810522827733922, "acc_norm,none": 0.3877248155817996, "acc_norm_stderr,none": 0.009903004691484696, "f1,score-first": 0.5390263032489272, "f1_stderr,score-first": 0.09713328442043558, "accuracy,score-first": 0.5390263032489272, "accuracy_stderr,score-first": 0.09713328442043558, "alias": "polish" } }, "configs": { "belebele_pol_Latn": { "task": "belebele_pol_Latn", "group": "belebele", "dataset_path": "facebook/belebele", "test_split": "pol_Latn", "fewshot_split": "pol_Latn", "doc_to_text": "P: {{flores_passage}}\nQ: {{question.strip()}}\nA: {{mc_answer1}}\nB: {{mc_answer2}}\nC: {{mc_answer3}}\nD: {{mc_answer4}}\nAnswer:", "doc_to_target": "{{['1', '2', '3', '4'].index(correct_answer_num)}}", "doc_to_choice": [ "A", "B", "C", "D" ], "description": "", "target_delimiter": " ", "fewshot_delimiter": "\n\n", "fewshot_config": { "sampler": "first_n" }, "num_fewshot": 0, "metric_list": [ { "metric": "acc", "aggregation": "mean", "higher_is_better": true }, { "metric": "acc_norm", "aggregation": "mean", "higher_is_better": true } ], "output_type": "multiple_choice", "repeats": 1, "should_decontaminate": true, "doc_to_decontamination_query": "{{question}}", "metadata": { "version": 0.0 } }, "polemo2_in": { "task": "polemo2_in", "group": [ "polemo2" ], "dataset_path": "allegro/klej-polemo2-in", "training_split": "train", "validation_split": "validation", "test_split": "test", "doc_to_text": "Opinia: \"{{sentence}}\"\nOkreśl sentyment podanej opinii. Możliwe odpowiedzi:\nA - Neutralny\nB - Negatywny\nC - Pozytywny\nD - Niejednoznaczny\nPrawidłowa odpowiedź:", "doc_to_target": "{{['__label__meta_zero', '__label__meta_minus_m', '__label__meta_plus_m', '__label__meta_amb'].index(target)}}", "description": "", "target_delimiter": " ", "fewshot_delimiter": "\n\n", "num_fewshot": 0, "metric_list": [ { "metric": "f1", "aggregation": "mean", "higher_is_better": true, "hf_evaluate": true, "average": "micro" }, { "metric": "accuracy", "aggregation": "mean", "higher_is_better": true } ], "output_type": "generate_until", "generation_kwargs": { "until": [ ".", "," ], "do_sample": false, "temperature": 0.0, "max_gen_toks": 50 }, "repeats": 1, "filter_list": [ { "name": "score-first", "filter": [ { "function": "regex", "regex_pattern": "(\\b[ABCD]\\b)" }, { "function": "take_first" }, { "function": "map", "mapping_dict": { "A": 0, "B": 1, "C": 2, "D": 3 }, "default_value": -1 }, { "function": "take_first" } ] } ], "should_decontaminate": true, "doc_to_decontamination_query": "{{sentence}}", "metadata": { "version": 1.0 } }, "polemo2_in_multiple_choice": { "task": "polemo2_in_multiple_choice", "group": [ "polemo2_mc" ], "dataset_path": "allegro/klej-polemo2-in", "training_split": "train", "validation_split": "validation", "test_split": "test", "doc_to_text": "Opinia: \"{{sentence}}\"\nOkreśl sentyment podanej opinii: Neutralny, Negatywny, Pozytywny, Niejednoznaczny.\nSentyment:", "doc_to_target": "{{['__label__meta_zero', '__label__meta_minus_m', '__label__meta_plus_m', '__label__meta_amb'].index(target)}}", "doc_to_choice": [ "Neutralny", "Negatywny", "Pozytywny", "Niejednoznaczny" ], "description": "", "target_delimiter": " ", "fewshot_delimiter": "\n\n", "num_fewshot": 0, "metric_list": [ { "metric": "acc", "aggregation": "mean", "higher_is_better": true }, { "metric": "acc_norm", "aggregation": "mean", "higher_is_better": true } ], "output_type": "multiple_choice", "repeats": 1, "should_decontaminate": true, "doc_to_decontamination_query": "{{sentence}}" }, "polemo2_out": { "task": "polemo2_out", "group": [ "polemo2" ], "dataset_path": "allegro/klej-polemo2-out", "training_split": "train", "validation_split": "validation", "test_split": "test", "doc_to_text": "Opinia: \"{{sentence}}\"\nOkreśl sentyment podanej opinii. Możliwe odpowiedzi:\nA - Neutralny\nB - Negatywny\nC - Pozytywny\nD - Niejednoznaczny\nPrawidłowa odpowiedź:", "doc_to_target": "{{['__label__meta_zero', '__label__meta_minus_m', '__label__meta_plus_m', '__label__meta_amb'].index(target)}}", "description": "", "target_delimiter": " ", "fewshot_delimiter": "\n\n", "num_fewshot": 0, "metric_list": [ { "metric": "f1", "aggregation": "mean", "higher_is_better": true, "hf_evaluate": true, "average": "micro" }, { "metric": "accuracy", "aggregation": "mean", "higher_is_better": true } ], "output_type": "generate_until", "generation_kwargs": { "until": [ ".", "," ], "do_sample": false, "temperature": 0.0, "max_gen_toks": 50 }, "repeats": 1, "filter_list": [ { "name": "score-first", "filter": [ { "function": "regex", "regex_pattern": "(\\b[ABCD]\\b)" }, { "function": "take_first" }, { "function": "map", "mapping_dict": { "A": 0, "B": 1, "C": 2, "D": 3 }, "default_value": -1 }, { "function": "take_first" } ] } ], "should_decontaminate": true, "doc_to_decontamination_query": "{{sentence}}", "metadata": { "version": 1.0 } }, "polemo2_out_multiple_choice": { "task": "polemo2_out_multiple_choice", "group": [ "polemo2_mc" ], "dataset_path": "allegro/klej-polemo2-out", "training_split": "train", "validation_split": "validation", "test_split": "test", "doc_to_text": "Opinia: \"{{sentence}}\"\nOkreśl sentyment podanej opinii: Neutralny, Negatywny, Pozytywny, Niejednoznaczny.\nSentyment:", "doc_to_target": "{{['__label__meta_zero', '__label__meta_minus_m', '__label__meta_plus_m', '__label__meta_amb'].index(target)}}", "doc_to_choice": [ "Neutralny", "Negatywny", "Pozytywny", "Niejednoznaczny" ], "description": "", "target_delimiter": " ", "fewshot_delimiter": "\n\n", "num_fewshot": 0, "metric_list": [ { "metric": "acc", "aggregation": "mean", "higher_is_better": true }, { "metric": "acc_norm", "aggregation": "mean", "higher_is_better": true } ], "output_type": "multiple_choice", "repeats": 1, "should_decontaminate": true, "doc_to_decontamination_query": "{{sentence}}" }, "polish_8tags_multiple_choice": { "task": "polish_8tags_multiple_choice", "dataset_path": "djstrong/8tags", "training_split": "train", "test_split": "test", "fewshot_split": "train", "doc_to_text": "Tytuł: \"{{sentence}}\"\nDo podanego tytułu przyporządkuj jedną najlepiej pasującą kategorię z podanych: Film, Historia, Jedzenie, Medycyna, Motoryzacja, Praca, Sport, Technologie.\nKategoria:", "doc_to_target": "{{label|int}}", "doc_to_choice": [ "Film", "Historia", "Jedzenie", "Medycyna", "Motoryzacja", "Praca", "Sport", "Technologie" ], "description": "", "target_delimiter": " ", "fewshot_delimiter": "\n\n", "num_fewshot": 0, "metric_list": [ { "metric": "acc", "aggregation": "mean", "higher_is_better": true }, { "metric": "acc_norm", "aggregation": "mean", "higher_is_better": true } ], "output_type": "multiple_choice", "repeats": 1, "should_decontaminate": true, "doc_to_decontamination_query": "{{sentence}}" }, "polish_8tags_regex": { "task": "polish_8tags_regex", "dataset_path": "djstrong/8tags", "training_split": "train", "validation_split": "validation", "test_split": "test", "doc_to_text": "Tytuł: \"{{sentence}}\"\nPytanie: jaka kategoria najlepiej pasuje do podanego tytułu?\nMożliwe odpowiedzi:\nA - film\nB - historia\nC - jedzenie\nD - medycyna\nE - motoryzacja\nF - praca\nG - sport\nH - technologie\nPrawidłowa odpowiedź:", "doc_to_target": "{{label|int}}", "description": "", "target_delimiter": " ", "fewshot_delimiter": "\n\n", "num_fewshot": 0, "metric_list": [ { "metric": "f1", "aggregation": "mean", "higher_is_better": true, "hf_evaluate": true, "average": "micro" }, { "metric": "accuracy", "aggregation": "mean", "higher_is_better": true } ], "output_type": "generate_until", "generation_kwargs": { "until": [ ".", "," ], "do_sample": false, "temperature": 0.0, "max_gen_toks": 50 }, "repeats": 1, "filter_list": [ { "name": "score-first", "filter": [ { "function": "regex", "regex_pattern": "(\\b[ABCDEFGH]\\b)" }, { "function": "take_first" }, { "function": "map", "mapping_dict": { "A": 0, "B": 1, "C": 2, "D": 3, "E": 4, "F": 5, "G": 6, "H": 7 }, "default_value": -1 }, { "function": "take_first" } ] } ], "should_decontaminate": true, "doc_to_decontamination_query": "{{sentence}}" }, "polish_belebele_regex": { "task": "polish_belebele_regex", "dataset_path": "facebook/belebele", "test_split": "pol_Latn", "doc_to_text": "Fragment: \"{{flores_passage}}\"\nPytanie: \"{{question}}\"\nMożliwe odpowiedzi:\nA - {{mc_answer1}}\nB - {{mc_answer2}}\nC - {{mc_answer3}}\nD - {{mc_answer4}}\nPrawidłowa odpowiedź:", "doc_to_target": "{{correct_answer_num|int - 1}}", "description": "", "target_delimiter": " ", "fewshot_delimiter": "\n\n", "num_fewshot": 0, "metric_list": [ { "metric": "f1", "aggregation": "mean", "higher_is_better": true, "hf_evaluate": true, "average": "micro" }, { "metric": "accuracy", "aggregation": "mean", "higher_is_better": true } ], "output_type": "generate_until", "generation_kwargs": { "until": [ ".", "," ], "do_sample": false, "temperature": 0.0, "max_gen_toks": 50 }, "repeats": 1, "filter_list": [ { "name": "score-first", "filter": [ { "function": "regex", "regex_pattern": "(\\b[ABCD]\\b)" }, { "function": "take_first" }, { "function": "map", "mapping_dict": { "A": 0, "B": 1, "C": 2, "D": 3 }, "default_value": -1 }, { "function": "take_first" } ] } ], "should_decontaminate": true, "doc_to_decontamination_query": "{{flores_passage}} {{question}} {{mc_answer1}} {{mc_answer2}} {{mc_answer3}} {{mc_answer4}}" }, "polish_dyk_multiple_choice": { "task": "polish_dyk_multiple_choice", "dataset_path": "allegro/klej-dyk", "training_split": "train", "test_split": "test", "doc_to_text": "Pytanie: \"{{question}}\"\nSugerowana odpowiedź: \"{{answer}}\"\nPytanie: Czy sugerowana odpowiedź na zadane pytanie jest poprawna?\nOdpowiedz krótko \"Tak\" lub \"Nie\". Prawidłowa odpowiedź:", "doc_to_target": "{{target|int}}", "doc_to_choice": [ "Nie", "Tak" ], "description": "", "target_delimiter": " ", "fewshot_delimiter": "\n\n", "num_fewshot": 0, "metric_list": [ { "metric": "acc", "aggregation": "mean", "higher_is_better": true }, { "metric": "acc_norm", "aggregation": "mean", "higher_is_better": true } ], "output_type": "multiple_choice", "repeats": 1, "should_decontaminate": true, "doc_to_decontamination_query": "{{question}} {{answer}}" }, "polish_dyk_regex": { "task": "polish_dyk_regex", "dataset_path": "allegro/klej-dyk", "training_split": "train", "test_split": "test", "doc_to_text": "Pytanie: \"{{question}}\"\nSugerowana odpowiedź: \"{{answer}}\"\nCzy sugerowana odpowiedź na zadane pytanie jest poprawna? Możliwe opcje:\nA - brakuje sugerowanej odpowiedzi\nB - nie, sugerowana odpowiedź nie jest poprawna\nC - tak, sugerowana odpowiedź jest poprawna\nD - brakuje pytania\nPrawidłowa opcja:", "doc_to_target": "{{target|int+1}}", "description": "", "target_delimiter": " ", "fewshot_delimiter": "\n\n", "num_fewshot": 0, "metric_list": [ { "metric": "f1", "aggregation": "mean", "higher_is_better": true, "hf_evaluate": true, "average": "micro" }, { "metric": "accuracy", "aggregation": "mean", "higher_is_better": true } ], "output_type": "generate_until", "generation_kwargs": { "until": [ ".", "," ], "do_sample": false, "temperature": 0.0, "max_gen_toks": 50 }, "repeats": 1, "filter_list": [ { "name": "score-first", "filter": [ { "function": "regex", "regex_pattern": "(\\b[ABCD]\\b)" }, { "function": "take_first" }, { "function": "map", "mapping_dict": { "A": 0, "B": 1, "C": 2, "D": 3 }, "default_value": -1 }, { "function": "take_first" } ] } ], "should_decontaminate": true, "doc_to_decontamination_query": "{{question}} {{answer}}" }, "polish_ppc_multiple_choice": { "task": "polish_ppc_multiple_choice", "dataset_path": "djstrong/ppc", "training_split": "train", "validation_split": "validation", "test_split": "test", "doc_to_text": "Zdanie A: \"{{sentence_A}}\"\nZdanie B: \"{{sentence_B}}\"\nPytanie: jaka jest zależność między zdaniami A i B? Możliwe odpowiedzi:\nA - znaczą dokładnie to samo\nB - mają podobne znaczenie\nC - mają różne znaczenie\nPrawidłowa odpowiedź:", "doc_to_target": "{{label|int - 1}}", "doc_to_choice": [ "A", "B", "C" ], "description": "", "target_delimiter": " ", "fewshot_delimiter": "\n\n", "num_fewshot": 0, "metric_list": [ { "metric": "acc", "aggregation": "mean", "higher_is_better": true }, { "metric": "acc_norm", "aggregation": "mean", "higher_is_better": true } ], "output_type": "multiple_choice", "repeats": 1, "should_decontaminate": true, "doc_to_decontamination_query": "{{sentence_A}} {{sentence_B}}" }, "polish_ppc_regex": { "task": "polish_ppc_regex", "dataset_path": "djstrong/ppc", "training_split": "train", "validation_split": "validation", "test_split": "test", "doc_to_text": "Zdanie A: \"{{sentence_A}}\"\nZdanie B: \"{{sentence_B}}\"\nPytanie: jaka jest zależność między zdaniami A i B? Możliwe odpowiedzi:\nA - wszystkie odpowiedzi poprawne\nB - znaczą dokładnie to samo\nC - mają podobne znaczenie\nD - mają różne znaczenie\nPrawidłowa odpowiedź:", "doc_to_target": "{{label|int}}", "description": "", "target_delimiter": " ", "fewshot_delimiter": "\n\n", "num_fewshot": 0, "metric_list": [ { "metric": "f1", "aggregation": "mean", "higher_is_better": true, "hf_evaluate": true, "average": "micro" }, { "metric": "accuracy", "aggregation": "mean", "higher_is_better": true } ], "output_type": "generate_until", "generation_kwargs": { "until": [ ".", "," ], "do_sample": false, "temperature": 0.0, "max_gen_toks": 50 }, "repeats": 1, "filter_list": [ { "name": "score-first", "filter": [ { "function": "regex", "regex_pattern": "(\\b[ABCD]\\b)" }, { "function": "take_first" }, { "function": "map", "mapping_dict": { "A": 0, "B": 1, "C": 2, "D": 3 }, "default_value": -1 }, { "function": "take_first" } ] } ], "should_decontaminate": true, "doc_to_decontamination_query": "{{sentence_A}} {{sentence_B}}" }, "polish_psc_multiple_choice": { "task": "polish_psc_multiple_choice", "dataset_path": "allegro/klej-psc", "training_split": "train", "test_split": "test", "doc_to_text": "Tekst: \"{{extract_text}}\"\nPodsumowanie: \"{{summary_text}}\"\nPytanie: Czy podsumowanie dla podanego tekstu jest poprawne?\nOdpowiedz krótko \"Tak\" lub \"Nie\". Prawidłowa odpowiedź:", "doc_to_target": "{{label|int}}", "doc_to_choice": [ "Nie", "Tak" ], "description": "", "target_delimiter": " ", "fewshot_delimiter": "\n\n", "num_fewshot": 0, "metric_list": [ { "metric": "acc", "aggregation": "mean", "higher_is_better": true }, { "metric": "acc_norm", "aggregation": "mean", "higher_is_better": true } ], "output_type": "multiple_choice", "repeats": 1, "should_decontaminate": true, "doc_to_decontamination_query": "{{extract_text}} {{summary_text}}" }, "polish_psc_regex": { "task": "polish_psc_regex", "dataset_path": "allegro/klej-psc", "training_split": "train", "test_split": "test", "doc_to_text": "Fragment 1: \"{{extract_text}}\"\nFragment 2: \"{{summary_text}}\"\nPytanie: jaka jest zależność między fragmentami 1 i 2?\nMożliwe odpowiedzi:\nA - wszystkie odpowiedzi poprawne\nB - dotyczą tego samego artykułu\nC - dotyczą różnych artykułów\nD - brak poprawnej odpowiedzi\nPrawidłowa odpowiedź:", "doc_to_target": "{{label|int + 1}}", "description": "", "target_delimiter": " ", "fewshot_delimiter": "\n\n", "num_fewshot": 0, "metric_list": [ { "metric": "f1", "aggregation": "mean", "higher_is_better": true, "hf_evaluate": true, "average": "micro" }, { "metric": "accuracy", "aggregation": "mean", "higher_is_better": true } ], "output_type": "generate_until", "generation_kwargs": { "until": [ ".", "," ], "do_sample": false, "temperature": 0.0, "max_gen_toks": 50 }, "repeats": 1, "filter_list": [ { "name": "score-first", "filter": [ { "function": "regex", "regex_pattern": "(\\b[ABCD]\\b)" }, { "function": "take_first" }, { "function": "map", "mapping_dict": { "A": 0, "B": 1, "C": 2, "D": 3 }, "default_value": -1 }, { "function": "take_first" } ] } ], "should_decontaminate": true, "doc_to_decontamination_query": "{{extract_text}} {{summary_text}}" } }, "versions": { "belebele_pol_Latn": 0.0, "polemo2_in": 1.0, "polemo2_in_multiple_choice": "Yaml", "polemo2_out": 1.0, "polemo2_out_multiple_choice": "Yaml", "polish": "N/A", "polish_8tags_multiple_choice": "Yaml", "polish_8tags_regex": "Yaml", "polish_belebele_regex": "Yaml", "polish_dyk_multiple_choice": "Yaml", "polish_dyk_regex": "Yaml", "polish_ppc_multiple_choice": "Yaml", "polish_ppc_regex": "Yaml", "polish_psc_multiple_choice": "Yaml", "polish_psc_regex": "Yaml" }, "n-shot": { "belebele_pol_Latn": 0, "polemo2_in": 0, "polemo2_in_multiple_choice": 0, "polemo2_out": 0, "polemo2_out_multiple_choice": 0, "polish": 0, "polish_8tags_multiple_choice": 0, "polish_8tags_regex": 0, "polish_belebele_regex": 0, "polish_dyk_multiple_choice": 0, "polish_dyk_regex": 0, "polish_ppc_multiple_choice": 0, "polish_ppc_regex": 0, "polish_psc_multiple_choice": 0, "polish_psc_regex": 0 }, "config": { "model": "hf", "model_args": "pretrained=Voicelab/trurl-2-7b", "batch_size": "auto", "batch_sizes": [ 32 ], "device": "cuda:0", "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, "git_hash": null }