{ "config_general": { "lighteval_sha": "?", "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null, "job_id": "", "start_time": 536.764640064, "end_time": 2862.164398973, "total_evaluation_time_secondes": "2325.399758909", "model_name": "sambanovasystems/SambaLingo-Arabic-Chat", "model_sha": "b35294fc67a7fafd58f4bbf02cdce073ab182fd0", "model_dtype": "torch.bfloat16", "model_size": "12.94 GB", "config": null }, "results": { "community|acva:Algeria|0": { "acc_norm": 0.5230769230769231, "acc_norm_stderr": 0.0358596530894741 }, "community|acva:Ancient_Egypt|0": { "acc_norm": 0.050793650793650794, "acc_norm_stderr": 0.01239139518482262 }, "community|acva:Arab_Empire|0": { "acc_norm": 0.30566037735849055, "acc_norm_stderr": 0.028353298073322666 }, "community|acva:Arabic_Architecture|0": { "acc_norm": 0.4564102564102564, "acc_norm_stderr": 0.035761230969912135 }, "community|acva:Arabic_Art|0": { "acc_norm": 0.3384615384615385, "acc_norm_stderr": 0.03397280032734094 }, "community|acva:Arabic_Astronomy|0": { "acc_norm": 0.4666666666666667, "acc_norm_stderr": 0.03581804596782233 }, "community|acva:Arabic_Calligraphy|0": { "acc_norm": 0.6784313725490196, "acc_norm_stderr": 0.02930710626317035 }, "community|acva:Arabic_Ceremony|0": { "acc_norm": 0.5297297297297298, "acc_norm_stderr": 0.036795272555679256 }, "community|acva:Arabic_Clothing|0": { "acc_norm": 0.517948717948718, "acc_norm_stderr": 0.03587477098773825 }, "community|acva:Arabic_Culture|0": { "acc_norm": 0.2358974358974359, "acc_norm_stderr": 0.030481516761721537 }, "community|acva:Arabic_Food|0": { "acc_norm": 0.441025641025641, "acc_norm_stderr": 0.0356473293185358 }, "community|acva:Arabic_Funeral|0": { "acc_norm": 0.4, "acc_norm_stderr": 0.050529115263991134 }, "community|acva:Arabic_Geography|0": { "acc_norm": 0.6344827586206897, "acc_norm_stderr": 0.04013124195424385 }, "community|acva:Arabic_History|0": { "acc_norm": 0.30256410256410254, "acc_norm_stderr": 0.03298070870085619 }, "community|acva:Arabic_Language_Origin|0": { "acc_norm": 0.5368421052631579, "acc_norm_stderr": 0.05143087276324537 }, "community|acva:Arabic_Literature|0": { "acc_norm": 0.47586206896551725, "acc_norm_stderr": 0.041618085035015295 }, "community|acva:Arabic_Math|0": { "acc_norm": 0.30256410256410254, "acc_norm_stderr": 0.03298070870085618 }, "community|acva:Arabic_Medicine|0": { "acc_norm": 0.46206896551724136, "acc_norm_stderr": 0.041546596717075474 }, "community|acva:Arabic_Music|0": { "acc_norm": 0.23741007194244604, "acc_norm_stderr": 0.036220593237998276 }, "community|acva:Arabic_Ornament|0": { "acc_norm": 0.5384615384615384, "acc_norm_stderr": 0.03579154352544572 }, "community|acva:Arabic_Philosophy|0": { "acc_norm": 0.5793103448275863, "acc_norm_stderr": 0.0411391498118926 }, "community|acva:Arabic_Physics_and_Chemistry|0": { "acc_norm": 0.5333333333333333, "acc_norm_stderr": 0.03581804596782232 }, "community|acva:Arabic_Wedding|0": { "acc_norm": 0.46153846153846156, "acc_norm_stderr": 0.0357915435254457 }, "community|acva:Bahrain|0": { "acc_norm": 0.3111111111111111, "acc_norm_stderr": 0.06979205927323111 }, "community|acva:Comoros|0": { "acc_norm": 0.37777777777777777, "acc_norm_stderr": 0.07309112127323451 }, "community|acva:Egypt_modern|0": { "acc_norm": 0.3157894736842105, "acc_norm_stderr": 0.04794350420740798 }, "community|acva:InfluenceFromAncientEgypt|0": { "acc_norm": 0.5641025641025641, "acc_norm_stderr": 0.035601666623466345 }, "community|acva:InfluenceFromByzantium|0": { "acc_norm": 0.7172413793103448, "acc_norm_stderr": 0.03752833958003337 }, "community|acva:InfluenceFromChina|0": { "acc_norm": 0.26666666666666666, "acc_norm_stderr": 0.0317493043641267 }, "community|acva:InfluenceFromGreece|0": { "acc_norm": 0.6307692307692307, "acc_norm_stderr": 0.034648411418637566 }, "community|acva:InfluenceFromIslam|0": { "acc_norm": 0.296551724137931, "acc_norm_stderr": 0.03806142687309993 }, "community|acva:InfluenceFromPersia|0": { "acc_norm": 0.6914285714285714, "acc_norm_stderr": 0.03501683519910117 }, "community|acva:InfluenceFromRome|0": { "acc_norm": 0.5743589743589743, "acc_norm_stderr": 0.03549871080367708 }, "community|acva:Iraq|0": { "acc_norm": 0.5294117647058824, "acc_norm_stderr": 0.0544600058689736 }, "community|acva:Islam_Education|0": { "acc_norm": 0.4564102564102564, "acc_norm_stderr": 0.03576123096991215 }, "community|acva:Islam_branches_and_schools|0": { "acc_norm": 0.4342857142857143, "acc_norm_stderr": 0.037576101528126626 }, "community|acva:Islamic_law_system|0": { "acc_norm": 0.4256410256410256, "acc_norm_stderr": 0.035498710803677086 }, "community|acva:Jordan|0": { "acc_norm": 0.3333333333333333, "acc_norm_stderr": 0.07106690545187012 }, "community|acva:Kuwait|0": { "acc_norm": 0.26666666666666666, "acc_norm_stderr": 0.06666666666666667 }, "community|acva:Lebanon|0": { "acc_norm": 0.2222222222222222, "acc_norm_stderr": 0.06267511942419626 }, "community|acva:Libya|0": { "acc_norm": 0.4444444444444444, "acc_norm_stderr": 0.07491109582924914 }, "community|acva:Mauritania|0": { "acc_norm": 0.4666666666666667, "acc_norm_stderr": 0.0752101433090355 }, "community|acva:Mesopotamia_civilization|0": { "acc_norm": 0.5225806451612903, "acc_norm_stderr": 0.0402500394824441 }, "community|acva:Morocco|0": { "acc_norm": 0.24444444444444444, "acc_norm_stderr": 0.06478835438716998 }, "community|acva:Oman|0": { "acc_norm": 0.2, "acc_norm_stderr": 0.06030226891555273 }, "community|acva:Palestine|0": { "acc_norm": 0.24705882352941178, "acc_norm_stderr": 0.047058823529411785 }, "community|acva:Qatar|0": { "acc_norm": 0.4, "acc_norm_stderr": 0.07385489458759964 }, "community|acva:Saudi_Arabia|0": { "acc_norm": 0.3282051282051282, "acc_norm_stderr": 0.03371243782413707 }, "community|acva:Somalia|0": { "acc_norm": 0.35555555555555557, "acc_norm_stderr": 0.07216392363431012 }, "community|acva:Sudan|0": { "acc_norm": 0.35555555555555557, "acc_norm_stderr": 0.07216392363431012 }, "community|acva:Syria|0": { "acc_norm": 0.3333333333333333, "acc_norm_stderr": 0.07106690545187012 }, "community|acva:Tunisia|0": { "acc_norm": 0.3111111111111111, "acc_norm_stderr": 0.06979205927323111 }, "community|acva:United_Arab_Emirates|0": { "acc_norm": 0.24705882352941178, "acc_norm_stderr": 0.047058823529411785 }, "community|acva:Yemen|0": { "acc_norm": 0.2, "acc_norm_stderr": 0.13333333333333333 }, "community|acva:communication|0": { "acc_norm": 0.42857142857142855, "acc_norm_stderr": 0.025974025974025955 }, "community|acva:computer_and_phone|0": { "acc_norm": 0.45084745762711864, "acc_norm_stderr": 0.02901934773187137 }, "community|acva:daily_life|0": { "acc_norm": 0.18694362017804153, "acc_norm_stderr": 0.021268948348414647 }, "community|acva:entertainment|0": { "acc_norm": 0.23389830508474577, "acc_norm_stderr": 0.024687839412166384 }, "community|alghafa:mcq_exams_test_ar|0": { "acc_norm": 0.3303411131059246, "acc_norm_stderr": 0.019946685327935997 }, "community|alghafa:meta_ar_dialects|0": { "acc_norm": 0.317516218721038, "acc_norm_stderr": 0.0063383125866490805 }, "community|alghafa:meta_ar_msa|0": { "acc_norm": 0.3553072625698324, "acc_norm_stderr": 0.016006989934803185 }, "community|alghafa:multiple_choice_facts_truefalse_balanced_task|0": { "acc_norm": 0.8933333333333333, "acc_norm_stderr": 0.03588436550487813 }, "community|alghafa:multiple_choice_grounded_statement_soqal_task|0": { "acc_norm": 0.47333333333333333, "acc_norm_stderr": 0.040903298047964304 }, "community|alghafa:multiple_choice_grounded_statement_xglue_mlqa_task|0": { "acc_norm": 0.36666666666666664, "acc_norm_stderr": 0.039478328284971595 }, "community|alghafa:multiple_choice_rating_sentiment_no_neutral_task|0": { "acc_norm": 0.6571607254534084, "acc_norm_stderr": 0.00530883236434125 }, "community|alghafa:multiple_choice_rating_sentiment_task|0": { "acc_norm": 0.48623853211009177, "acc_norm_stderr": 0.006455755606229747 }, "community|alghafa:multiple_choice_sentiment_task|0": { "acc_norm": 0.35348837209302325, "acc_norm_stderr": 0.011530225420513065 }, "community|arabic_exams|0": { "acc_norm": 0.30726256983240224, "acc_norm_stderr": 0.019927680956566315 }, "community|arabic_mmlu:abstract_algebra|0": { "acc_norm": 0.25, "acc_norm_stderr": 0.04351941398892446 }, "community|arabic_mmlu:anatomy|0": { "acc_norm": 0.3037037037037037, "acc_norm_stderr": 0.03972552884785136 }, "community|arabic_mmlu:astronomy|0": { "acc_norm": 0.3355263157894737, "acc_norm_stderr": 0.038424985593952694 }, "community|arabic_mmlu:business_ethics|0": { "acc_norm": 0.26, "acc_norm_stderr": 0.0440844002276808 }, "community|arabic_mmlu:clinical_knowledge|0": { "acc_norm": 0.3169811320754717, "acc_norm_stderr": 0.028637235639800928 }, "community|arabic_mmlu:college_biology|0": { "acc_norm": 0.3055555555555556, "acc_norm_stderr": 0.03852084696008534 }, "community|arabic_mmlu:college_chemistry|0": { "acc_norm": 0.26, "acc_norm_stderr": 0.0440844002276808 }, "community|arabic_mmlu:college_computer_science|0": { "acc_norm": 0.2, "acc_norm_stderr": 0.04020151261036844 }, "community|arabic_mmlu:college_mathematics|0": { "acc_norm": 0.2, "acc_norm_stderr": 0.04020151261036845 }, "community|arabic_mmlu:college_medicine|0": { "acc_norm": 0.26011560693641617, "acc_norm_stderr": 0.03345036916788992 }, "community|arabic_mmlu:college_physics|0": { "acc_norm": 0.27450980392156865, "acc_norm_stderr": 0.044405219061793275 }, "community|arabic_mmlu:computer_security|0": { "acc_norm": 0.34, "acc_norm_stderr": 0.047609522856952365 }, "community|arabic_mmlu:conceptual_physics|0": { "acc_norm": 0.32340425531914896, "acc_norm_stderr": 0.030579442773610337 }, "community|arabic_mmlu:econometrics|0": { "acc_norm": 0.2543859649122807, "acc_norm_stderr": 0.040969851398436716 }, "community|arabic_mmlu:electrical_engineering|0": { "acc_norm": 0.31724137931034485, "acc_norm_stderr": 0.038783523721386215 }, "community|arabic_mmlu:elementary_mathematics|0": { "acc_norm": 0.22486772486772486, "acc_norm_stderr": 0.02150209607822914 }, "community|arabic_mmlu:formal_logic|0": { "acc_norm": 0.21428571428571427, "acc_norm_stderr": 0.03670066451047182 }, "community|arabic_mmlu:global_facts|0": { "acc_norm": 0.28, "acc_norm_stderr": 0.04512608598542128 }, "community|arabic_mmlu:high_school_biology|0": { "acc_norm": 0.3096774193548387, "acc_norm_stderr": 0.026302774983517418 }, "community|arabic_mmlu:high_school_chemistry|0": { "acc_norm": 0.28078817733990147, "acc_norm_stderr": 0.031618563353586114 }, "community|arabic_mmlu:high_school_computer_science|0": { "acc_norm": 0.31, "acc_norm_stderr": 0.04648231987117316 }, "community|arabic_mmlu:high_school_european_history|0": { "acc_norm": 0.2727272727272727, "acc_norm_stderr": 0.03477691162163659 }, "community|arabic_mmlu:high_school_geography|0": { "acc_norm": 0.30808080808080807, "acc_norm_stderr": 0.032894773300986155 }, "community|arabic_mmlu:high_school_government_and_politics|0": { "acc_norm": 0.3316062176165803, "acc_norm_stderr": 0.03397636541089117 }, "community|arabic_mmlu:high_school_macroeconomics|0": { "acc_norm": 0.31025641025641026, "acc_norm_stderr": 0.023454674889404288 }, "community|arabic_mmlu:high_school_mathematics|0": { "acc_norm": 0.25925925925925924, "acc_norm_stderr": 0.02671924078371216 }, "community|arabic_mmlu:high_school_microeconomics|0": { "acc_norm": 0.24789915966386555, "acc_norm_stderr": 0.028047967224176892 }, "community|arabic_mmlu:high_school_physics|0": { "acc_norm": 0.19205298013245034, "acc_norm_stderr": 0.03216298420593614 }, "community|arabic_mmlu:high_school_psychology|0": { "acc_norm": 0.28623853211009176, "acc_norm_stderr": 0.019379436628919965 }, "community|arabic_mmlu:high_school_statistics|0": { "acc_norm": 0.24537037037037038, "acc_norm_stderr": 0.029346665094372944 }, "community|arabic_mmlu:high_school_us_history|0": { "acc_norm": 0.22549019607843138, "acc_norm_stderr": 0.02933116229425172 }, "community|arabic_mmlu:high_school_world_history|0": { "acc_norm": 0.270042194092827, "acc_norm_stderr": 0.028900721906293426 }, "community|arabic_mmlu:human_aging|0": { "acc_norm": 0.29596412556053814, "acc_norm_stderr": 0.0306365913486998 }, "community|arabic_mmlu:human_sexuality|0": { "acc_norm": 0.25190839694656486, "acc_norm_stderr": 0.03807387116306085 }, "community|arabic_mmlu:international_law|0": { "acc_norm": 0.4214876033057851, "acc_norm_stderr": 0.04507732278775094 }, "community|arabic_mmlu:jurisprudence|0": { "acc_norm": 0.37037037037037035, "acc_norm_stderr": 0.0466840803302493 }, "community|arabic_mmlu:logical_fallacies|0": { "acc_norm": 0.3006134969325153, "acc_norm_stderr": 0.03602511318806771 }, "community|arabic_mmlu:machine_learning|0": { "acc_norm": 0.25, "acc_norm_stderr": 0.04109974682633932 }, "community|arabic_mmlu:management|0": { "acc_norm": 0.33980582524271846, "acc_norm_stderr": 0.046897659372781335 }, "community|arabic_mmlu:marketing|0": { "acc_norm": 0.36752136752136755, "acc_norm_stderr": 0.03158539157745636 }, "community|arabic_mmlu:medical_genetics|0": { "acc_norm": 0.35, "acc_norm_stderr": 0.0479372485441102 }, "community|arabic_mmlu:miscellaneous|0": { "acc_norm": 0.3397190293742018, "acc_norm_stderr": 0.016936394114301652 }, "community|arabic_mmlu:moral_disputes|0": { "acc_norm": 0.2832369942196532, "acc_norm_stderr": 0.024257901705323378 }, "community|arabic_mmlu:moral_scenarios|0": { "acc_norm": 0.2435754189944134, "acc_norm_stderr": 0.01435591196476786 }, "community|arabic_mmlu:nutrition|0": { "acc_norm": 0.31699346405228757, "acc_norm_stderr": 0.02664327847450875 }, "community|arabic_mmlu:philosophy|0": { "acc_norm": 0.34726688102893893, "acc_norm_stderr": 0.027040745502307336 }, "community|arabic_mmlu:prehistory|0": { "acc_norm": 0.3055555555555556, "acc_norm_stderr": 0.02563082497562135 }, "community|arabic_mmlu:professional_accounting|0": { "acc_norm": 0.3049645390070922, "acc_norm_stderr": 0.027464708442022142 }, "community|arabic_mmlu:professional_law|0": { "acc_norm": 0.27640156453715775, "acc_norm_stderr": 0.011422153194553577 }, "community|arabic_mmlu:professional_medicine|0": { "acc_norm": 0.20588235294117646, "acc_norm_stderr": 0.024562204314142317 }, "community|arabic_mmlu:professional_psychology|0": { "acc_norm": 0.30718954248366015, "acc_norm_stderr": 0.018663359671463663 }, "community|arabic_mmlu:public_relations|0": { "acc_norm": 0.35454545454545455, "acc_norm_stderr": 0.04582004841505415 }, "community|arabic_mmlu:security_studies|0": { "acc_norm": 0.30612244897959184, "acc_norm_stderr": 0.029504896454595957 }, "community|arabic_mmlu:sociology|0": { "acc_norm": 0.3283582089552239, "acc_norm_stderr": 0.033206858897443244 }, "community|arabic_mmlu:us_foreign_policy|0": { "acc_norm": 0.31, "acc_norm_stderr": 0.04648231987117316 }, "community|arabic_mmlu:virology|0": { "acc_norm": 0.3253012048192771, "acc_norm_stderr": 0.03647168523683228 }, "community|arabic_mmlu:world_religions|0": { "acc_norm": 0.29239766081871343, "acc_norm_stderr": 0.034886477134579215 }, "community|arc_challenge_okapi_ar|0": { "acc_norm": 0.33448275862068966, "acc_norm_stderr": 0.01385877378068539 }, "community|arc_easy_ar|0": { "acc_norm": 0.38071065989847713, "acc_norm_stderr": 0.009988780276735732 }, "community|boolq_ar|0": { "acc_norm": 0.695398773006135, "acc_norm_stderr": 0.008061959349756813 }, "community|copa_ext_ar|0": { "acc_norm": 0.4777777777777778, "acc_norm_stderr": 0.05294752255076824 }, "community|hellaswag_okapi_ar|0": { "acc_norm": 0.26507469196379896, "acc_norm_stderr": 0.004609152381455368 }, "community|openbook_qa_ext_ar|0": { "acc_norm": 0.41818181818181815, "acc_norm_stderr": 0.022192832902420363 }, "community|piqa_ar|0": { "acc_norm": 0.5417348608837971, "acc_norm_stderr": 0.011640967256247612 }, "community|race_ar|0": { "acc_norm": 0.37532968147697304, "acc_norm_stderr": 0.006897576738971542 }, "community|sciq_ar|0": { "acc_norm": 0.39396984924623113, "acc_norm_stderr": 0.015498348296857803 }, "community|toxigen_ar|0": { "acc_norm": 0.4320855614973262, "acc_norm_stderr": 0.01620887578524445 }, "lighteval|xstory_cloze:ar|0": { "acc": 0.6657842488418266, "acc_stderr": 0.012139246810918231 }, "community|acva:_average|0": { "acc_norm": 0.403079033847007, "acc_norm_stderr": 0.045956791952076054 }, "community|alghafa:_average|0": { "acc_norm": 0.4703761730429613, "acc_norm_stderr": 0.020205865897587372 }, "community|arabic_mmlu:_average|0": { "acc_norm": 0.29061837999917134, "acc_norm_stderr": 0.033812069602402954 }, "all": { "acc_norm": 0.3614757495026021, "acc_norm_stderr": 0.03671457158452871, "acc": 0.6657842488418266, "acc_stderr": 0.012139246810918231 } }, "versions": { "community|acva:Algeria|0": 0, "community|acva:Ancient_Egypt|0": 0, "community|acva:Arab_Empire|0": 0, "community|acva:Arabic_Architecture|0": 0, "community|acva:Arabic_Art|0": 0, "community|acva:Arabic_Astronomy|0": 0, "community|acva:Arabic_Calligraphy|0": 0, "community|acva:Arabic_Ceremony|0": 0, "community|acva:Arabic_Clothing|0": 0, "community|acva:Arabic_Culture|0": 0, "community|acva:Arabic_Food|0": 0, "community|acva:Arabic_Funeral|0": 0, "community|acva:Arabic_Geography|0": 0, "community|acva:Arabic_History|0": 0, "community|acva:Arabic_Language_Origin|0": 0, "community|acva:Arabic_Literature|0": 0, "community|acva:Arabic_Math|0": 0, "community|acva:Arabic_Medicine|0": 0, "community|acva:Arabic_Music|0": 0, "community|acva:Arabic_Ornament|0": 0, "community|acva:Arabic_Philosophy|0": 0, "community|acva:Arabic_Physics_and_Chemistry|0": 0, "community|acva:Arabic_Wedding|0": 0, "community|acva:Bahrain|0": 0, "community|acva:Comoros|0": 0, "community|acva:Egypt_modern|0": 0, "community|acva:InfluenceFromAncientEgypt|0": 0, "community|acva:InfluenceFromByzantium|0": 0, "community|acva:InfluenceFromChina|0": 0, "community|acva:InfluenceFromGreece|0": 0, "community|acva:InfluenceFromIslam|0": 0, "community|acva:InfluenceFromPersia|0": 0, "community|acva:InfluenceFromRome|0": 0, "community|acva:Iraq|0": 0, "community|acva:Islam_Education|0": 0, "community|acva:Islam_branches_and_schools|0": 0, "community|acva:Islamic_law_system|0": 0, "community|acva:Jordan|0": 0, "community|acva:Kuwait|0": 0, "community|acva:Lebanon|0": 0, "community|acva:Libya|0": 0, "community|acva:Mauritania|0": 0, "community|acva:Mesopotamia_civilization|0": 0, "community|acva:Morocco|0": 0, "community|acva:Oman|0": 0, "community|acva:Palestine|0": 0, "community|acva:Qatar|0": 0, "community|acva:Saudi_Arabia|0": 0, "community|acva:Somalia|0": 0, "community|acva:Sudan|0": 0, "community|acva:Syria|0": 0, "community|acva:Tunisia|0": 0, "community|acva:United_Arab_Emirates|0": 0, "community|acva:Yemen|0": 0, "community|acva:communication|0": 0, "community|acva:computer_and_phone|0": 0, "community|acva:daily_life|0": 0, "community|acva:entertainment|0": 0, "community|alghafa:mcq_exams_test_ar|0": 0, "community|alghafa:meta_ar_dialects|0": 0, "community|alghafa:meta_ar_msa|0": 0, "community|alghafa:multiple_choice_facts_truefalse_balanced_task|0": 0, "community|alghafa:multiple_choice_grounded_statement_soqal_task|0": 0, "community|alghafa:multiple_choice_grounded_statement_xglue_mlqa_task|0": 0, "community|alghafa:multiple_choice_rating_sentiment_no_neutral_task|0": 0, "community|alghafa:multiple_choice_rating_sentiment_task|0": 0, "community|alghafa:multiple_choice_sentiment_task|0": 0, "community|arabic_exams|0": 0, "community|arabic_mmlu:abstract_algebra|0": 0, "community|arabic_mmlu:anatomy|0": 0, "community|arabic_mmlu:astronomy|0": 0, "community|arabic_mmlu:business_ethics|0": 0, "community|arabic_mmlu:clinical_knowledge|0": 0, "community|arabic_mmlu:college_biology|0": 0, "community|arabic_mmlu:college_chemistry|0": 0, "community|arabic_mmlu:college_computer_science|0": 0, "community|arabic_mmlu:college_mathematics|0": 0, "community|arabic_mmlu:college_medicine|0": 0, "community|arabic_mmlu:college_physics|0": 0, "community|arabic_mmlu:computer_security|0": 0, "community|arabic_mmlu:conceptual_physics|0": 0, "community|arabic_mmlu:econometrics|0": 0, "community|arabic_mmlu:electrical_engineering|0": 0, "community|arabic_mmlu:elementary_mathematics|0": 0, "community|arabic_mmlu:formal_logic|0": 0, "community|arabic_mmlu:global_facts|0": 0, "community|arabic_mmlu:high_school_biology|0": 0, "community|arabic_mmlu:high_school_chemistry|0": 0, "community|arabic_mmlu:high_school_computer_science|0": 0, "community|arabic_mmlu:high_school_european_history|0": 0, "community|arabic_mmlu:high_school_geography|0": 0, "community|arabic_mmlu:high_school_government_and_politics|0": 0, "community|arabic_mmlu:high_school_macroeconomics|0": 0, "community|arabic_mmlu:high_school_mathematics|0": 0, "community|arabic_mmlu:high_school_microeconomics|0": 0, "community|arabic_mmlu:high_school_physics|0": 0, "community|arabic_mmlu:high_school_psychology|0": 0, "community|arabic_mmlu:high_school_statistics|0": 0, "community|arabic_mmlu:high_school_us_history|0": 0, "community|arabic_mmlu:high_school_world_history|0": 0, "community|arabic_mmlu:human_aging|0": 0, "community|arabic_mmlu:human_sexuality|0": 0, "community|arabic_mmlu:international_law|0": 0, "community|arabic_mmlu:jurisprudence|0": 0, "community|arabic_mmlu:logical_fallacies|0": 0, "community|arabic_mmlu:machine_learning|0": 0, "community|arabic_mmlu:management|0": 0, "community|arabic_mmlu:marketing|0": 0, "community|arabic_mmlu:medical_genetics|0": 0, "community|arabic_mmlu:miscellaneous|0": 0, "community|arabic_mmlu:moral_disputes|0": 0, "community|arabic_mmlu:moral_scenarios|0": 0, "community|arabic_mmlu:nutrition|0": 0, "community|arabic_mmlu:philosophy|0": 0, "community|arabic_mmlu:prehistory|0": 0, "community|arabic_mmlu:professional_accounting|0": 0, "community|arabic_mmlu:professional_law|0": 0, "community|arabic_mmlu:professional_medicine|0": 0, "community|arabic_mmlu:professional_psychology|0": 0, "community|arabic_mmlu:public_relations|0": 0, "community|arabic_mmlu:security_studies|0": 0, "community|arabic_mmlu:sociology|0": 0, "community|arabic_mmlu:us_foreign_policy|0": 0, "community|arabic_mmlu:virology|0": 0, "community|arabic_mmlu:world_religions|0": 0, "community|arc_challenge_okapi_ar|0": 0, "community|arc_easy_ar|0": 0, "community|boolq_ar|0": 0, "community|copa_ext_ar|0": 0, "community|hellaswag_okapi_ar|0": 0, "community|openbook_qa_ext_ar|0": 0, "community|piqa_ar|0": 0, "community|race_ar|0": 0, "community|sciq_ar|0": 0, "community|toxigen_ar|0": 0, "lighteval|xstory_cloze:ar|0": 0 }, "config_tasks": { "community|acva:Algeria": { "name": "acva:Algeria", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Algeria", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Ancient_Egypt": { "name": "acva:Ancient_Egypt", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Ancient_Egypt", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 315, "effective_num_docs": 315, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arab_Empire": { "name": "acva:Arab_Empire", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arab_Empire", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 265, "effective_num_docs": 265, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Architecture": { "name": "acva:Arabic_Architecture", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Architecture", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Art": { "name": "acva:Arabic_Art", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Art", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Astronomy": { "name": "acva:Arabic_Astronomy", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Astronomy", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Calligraphy": { "name": "acva:Arabic_Calligraphy", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Calligraphy", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 255, "effective_num_docs": 255, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Ceremony": { "name": "acva:Arabic_Ceremony", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Ceremony", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 185, "effective_num_docs": 185, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Clothing": { "name": "acva:Arabic_Clothing", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Clothing", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Culture": { "name": "acva:Arabic_Culture", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Culture", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Food": { "name": "acva:Arabic_Food", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Food", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Funeral": { "name": "acva:Arabic_Funeral", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Funeral", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 95, "effective_num_docs": 95, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Geography": { "name": "acva:Arabic_Geography", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Geography", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 145, "effective_num_docs": 145, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_History": { "name": "acva:Arabic_History", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_History", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Language_Origin": { "name": "acva:Arabic_Language_Origin", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Language_Origin", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 95, "effective_num_docs": 95, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Literature": { "name": "acva:Arabic_Literature", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Literature", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 145, "effective_num_docs": 145, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Math": { "name": "acva:Arabic_Math", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Math", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Medicine": { "name": "acva:Arabic_Medicine", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Medicine", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 145, "effective_num_docs": 145, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Music": { "name": "acva:Arabic_Music", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Music", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 139, "effective_num_docs": 139, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Ornament": { "name": "acva:Arabic_Ornament", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Ornament", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Philosophy": { "name": "acva:Arabic_Philosophy", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Philosophy", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 145, "effective_num_docs": 145, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Physics_and_Chemistry": { "name": "acva:Arabic_Physics_and_Chemistry", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Physics_and_Chemistry", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Wedding": { "name": "acva:Arabic_Wedding", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Wedding", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Bahrain": { "name": "acva:Bahrain", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Bahrain", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 45, "effective_num_docs": 45, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Comoros": { "name": "acva:Comoros", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Comoros", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 45, "effective_num_docs": 45, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Egypt_modern": { "name": "acva:Egypt_modern", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Egypt_modern", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 95, "effective_num_docs": 95, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:InfluenceFromAncientEgypt": { "name": "acva:InfluenceFromAncientEgypt", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "InfluenceFromAncientEgypt", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:InfluenceFromByzantium": { "name": "acva:InfluenceFromByzantium", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "InfluenceFromByzantium", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 145, "effective_num_docs": 145, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:InfluenceFromChina": { "name": "acva:InfluenceFromChina", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "InfluenceFromChina", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:InfluenceFromGreece": { "name": "acva:InfluenceFromGreece", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "InfluenceFromGreece", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:InfluenceFromIslam": { "name": "acva:InfluenceFromIslam", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "InfluenceFromIslam", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 145, "effective_num_docs": 145, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:InfluenceFromPersia": { "name": "acva:InfluenceFromPersia", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "InfluenceFromPersia", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 175, "effective_num_docs": 175, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:InfluenceFromRome": { "name": "acva:InfluenceFromRome", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "InfluenceFromRome", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Iraq": { "name": "acva:Iraq", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Iraq", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 85, "effective_num_docs": 85, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Islam_Education": { "name": "acva:Islam_Education", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Islam_Education", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Islam_branches_and_schools": { "name": "acva:Islam_branches_and_schools", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Islam_branches_and_schools", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 175, "effective_num_docs": 175, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Islamic_law_system": { "name": "acva:Islamic_law_system", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Islamic_law_system", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Jordan": { "name": "acva:Jordan", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Jordan", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 45, "effective_num_docs": 45, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Kuwait": { "name": "acva:Kuwait", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Kuwait", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 45, "effective_num_docs": 45, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Lebanon": { "name": "acva:Lebanon", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Lebanon", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 45, "effective_num_docs": 45, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Libya": { "name": "acva:Libya", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Libya", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 45, "effective_num_docs": 45, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Mauritania": { "name": "acva:Mauritania", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Mauritania", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 45, "effective_num_docs": 45, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Mesopotamia_civilization": { "name": "acva:Mesopotamia_civilization", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Mesopotamia_civilization", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 155, "effective_num_docs": 155, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Morocco": { "name": "acva:Morocco", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Morocco", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 45, "effective_num_docs": 45, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Oman": { "name": "acva:Oman", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Oman", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 45, "effective_num_docs": 45, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Palestine": { "name": "acva:Palestine", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Palestine", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 85, "effective_num_docs": 85, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Qatar": { "name": "acva:Qatar", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Qatar", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 45, "effective_num_docs": 45, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Saudi_Arabia": { "name": "acva:Saudi_Arabia", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Saudi_Arabia", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Somalia": { "name": "acva:Somalia", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Somalia", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 45, "effective_num_docs": 45, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Sudan": { "name": "acva:Sudan", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Sudan", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 45, "effective_num_docs": 45, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Syria": { "name": "acva:Syria", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Syria", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 45, "effective_num_docs": 45, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Tunisia": { "name": "acva:Tunisia", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Tunisia", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 45, "effective_num_docs": 45, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:United_Arab_Emirates": { "name": "acva:United_Arab_Emirates", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "United_Arab_Emirates", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 85, "effective_num_docs": 85, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Yemen": { "name": "acva:Yemen", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Yemen", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 10, "effective_num_docs": 10, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:communication": { "name": "acva:communication", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "communication", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 364, "effective_num_docs": 364, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:computer_and_phone": { "name": "acva:computer_and_phone", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "computer_and_phone", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 295, "effective_num_docs": 295, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:daily_life": { "name": "acva:daily_life", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "daily_life", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 337, "effective_num_docs": 337, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:entertainment": { "name": "acva:entertainment", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "entertainment", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 295, "effective_num_docs": 295, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|alghafa:mcq_exams_test_ar": { "name": "alghafa:mcq_exams_test_ar", "prompt_function": "alghafa_prompt", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native", "hf_subset": "mcq_exams_test_ar", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 557, "effective_num_docs": 557, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|alghafa:meta_ar_dialects": { "name": "alghafa:meta_ar_dialects", "prompt_function": "alghafa_prompt", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native", "hf_subset": "meta_ar_dialects", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 5395, "effective_num_docs": 5395, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|alghafa:meta_ar_msa": { "name": "alghafa:meta_ar_msa", "prompt_function": "alghafa_prompt", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native", "hf_subset": "meta_ar_msa", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 895, "effective_num_docs": 895, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|alghafa:multiple_choice_facts_truefalse_balanced_task": { "name": "alghafa:multiple_choice_facts_truefalse_balanced_task", "prompt_function": "alghafa_prompt", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native", "hf_subset": "multiple_choice_facts_truefalse_balanced_task", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 75, "effective_num_docs": 75, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|alghafa:multiple_choice_grounded_statement_soqal_task": { "name": "alghafa:multiple_choice_grounded_statement_soqal_task", "prompt_function": "alghafa_prompt", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native", "hf_subset": "multiple_choice_grounded_statement_soqal_task", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 150, "effective_num_docs": 150, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|alghafa:multiple_choice_grounded_statement_xglue_mlqa_task": { "name": "alghafa:multiple_choice_grounded_statement_xglue_mlqa_task", "prompt_function": "alghafa_prompt", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native", "hf_subset": "multiple_choice_grounded_statement_xglue_mlqa_task", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 150, "effective_num_docs": 150, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|alghafa:multiple_choice_rating_sentiment_no_neutral_task": { "name": "alghafa:multiple_choice_rating_sentiment_no_neutral_task", "prompt_function": "alghafa_prompt", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native", "hf_subset": "multiple_choice_rating_sentiment_no_neutral_task", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 7995, "effective_num_docs": 7995, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|alghafa:multiple_choice_rating_sentiment_task": { "name": "alghafa:multiple_choice_rating_sentiment_task", "prompt_function": "alghafa_prompt", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native", "hf_subset": "multiple_choice_rating_sentiment_task", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 5995, "effective_num_docs": 5995, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|alghafa:multiple_choice_sentiment_task": { "name": "alghafa:multiple_choice_sentiment_task", "prompt_function": "alghafa_prompt", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native", "hf_subset": "multiple_choice_sentiment_task", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 1720, "effective_num_docs": 1720, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_exams": { "name": "arabic_exams", "prompt_function": "arabic_exams", "hf_repo": "OALL/Arabic_EXAMS", "hf_subset": "default", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": null, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 537, "effective_num_docs": 537, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:abstract_algebra": { "name": "arabic_mmlu:abstract_algebra", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "abstract_algebra", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:anatomy": { "name": "arabic_mmlu:anatomy", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "anatomy", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 135, "effective_num_docs": 135, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:astronomy": { "name": "arabic_mmlu:astronomy", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "astronomy", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 152, "effective_num_docs": 152, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:business_ethics": { "name": "arabic_mmlu:business_ethics", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "business_ethics", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:clinical_knowledge": { "name": "arabic_mmlu:clinical_knowledge", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "clinical_knowledge", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 265, "effective_num_docs": 265, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:college_biology": { "name": "arabic_mmlu:college_biology", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "college_biology", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 144, "effective_num_docs": 144, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:college_chemistry": { "name": "arabic_mmlu:college_chemistry", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "college_chemistry", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:college_computer_science": { "name": "arabic_mmlu:college_computer_science", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "college_computer_science", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:college_mathematics": { "name": "arabic_mmlu:college_mathematics", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "college_mathematics", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:college_medicine": { "name": "arabic_mmlu:college_medicine", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "college_medicine", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 173, "effective_num_docs": 173, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:college_physics": { "name": "arabic_mmlu:college_physics", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "college_physics", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 102, "effective_num_docs": 102, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:computer_security": { "name": "arabic_mmlu:computer_security", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "computer_security", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:conceptual_physics": { "name": "arabic_mmlu:conceptual_physics", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "conceptual_physics", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 235, "effective_num_docs": 235, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:econometrics": { "name": "arabic_mmlu:econometrics", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "econometrics", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 114, "effective_num_docs": 114, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:electrical_engineering": { "name": "arabic_mmlu:electrical_engineering", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "electrical_engineering", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 145, "effective_num_docs": 145, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:elementary_mathematics": { "name": "arabic_mmlu:elementary_mathematics", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "elementary_mathematics", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 378, "effective_num_docs": 378, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:formal_logic": { "name": "arabic_mmlu:formal_logic", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "formal_logic", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 126, "effective_num_docs": 126, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:global_facts": { "name": "arabic_mmlu:global_facts", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "global_facts", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:high_school_biology": { "name": "arabic_mmlu:high_school_biology", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "high_school_biology", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 310, "effective_num_docs": 310, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:high_school_chemistry": { "name": "arabic_mmlu:high_school_chemistry", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "high_school_chemistry", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 203, "effective_num_docs": 203, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:high_school_computer_science": { "name": "arabic_mmlu:high_school_computer_science", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "high_school_computer_science", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:high_school_european_history": { "name": "arabic_mmlu:high_school_european_history", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "high_school_european_history", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 165, "effective_num_docs": 165, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:high_school_geography": { "name": "arabic_mmlu:high_school_geography", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "high_school_geography", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 198, "effective_num_docs": 198, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:high_school_government_and_politics": { "name": "arabic_mmlu:high_school_government_and_politics", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "high_school_government_and_politics", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 193, "effective_num_docs": 193, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:high_school_macroeconomics": { "name": "arabic_mmlu:high_school_macroeconomics", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "high_school_macroeconomics", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 390, "effective_num_docs": 390, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:high_school_mathematics": { "name": "arabic_mmlu:high_school_mathematics", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "high_school_mathematics", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 270, "effective_num_docs": 270, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:high_school_microeconomics": { "name": "arabic_mmlu:high_school_microeconomics", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "high_school_microeconomics", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 238, "effective_num_docs": 238, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:high_school_physics": { "name": "arabic_mmlu:high_school_physics", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "high_school_physics", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 151, "effective_num_docs": 151, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:high_school_psychology": { "name": "arabic_mmlu:high_school_psychology", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "high_school_psychology", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 545, "effective_num_docs": 545, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:high_school_statistics": { "name": "arabic_mmlu:high_school_statistics", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "high_school_statistics", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 216, "effective_num_docs": 216, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:high_school_us_history": { "name": "arabic_mmlu:high_school_us_history", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "high_school_us_history", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 204, "effective_num_docs": 204, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:high_school_world_history": { "name": "arabic_mmlu:high_school_world_history", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "high_school_world_history", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 237, "effective_num_docs": 237, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:human_aging": { "name": "arabic_mmlu:human_aging", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "human_aging", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 223, "effective_num_docs": 223, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:human_sexuality": { "name": "arabic_mmlu:human_sexuality", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "human_sexuality", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 131, "effective_num_docs": 131, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:international_law": { "name": "arabic_mmlu:international_law", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "international_law", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 121, "effective_num_docs": 121, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:jurisprudence": { "name": "arabic_mmlu:jurisprudence", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "jurisprudence", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 108, "effective_num_docs": 108, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:logical_fallacies": { "name": "arabic_mmlu:logical_fallacies", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "logical_fallacies", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 163, "effective_num_docs": 163, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:machine_learning": { "name": "arabic_mmlu:machine_learning", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "machine_learning", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 112, "effective_num_docs": 112, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:management": { "name": "arabic_mmlu:management", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "management", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 103, "effective_num_docs": 103, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:marketing": { "name": "arabic_mmlu:marketing", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "marketing", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 234, "effective_num_docs": 234, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:medical_genetics": { "name": "arabic_mmlu:medical_genetics", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "medical_genetics", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:miscellaneous": { "name": "arabic_mmlu:miscellaneous", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "miscellaneous", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 783, "effective_num_docs": 783, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:moral_disputes": { "name": "arabic_mmlu:moral_disputes", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "moral_disputes", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 346, "effective_num_docs": 346, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:moral_scenarios": { "name": "arabic_mmlu:moral_scenarios", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "moral_scenarios", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 895, "effective_num_docs": 895, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:nutrition": { "name": "arabic_mmlu:nutrition", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "nutrition", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 306, "effective_num_docs": 306, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:philosophy": { "name": "arabic_mmlu:philosophy", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "philosophy", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 311, "effective_num_docs": 311, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:prehistory": { "name": "arabic_mmlu:prehistory", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "prehistory", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 324, "effective_num_docs": 324, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:professional_accounting": { "name": "arabic_mmlu:professional_accounting", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "professional_accounting", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 282, "effective_num_docs": 282, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:professional_law": { "name": "arabic_mmlu:professional_law", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "professional_law", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 1534, "effective_num_docs": 1534, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:professional_medicine": { "name": "arabic_mmlu:professional_medicine", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "professional_medicine", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 272, "effective_num_docs": 272, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:professional_psychology": { "name": "arabic_mmlu:professional_psychology", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "professional_psychology", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 612, "effective_num_docs": 612, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:public_relations": { "name": "arabic_mmlu:public_relations", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "public_relations", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 110, "effective_num_docs": 110, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:security_studies": { "name": "arabic_mmlu:security_studies", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "security_studies", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 245, "effective_num_docs": 245, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:sociology": { "name": "arabic_mmlu:sociology", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "sociology", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 201, "effective_num_docs": 201, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:us_foreign_policy": { "name": "arabic_mmlu:us_foreign_policy", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "us_foreign_policy", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:virology": { "name": "arabic_mmlu:virology", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "virology", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 166, "effective_num_docs": 166, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:world_religions": { "name": "arabic_mmlu:world_religions", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "world_religions", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 171, "effective_num_docs": 171, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arc_challenge_okapi_ar": { "name": "arc_challenge_okapi_ar", "prompt_function": "alghafa_prompt", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Translated", "hf_subset": "arc_challenge_okapi_ar", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": null, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 1160, "effective_num_docs": 1160, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arc_easy_ar": { "name": "arc_easy_ar", "prompt_function": "alghafa_prompt", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Translated", "hf_subset": "arc_easy_ar", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": null, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 2364, "effective_num_docs": 2364, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|boolq_ar": { "name": "boolq_ar", "prompt_function": "boolq_prompt_arabic", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Translated", "hf_subset": "boolq_ar", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": null, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 3260, "effective_num_docs": 3260, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|copa_ext_ar": { "name": "copa_ext_ar", "prompt_function": "copa_prompt_arabic", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Translated", "hf_subset": "copa_ext_ar", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": null, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 90, "effective_num_docs": 90, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|hellaswag_okapi_ar": { "name": "hellaswag_okapi_ar", "prompt_function": "hellaswag_prompt_arabic", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Translated", "hf_subset": "hellaswag_okapi_ar", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": null, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 9171, "effective_num_docs": 9171, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|openbook_qa_ext_ar": { "name": "openbook_qa_ext_ar", "prompt_function": "alghafa_prompt", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Translated", "hf_subset": "openbook_qa_ext_ar", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": null, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 495, "effective_num_docs": 495, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|piqa_ar": { "name": "piqa_ar", "prompt_function": "alghafa_prompt", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Translated", "hf_subset": "piqa_ar", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": null, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 1833, "effective_num_docs": 1833, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|race_ar": { "name": "race_ar", "prompt_function": "alghafa_prompt", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Translated", "hf_subset": "race_ar", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": null, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 4929, "effective_num_docs": 4929, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|sciq_ar": { "name": "sciq_ar", "prompt_function": "sciq_prompt_arabic", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Translated", "hf_subset": "sciq_ar", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": null, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 995, "effective_num_docs": 995, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|toxigen_ar": { "name": "toxigen_ar", "prompt_function": "toxigen_prompt_arabic", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Translated", "hf_subset": "toxigen_ar", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": null, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 935, "effective_num_docs": 935, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "lighteval|xstory_cloze:ar": { "name": "xstory_cloze:ar", "prompt_function": "storycloze", "hf_repo": "juletxara/xstory_cloze", "hf_subset": "ar", "metric": [ "loglikelihood_acc" ], "hf_avail_splits": [ "training", "eval" ], "evaluation_splits": [ "eval" ], "few_shots_split": null, "few_shots_select": null, "generation_size": -1, "stop_sequence": [ "\n" ], "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "lighteval" ], "original_num_docs": 1511, "effective_num_docs": 1511, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 } }, "summary_tasks": { "community|acva:Algeria|0": { "hashes": { "hash_examples": "da5a3003cd46f6f9", "hash_full_prompts": "da5a3003cd46f6f9", "hash_input_tokens": "503935e6c18624b6", "hash_cont_tokens": "df2048053ef0446e" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Ancient_Egypt|0": { "hashes": { "hash_examples": "52d6f767fede195b", "hash_full_prompts": "52d6f767fede195b", "hash_input_tokens": "e1afd9ee12ddc273", "hash_cont_tokens": "6be3ab271a2cb744" }, "truncated": 0, "non_truncated": 315, "padded": 630, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arab_Empire|0": { "hashes": { "hash_examples": "8dacff6a79804a75", "hash_full_prompts": "8dacff6a79804a75", "hash_input_tokens": "5b5d051012c731a4", "hash_cont_tokens": "3f10541eccd3d046" }, "truncated": 0, "non_truncated": 265, "padded": 530, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Architecture|0": { "hashes": { "hash_examples": "df286cd862d9f6bb", "hash_full_prompts": "df286cd862d9f6bb", "hash_input_tokens": "737ea783eaf6a5e9", "hash_cont_tokens": "df2048053ef0446e" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Art|0": { "hashes": { "hash_examples": "112883d764118a49", "hash_full_prompts": "112883d764118a49", "hash_input_tokens": "c2e86242b70bc56d", "hash_cont_tokens": "df2048053ef0446e" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Astronomy|0": { "hashes": { "hash_examples": "20dcdf2454bf8671", "hash_full_prompts": "20dcdf2454bf8671", "hash_input_tokens": "2523571a0fa48e8f", "hash_cont_tokens": "df2048053ef0446e" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Calligraphy|0": { "hashes": { "hash_examples": "3a9f9d1ebe868a15", "hash_full_prompts": "3a9f9d1ebe868a15", "hash_input_tokens": "117aa239a226ce0d", "hash_cont_tokens": "2ea2346ae04a04bd" }, "truncated": 0, "non_truncated": 255, "padded": 510, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Ceremony|0": { "hashes": { "hash_examples": "c927630f8d2f44da", "hash_full_prompts": "c927630f8d2f44da", "hash_input_tokens": "2f18ab10490e2fa9", "hash_cont_tokens": "410fbdbbe5c6a1ba" }, "truncated": 0, "non_truncated": 185, "padded": 370, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Clothing|0": { "hashes": { "hash_examples": "6ad0740c2ac6ac92", "hash_full_prompts": "6ad0740c2ac6ac92", "hash_input_tokens": "b024b698b99505b5", "hash_cont_tokens": "df2048053ef0446e" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Culture|0": { "hashes": { "hash_examples": "2177bd857ad872ae", "hash_full_prompts": "2177bd857ad872ae", "hash_input_tokens": "5f79f9e8a9fd08c5", "hash_cont_tokens": "df2048053ef0446e" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Food|0": { "hashes": { "hash_examples": "a6ada65b71d7c9c5", "hash_full_prompts": "a6ada65b71d7c9c5", "hash_input_tokens": "3885b6e25a069cef", "hash_cont_tokens": "df2048053ef0446e" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Funeral|0": { "hashes": { "hash_examples": "fcee39dc29eaae91", "hash_full_prompts": "fcee39dc29eaae91", "hash_input_tokens": "78e7d59ebcaf7989", "hash_cont_tokens": "f0c5cda13b476af9" }, "truncated": 0, "non_truncated": 95, "padded": 190, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Geography|0": { "hashes": { "hash_examples": "d36eda7c89231c02", "hash_full_prompts": "d36eda7c89231c02", "hash_input_tokens": "6e9baccfff4536cd", "hash_cont_tokens": "0755257cc7305005" }, "truncated": 0, "non_truncated": 145, "padded": 290, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_History|0": { "hashes": { "hash_examples": "6354ac0d6db6a5fc", "hash_full_prompts": "6354ac0d6db6a5fc", "hash_input_tokens": "53ca2529585d8637", "hash_cont_tokens": "df2048053ef0446e" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Language_Origin|0": { "hashes": { "hash_examples": "ddc967c8aca34402", "hash_full_prompts": "ddc967c8aca34402", "hash_input_tokens": "a3cb206ffeb63ba2", "hash_cont_tokens": "f0c5cda13b476af9" }, "truncated": 0, "non_truncated": 95, "padded": 190, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Literature|0": { "hashes": { "hash_examples": "4305379fd46be5d8", "hash_full_prompts": "4305379fd46be5d8", "hash_input_tokens": "2652fcd6edad36a7", "hash_cont_tokens": "0755257cc7305005" }, "truncated": 0, "non_truncated": 145, "padded": 290, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Math|0": { "hashes": { "hash_examples": "dec621144f4d28be", "hash_full_prompts": "dec621144f4d28be", "hash_input_tokens": "200b5bec48aac8db", "hash_cont_tokens": "df2048053ef0446e" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Medicine|0": { "hashes": { "hash_examples": "2b344cdae9495ff2", "hash_full_prompts": "2b344cdae9495ff2", "hash_input_tokens": "274a98a2fdabb049", "hash_cont_tokens": "0755257cc7305005" }, "truncated": 0, "non_truncated": 145, "padded": 290, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Music|0": { "hashes": { "hash_examples": "0c54624d881944ce", "hash_full_prompts": "0c54624d881944ce", "hash_input_tokens": "306469dc6d45872d", "hash_cont_tokens": "10c5e10b463fd8d4" }, "truncated": 0, "non_truncated": 139, "padded": 278, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Ornament|0": { "hashes": { "hash_examples": "251a4a84289d8bc1", "hash_full_prompts": "251a4a84289d8bc1", "hash_input_tokens": "c93b68fc0bc9eca4", "hash_cont_tokens": "df2048053ef0446e" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Philosophy|0": { "hashes": { "hash_examples": "3f86fb9c94c13d22", "hash_full_prompts": "3f86fb9c94c13d22", "hash_input_tokens": "ab9807639025e5ce", "hash_cont_tokens": "0755257cc7305005" }, "truncated": 0, "non_truncated": 145, "padded": 290, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Physics_and_Chemistry|0": { "hashes": { "hash_examples": "8fec65af3695b62a", "hash_full_prompts": "8fec65af3695b62a", "hash_input_tokens": "ec191df1850ef542", "hash_cont_tokens": "df2048053ef0446e" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Wedding|0": { "hashes": { "hash_examples": "9cc3477184d7a4b8", "hash_full_prompts": "9cc3477184d7a4b8", "hash_input_tokens": "56bad6eec7d7d1d6", "hash_cont_tokens": "df2048053ef0446e" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Bahrain|0": { "hashes": { "hash_examples": "c92e803a0fa8b9e2", "hash_full_prompts": "c92e803a0fa8b9e2", "hash_input_tokens": "eeac285fa1a8382f", "hash_cont_tokens": "f25110480285096a" }, "truncated": 0, "non_truncated": 45, "padded": 90, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Comoros|0": { "hashes": { "hash_examples": "06e5d4bba8e54cae", "hash_full_prompts": "06e5d4bba8e54cae", "hash_input_tokens": "c035fbed0afa27d0", "hash_cont_tokens": "f25110480285096a" }, "truncated": 0, "non_truncated": 45, "padded": 90, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Egypt_modern|0": { "hashes": { "hash_examples": "c6ec369164f93446", "hash_full_prompts": "c6ec369164f93446", "hash_input_tokens": "569f1878bcd228b6", "hash_cont_tokens": "f0c5cda13b476af9" }, "truncated": 0, "non_truncated": 95, "padded": 190, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:InfluenceFromAncientEgypt|0": { "hashes": { "hash_examples": "b9d56d74818b9bd4", "hash_full_prompts": "b9d56d74818b9bd4", "hash_input_tokens": "b63c76706e068813", "hash_cont_tokens": "df2048053ef0446e" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:InfluenceFromByzantium|0": { "hashes": { "hash_examples": "5316c9624e7e59b8", "hash_full_prompts": "5316c9624e7e59b8", "hash_input_tokens": "d7d78fd2767c1d48", "hash_cont_tokens": "0755257cc7305005" }, "truncated": 0, "non_truncated": 145, "padded": 290, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:InfluenceFromChina|0": { "hashes": { "hash_examples": "87894bce95a56411", "hash_full_prompts": "87894bce95a56411", "hash_input_tokens": "474fa2194f35d60a", "hash_cont_tokens": "df2048053ef0446e" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:InfluenceFromGreece|0": { "hashes": { "hash_examples": "0baa78a27e469312", "hash_full_prompts": "0baa78a27e469312", "hash_input_tokens": "977c70c92b71b363", "hash_cont_tokens": "df2048053ef0446e" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:InfluenceFromIslam|0": { "hashes": { "hash_examples": "0c2532cde6541ff2", "hash_full_prompts": "0c2532cde6541ff2", "hash_input_tokens": "07557f2cb421292b", "hash_cont_tokens": "0755257cc7305005" }, "truncated": 0, "non_truncated": 145, "padded": 290, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:InfluenceFromPersia|0": { "hashes": { "hash_examples": "efcd8112dc53c6e5", "hash_full_prompts": "efcd8112dc53c6e5", "hash_input_tokens": "e7363de3e7e4ed8b", "hash_cont_tokens": "82f65a2c40e9602d" }, "truncated": 0, "non_truncated": 175, "padded": 350, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:InfluenceFromRome|0": { "hashes": { "hash_examples": "9db61480e2e85fd3", "hash_full_prompts": "9db61480e2e85fd3", "hash_input_tokens": "470b25965c404b4e", "hash_cont_tokens": "df2048053ef0446e" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Iraq|0": { "hashes": { "hash_examples": "96dac3dfa8d2f41f", "hash_full_prompts": "96dac3dfa8d2f41f", "hash_input_tokens": "c9a36b37100557ce", "hash_cont_tokens": "0da22f62ac434c4e" }, "truncated": 0, "non_truncated": 85, "padded": 170, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Islam_Education|0": { "hashes": { "hash_examples": "0d80355f6a4cb51b", "hash_full_prompts": "0d80355f6a4cb51b", "hash_input_tokens": "449d829677133c6b", "hash_cont_tokens": "df2048053ef0446e" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Islam_branches_and_schools|0": { "hashes": { "hash_examples": "5cedce1be2c3ad50", "hash_full_prompts": "5cedce1be2c3ad50", "hash_input_tokens": "e5881e19b982c470", "hash_cont_tokens": "82f65a2c40e9602d" }, "truncated": 0, "non_truncated": 175, "padded": 350, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Islamic_law_system|0": { "hashes": { "hash_examples": "c0e6db8bc84e105e", "hash_full_prompts": "c0e6db8bc84e105e", "hash_input_tokens": "cbe6cca36083a812", "hash_cont_tokens": "df2048053ef0446e" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Jordan|0": { "hashes": { "hash_examples": "33deb5b4e5ddd6a1", "hash_full_prompts": "33deb5b4e5ddd6a1", "hash_input_tokens": "4401790e7cfebbab", "hash_cont_tokens": "f25110480285096a" }, "truncated": 0, "non_truncated": 45, "padded": 90, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Kuwait|0": { "hashes": { "hash_examples": "eb41773346d7c46c", "hash_full_prompts": "eb41773346d7c46c", "hash_input_tokens": "c965592ab6275416", "hash_cont_tokens": "f25110480285096a" }, "truncated": 0, "non_truncated": 45, "padded": 90, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Lebanon|0": { "hashes": { "hash_examples": "25932dbf4c13d34f", "hash_full_prompts": "25932dbf4c13d34f", "hash_input_tokens": "d4440612c57b9b10", "hash_cont_tokens": "f25110480285096a" }, "truncated": 0, "non_truncated": 45, "padded": 90, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Libya|0": { "hashes": { "hash_examples": "f2c4db63cd402926", "hash_full_prompts": "f2c4db63cd402926", "hash_input_tokens": "4c12401210fb3c3d", "hash_cont_tokens": "f25110480285096a" }, "truncated": 0, "non_truncated": 45, "padded": 90, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Mauritania|0": { "hashes": { "hash_examples": "8723ab5fdf286b54", "hash_full_prompts": "8723ab5fdf286b54", "hash_input_tokens": "b527eb6393ee089f", "hash_cont_tokens": "f25110480285096a" }, "truncated": 0, "non_truncated": 45, "padded": 90, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Mesopotamia_civilization|0": { "hashes": { "hash_examples": "c33f5502a6130ca9", "hash_full_prompts": "c33f5502a6130ca9", "hash_input_tokens": "11ce41225328c1ae", "hash_cont_tokens": "c7cfbf3d87628357" }, "truncated": 0, "non_truncated": 155, "padded": 310, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Morocco|0": { "hashes": { "hash_examples": "588a5ed27904b1ae", "hash_full_prompts": "588a5ed27904b1ae", "hash_input_tokens": "41599d6129b22cde", "hash_cont_tokens": "f25110480285096a" }, "truncated": 0, "non_truncated": 45, "padded": 90, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Oman|0": { "hashes": { "hash_examples": "d447c52b94248b69", "hash_full_prompts": "d447c52b94248b69", "hash_input_tokens": "c603e0e6d52801c4", "hash_cont_tokens": "f25110480285096a" }, "truncated": 0, "non_truncated": 45, "padded": 90, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Palestine|0": { "hashes": { "hash_examples": "19197e076ad14ff5", "hash_full_prompts": "19197e076ad14ff5", "hash_input_tokens": "7eba46cf57559300", "hash_cont_tokens": "0da22f62ac434c4e" }, "truncated": 0, "non_truncated": 85, "padded": 170, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Qatar|0": { "hashes": { "hash_examples": "cf0736fa185b28f6", "hash_full_prompts": "cf0736fa185b28f6", "hash_input_tokens": "8dc45c023209c5e3", "hash_cont_tokens": "f25110480285096a" }, "truncated": 0, "non_truncated": 45, "padded": 90, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Saudi_Arabia|0": { "hashes": { "hash_examples": "69beda6e1b85a08d", "hash_full_prompts": "69beda6e1b85a08d", "hash_input_tokens": "16ca5b40e6c7aa1f", "hash_cont_tokens": "df2048053ef0446e" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Somalia|0": { "hashes": { "hash_examples": "b387940c65784fbf", "hash_full_prompts": "b387940c65784fbf", "hash_input_tokens": "08527468399605bb", "hash_cont_tokens": "f25110480285096a" }, "truncated": 0, "non_truncated": 45, "padded": 90, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Sudan|0": { "hashes": { "hash_examples": "e02c32b9d2dd0c3f", "hash_full_prompts": "e02c32b9d2dd0c3f", "hash_input_tokens": "164b8f374de60e91", "hash_cont_tokens": "f25110480285096a" }, "truncated": 0, "non_truncated": 45, "padded": 90, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Syria|0": { "hashes": { "hash_examples": "60a6f8fe73bda4bb", "hash_full_prompts": "60a6f8fe73bda4bb", "hash_input_tokens": "4d18abb875d5151a", "hash_cont_tokens": "f25110480285096a" }, "truncated": 0, "non_truncated": 45, "padded": 90, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Tunisia|0": { "hashes": { "hash_examples": "34bb15d3830c5649", "hash_full_prompts": "34bb15d3830c5649", "hash_input_tokens": "e00dec51260b3a6a", "hash_cont_tokens": "f25110480285096a" }, "truncated": 0, "non_truncated": 45, "padded": 90, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:United_Arab_Emirates|0": { "hashes": { "hash_examples": "98a0ba78172718ce", "hash_full_prompts": "98a0ba78172718ce", "hash_input_tokens": "8ad8386221e362ec", "hash_cont_tokens": "0da22f62ac434c4e" }, "truncated": 0, "non_truncated": 85, "padded": 170, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Yemen|0": { "hashes": { "hash_examples": "18e9bcccbb4ced7a", "hash_full_prompts": "18e9bcccbb4ced7a", "hash_input_tokens": "ad1de50a6870671f", "hash_cont_tokens": "146fa23edce78d6b" }, "truncated": 0, "non_truncated": 10, "padded": 20, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:communication|0": { "hashes": { "hash_examples": "9ff28ab5eab5c97b", "hash_full_prompts": "9ff28ab5eab5c97b", "hash_input_tokens": "b2060f8971a94fe3", "hash_cont_tokens": "3c78d3a38133f37b" }, "truncated": 0, "non_truncated": 364, "padded": 728, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:computer_and_phone|0": { "hashes": { "hash_examples": "37bac2f086aaf6c2", "hash_full_prompts": "37bac2f086aaf6c2", "hash_input_tokens": "2e52e3d1f1c3b65c", "hash_cont_tokens": "3bdd2d06ac92d4c1" }, "truncated": 0, "non_truncated": 295, "padded": 590, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:daily_life|0": { "hashes": { "hash_examples": "bf07363c1c252e2f", "hash_full_prompts": "bf07363c1c252e2f", "hash_input_tokens": "8713e036b9baa51b", "hash_cont_tokens": "ddb08393768e7592" }, "truncated": 0, "non_truncated": 337, "padded": 674, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:entertainment|0": { "hashes": { "hash_examples": "37077bc00f0ac56a", "hash_full_prompts": "37077bc00f0ac56a", "hash_input_tokens": "39a9baea2aa52252", "hash_cont_tokens": "3bdd2d06ac92d4c1" }, "truncated": 0, "non_truncated": 295, "padded": 590, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|alghafa:mcq_exams_test_ar|0": { "hashes": { "hash_examples": "c07a5e78c5c0b8fe", "hash_full_prompts": "c07a5e78c5c0b8fe", "hash_input_tokens": "eb3a84ec90485a2c", "hash_cont_tokens": "397440185e16e7dc" }, "truncated": 0, "non_truncated": 557, "padded": 2228, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|alghafa:meta_ar_dialects|0": { "hashes": { "hash_examples": "c0b6081f83e14064", "hash_full_prompts": "c0b6081f83e14064", "hash_input_tokens": "d3ae47d9938f6667", "hash_cont_tokens": "367b4b49650355e2" }, "truncated": 0, "non_truncated": 5395, "padded": 21457, "non_padded": 123, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|alghafa:meta_ar_msa|0": { "hashes": { "hash_examples": "64eb78a7c5b7484b", "hash_full_prompts": "64eb78a7c5b7484b", "hash_input_tokens": "31ed66292f6a9975", "hash_cont_tokens": "82ab2625160d0dcf" }, "truncated": 0, "non_truncated": 895, "padded": 3548, "non_padded": 32, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|alghafa:multiple_choice_facts_truefalse_balanced_task|0": { "hashes": { "hash_examples": "54fc3502c1c02c06", "hash_full_prompts": "54fc3502c1c02c06", "hash_input_tokens": "421ed1c386d8e4f4", "hash_cont_tokens": "87f39c4c80644a76" }, "truncated": 0, "non_truncated": 75, "padded": 148, "non_padded": 2, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|alghafa:multiple_choice_grounded_statement_soqal_task|0": { "hashes": { "hash_examples": "46572d83696552ae", "hash_full_prompts": "46572d83696552ae", "hash_input_tokens": "816a46c0e9a78a26", "hash_cont_tokens": "d0eb3cf49e97b2a0" }, "truncated": 0, "non_truncated": 150, "padded": 743, "non_padded": 7, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|alghafa:multiple_choice_grounded_statement_xglue_mlqa_task|0": { "hashes": { "hash_examples": "f430d97ff715bc1c", "hash_full_prompts": "f430d97ff715bc1c", "hash_input_tokens": "74005ec37454db29", "hash_cont_tokens": "2f0684addd4381fc" }, "truncated": 0, "non_truncated": 150, "padded": 744, "non_padded": 6, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|alghafa:multiple_choice_rating_sentiment_no_neutral_task|0": { "hashes": { "hash_examples": "6b70a7416584f98c", "hash_full_prompts": "6b70a7416584f98c", "hash_input_tokens": "d652b9bd9ee0113a", "hash_cont_tokens": "31a05c87a910bd45" }, "truncated": 0, "non_truncated": 7995, "padded": 15990, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|alghafa:multiple_choice_rating_sentiment_task|0": { "hashes": { "hash_examples": "bc2005cc9d2f436e", "hash_full_prompts": "bc2005cc9d2f436e", "hash_input_tokens": "5a7eae6bddc3298b", "hash_cont_tokens": "8fd1b65b30f88ff2" }, "truncated": 0, "non_truncated": 5995, "padded": 17985, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|alghafa:multiple_choice_sentiment_task|0": { "hashes": { "hash_examples": "6fb0e254ea5945d8", "hash_full_prompts": "6fb0e254ea5945d8", "hash_input_tokens": "a561e22bb68b0ed3", "hash_cont_tokens": "23cf5b9b88b866cc" }, "truncated": 0, "non_truncated": 1720, "padded": 5158, "non_padded": 2, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_exams|0": { "hashes": { "hash_examples": "6d721df351722656", "hash_full_prompts": "6d721df351722656", "hash_input_tokens": "7109fe6dbd9b70b7", "hash_cont_tokens": "cd403399a6437d4c" }, "truncated": 0, "non_truncated": 537, "padded": 2148, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:abstract_algebra|0": { "hashes": { "hash_examples": "f2ddca8f45c0a511", "hash_full_prompts": "f2ddca8f45c0a511", "hash_input_tokens": "8b944a91b59d6f10", "hash_cont_tokens": "4e81a5650d054535" }, "truncated": 0, "non_truncated": 100, "padded": 400, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:anatomy|0": { "hashes": { "hash_examples": "dfdbc1b83107668d", "hash_full_prompts": "dfdbc1b83107668d", "hash_input_tokens": "e57c9c9bcdb717a7", "hash_cont_tokens": "7fc4cc50abc94fab" }, "truncated": 0, "non_truncated": 135, "padded": 540, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:astronomy|0": { "hashes": { "hash_examples": "9736a606002a848e", "hash_full_prompts": "9736a606002a848e", "hash_input_tokens": "1670d5a7cbb0d133", "hash_cont_tokens": "3b563d68fea22fed" }, "truncated": 0, "non_truncated": 152, "padded": 608, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:business_ethics|0": { "hashes": { "hash_examples": "735e452fbb6dc63d", "hash_full_prompts": "735e452fbb6dc63d", "hash_input_tokens": "bfe92fe9f80a2df2", "hash_cont_tokens": "4e81a5650d054535" }, "truncated": 0, "non_truncated": 100, "padded": 400, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:clinical_knowledge|0": { "hashes": { "hash_examples": "6ab0ca4da98aedcf", "hash_full_prompts": "6ab0ca4da98aedcf", "hash_input_tokens": "312df4688e9f5cd7", "hash_cont_tokens": "30d2f27eabc17d43" }, "truncated": 0, "non_truncated": 265, "padded": 1060, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:college_biology|0": { "hashes": { "hash_examples": "17e4e390848018a4", "hash_full_prompts": "17e4e390848018a4", "hash_input_tokens": "8d88e1d1cf2b750c", "hash_cont_tokens": "877270cee1a485f4" }, "truncated": 0, "non_truncated": 144, "padded": 576, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:college_chemistry|0": { "hashes": { "hash_examples": "4abb169f6dfd234b", "hash_full_prompts": "4abb169f6dfd234b", "hash_input_tokens": "e8fda15a414404a3", "hash_cont_tokens": "4e81a5650d054535" }, "truncated": 0, "non_truncated": 100, "padded": 400, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:college_computer_science|0": { "hashes": { "hash_examples": "a369e2e941358a1e", "hash_full_prompts": "a369e2e941358a1e", "hash_input_tokens": "9e1b57290de8307d", "hash_cont_tokens": "4e81a5650d054535" }, "truncated": 0, "non_truncated": 100, "padded": 400, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:college_mathematics|0": { "hashes": { "hash_examples": "d7be03b8b6020bff", "hash_full_prompts": "d7be03b8b6020bff", "hash_input_tokens": "49a2edd579a65feb", "hash_cont_tokens": "4e81a5650d054535" }, "truncated": 0, "non_truncated": 100, "padded": 400, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:college_medicine|0": { "hashes": { "hash_examples": "0518a00f097346bf", "hash_full_prompts": "0518a00f097346bf", "hash_input_tokens": "36879e06dda749e3", "hash_cont_tokens": "d42e1575e54d677a" }, "truncated": 0, "non_truncated": 173, "padded": 692, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:college_physics|0": { "hashes": { "hash_examples": "5d842cd49bc70e12", "hash_full_prompts": "5d842cd49bc70e12", "hash_input_tokens": "0e13baeb3424360c", "hash_cont_tokens": "5b6bb337d096385e" }, "truncated": 0, "non_truncated": 102, "padded": 408, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:computer_security|0": { "hashes": { "hash_examples": "8e85d9f85be9b32f", "hash_full_prompts": "8e85d9f85be9b32f", "hash_input_tokens": "c062383957d378ee", "hash_cont_tokens": "4e81a5650d054535" }, "truncated": 0, "non_truncated": 100, "padded": 400, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:conceptual_physics|0": { "hashes": { "hash_examples": "7964b55a0a49502b", "hash_full_prompts": "7964b55a0a49502b", "hash_input_tokens": "4e0386d348c5440c", "hash_cont_tokens": "bf11249a64235593" }, "truncated": 0, "non_truncated": 235, "padded": 940, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:econometrics|0": { "hashes": { "hash_examples": "1e192eae38347257", "hash_full_prompts": "1e192eae38347257", "hash_input_tokens": "dac555c7ff9a1af3", "hash_cont_tokens": "379bc8af1b012d6d" }, "truncated": 0, "non_truncated": 114, "padded": 456, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:electrical_engineering|0": { "hashes": { "hash_examples": "cf97671d5c441da1", "hash_full_prompts": "cf97671d5c441da1", "hash_input_tokens": "07d8f7cdeec590de", "hash_cont_tokens": "d925470948182565" }, "truncated": 0, "non_truncated": 145, "padded": 580, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:elementary_mathematics|0": { "hashes": { "hash_examples": "6f49107ed43c40c5", "hash_full_prompts": "6f49107ed43c40c5", "hash_input_tokens": "a595ff836d347263", "hash_cont_tokens": "2311cea390157f8f" }, "truncated": 0, "non_truncated": 378, "padded": 1512, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:formal_logic|0": { "hashes": { "hash_examples": "7922c376008ba77b", "hash_full_prompts": "7922c376008ba77b", "hash_input_tokens": "ee8f2b9b29e2a346", "hash_cont_tokens": "ad34ef90308d0cb6" }, "truncated": 0, "non_truncated": 126, "padded": 504, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:global_facts|0": { "hashes": { "hash_examples": "11f9813185047d5b", "hash_full_prompts": "11f9813185047d5b", "hash_input_tokens": "e1d7b74a2de341a3", "hash_cont_tokens": "4e81a5650d054535" }, "truncated": 0, "non_truncated": 100, "padded": 400, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:high_school_biology|0": { "hashes": { "hash_examples": "2a804b1d90cbe66e", "hash_full_prompts": "2a804b1d90cbe66e", "hash_input_tokens": "3f6ff5e6f707e1e1", "hash_cont_tokens": "76f445fb779acf11" }, "truncated": 0, "non_truncated": 310, "padded": 1240, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:high_school_chemistry|0": { "hashes": { "hash_examples": "0032168adabc53b4", "hash_full_prompts": "0032168adabc53b4", "hash_input_tokens": "f9e3997f76289fb4", "hash_cont_tokens": "0f535f135c97cb24" }, "truncated": 0, "non_truncated": 203, "padded": 808, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:high_school_computer_science|0": { "hashes": { "hash_examples": "f2fb8740f9df980f", "hash_full_prompts": "f2fb8740f9df980f", "hash_input_tokens": "1252ce54c1a9de2f", "hash_cont_tokens": "4e81a5650d054535" }, "truncated": 0, "non_truncated": 100, "padded": 400, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:high_school_european_history|0": { "hashes": { "hash_examples": "73509021e7e66435", "hash_full_prompts": "73509021e7e66435", "hash_input_tokens": "b671539a14384656", "hash_cont_tokens": "dd25c24efc961a9f" }, "truncated": 0, "non_truncated": 165, "padded": 660, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:high_school_geography|0": { "hashes": { "hash_examples": "9e08d1894940ff42", "hash_full_prompts": "9e08d1894940ff42", "hash_input_tokens": "4f4ee652b85b1ca0", "hash_cont_tokens": "003ba9ddfb050dad" }, "truncated": 0, "non_truncated": 198, "padded": 792, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:high_school_government_and_politics|0": { "hashes": { "hash_examples": "64b7e97817ca6c76", "hash_full_prompts": "64b7e97817ca6c76", "hash_input_tokens": "a0e58ac00876e781", "hash_cont_tokens": "24e0a0a2c8966c25" }, "truncated": 0, "non_truncated": 193, "padded": 772, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:high_school_macroeconomics|0": { "hashes": { "hash_examples": "9f582da8534bd2ef", "hash_full_prompts": "9f582da8534bd2ef", "hash_input_tokens": "d0e6c8332c0ee2da", "hash_cont_tokens": "9f2551c392fb77e6" }, "truncated": 0, "non_truncated": 390, "padded": 1560, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:high_school_mathematics|0": { "hashes": { "hash_examples": "fd54f1c10d423c51", "hash_full_prompts": "fd54f1c10d423c51", "hash_input_tokens": "9ac488190b385956", "hash_cont_tokens": "a6e70ae02a6cf6d1" }, "truncated": 0, "non_truncated": 270, "padded": 1080, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:high_school_microeconomics|0": { "hashes": { "hash_examples": "7037896925aaf42f", "hash_full_prompts": "7037896925aaf42f", "hash_input_tokens": "cf7e43e43dbd324f", "hash_cont_tokens": "af5f8ec7c18336c3" }, "truncated": 0, "non_truncated": 238, "padded": 952, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:high_school_physics|0": { "hashes": { "hash_examples": "60c3776215167dae", "hash_full_prompts": "60c3776215167dae", "hash_input_tokens": "293f94e1be96821c", "hash_cont_tokens": "42eef2b2c6af8a8c" }, "truncated": 0, "non_truncated": 151, "padded": 604, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:high_school_psychology|0": { "hashes": { "hash_examples": "61176bfd5da1298f", "hash_full_prompts": "61176bfd5da1298f", "hash_input_tokens": "d37c791b6afcf817", "hash_cont_tokens": "5b6f8628f39e39fb" }, "truncated": 0, "non_truncated": 545, "padded": 2180, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:high_school_statistics|0": { "hashes": { "hash_examples": "40dfeebd1ea10f76", "hash_full_prompts": "40dfeebd1ea10f76", "hash_input_tokens": "777cb63c6cba7daa", "hash_cont_tokens": "ae3f2e7e623396c6" }, "truncated": 0, "non_truncated": 216, "padded": 860, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:high_school_us_history|0": { "hashes": { "hash_examples": "03daa510ba917f4d", "hash_full_prompts": "03daa510ba917f4d", "hash_input_tokens": "abaa5f30931fecf8", "hash_cont_tokens": "f0a9bb4a70fb64ab" }, "truncated": 0, "non_truncated": 204, "padded": 816, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:high_school_world_history|0": { "hashes": { "hash_examples": "be075ffd579f43c2", "hash_full_prompts": "be075ffd579f43c2", "hash_input_tokens": "7418e37be8da4736", "hash_cont_tokens": "36ab695092e41a42" }, "truncated": 0, "non_truncated": 237, "padded": 948, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:human_aging|0": { "hashes": { "hash_examples": "caa5b69f640bd1ef", "hash_full_prompts": "caa5b69f640bd1ef", "hash_input_tokens": "f94410e3073ad839", "hash_cont_tokens": "3a8d9d501baf9044" }, "truncated": 0, "non_truncated": 223, "padded": 892, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:human_sexuality|0": { "hashes": { "hash_examples": "5ed2e38fb25a3767", "hash_full_prompts": "5ed2e38fb25a3767", "hash_input_tokens": "82e312b2bd932179", "hash_cont_tokens": "9b97f9e5a094f320" }, "truncated": 0, "non_truncated": 131, "padded": 524, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:international_law|0": { "hashes": { "hash_examples": "4e3e9e28d1b96484", "hash_full_prompts": "4e3e9e28d1b96484", "hash_input_tokens": "38a9532d20929fdb", "hash_cont_tokens": "f68a3e36b43b851f" }, "truncated": 0, "non_truncated": 121, "padded": 484, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:jurisprudence|0": { "hashes": { "hash_examples": "e264b755366310b3", "hash_full_prompts": "e264b755366310b3", "hash_input_tokens": "066c3adfb30420c7", "hash_cont_tokens": "45126ebf1439dd8b" }, "truncated": 0, "non_truncated": 108, "padded": 432, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:logical_fallacies|0": { "hashes": { "hash_examples": "a4ab6965a3e38071", "hash_full_prompts": "a4ab6965a3e38071", "hash_input_tokens": "6de4596a9a390603", "hash_cont_tokens": "f8cbb08966a240c0" }, "truncated": 0, "non_truncated": 163, "padded": 652, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:machine_learning|0": { "hashes": { "hash_examples": "b92320efa6636b40", "hash_full_prompts": "b92320efa6636b40", "hash_input_tokens": "4371bf80c091f2d2", "hash_cont_tokens": "d50fff2aa6530403" }, "truncated": 0, "non_truncated": 112, "padded": 448, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:management|0": { "hashes": { "hash_examples": "c9ee4872a850fe20", "hash_full_prompts": "c9ee4872a850fe20", "hash_input_tokens": "aa370e6a912e584f", "hash_cont_tokens": "cd406f58542477f4" }, "truncated": 0, "non_truncated": 103, "padded": 412, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:marketing|0": { "hashes": { "hash_examples": "0c151b70f6a047e3", "hash_full_prompts": "0c151b70f6a047e3", "hash_input_tokens": "b6b419bbb131cbc2", "hash_cont_tokens": "1c70d08c5c33a0e0" }, "truncated": 0, "non_truncated": 234, "padded": 936, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:medical_genetics|0": { "hashes": { "hash_examples": "513f6cb8fca3a24e", "hash_full_prompts": "513f6cb8fca3a24e", "hash_input_tokens": "df7a53920afc28b8", "hash_cont_tokens": "4e81a5650d054535" }, "truncated": 0, "non_truncated": 100, "padded": 400, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:miscellaneous|0": { "hashes": { "hash_examples": "259a190d635331db", "hash_full_prompts": "259a190d635331db", "hash_input_tokens": "775a5731397f0895", "hash_cont_tokens": "09eb6572f8454e64" }, "truncated": 0, "non_truncated": 783, "padded": 3132, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:moral_disputes|0": { "hashes": { "hash_examples": "b85052c48a0b7bc3", "hash_full_prompts": "b85052c48a0b7bc3", "hash_input_tokens": "262d8045bdeb9bf6", "hash_cont_tokens": "2bfa7044a1112e43" }, "truncated": 0, "non_truncated": 346, "padded": 1384, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:moral_scenarios|0": { "hashes": { "hash_examples": "28d0b069ef00dd00", "hash_full_prompts": "28d0b069ef00dd00", "hash_input_tokens": "04d5a277879b23d6", "hash_cont_tokens": "918dfd392887200c" }, "truncated": 0, "non_truncated": 895, "padded": 3580, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:nutrition|0": { "hashes": { "hash_examples": "00c9bc5f1d305b2f", "hash_full_prompts": "00c9bc5f1d305b2f", "hash_input_tokens": "96fe3e12a6d50395", "hash_cont_tokens": "8d828cbeed27e465" }, "truncated": 0, "non_truncated": 306, "padded": 1224, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:philosophy|0": { "hashes": { "hash_examples": "a458c08454a3fd5f", "hash_full_prompts": "a458c08454a3fd5f", "hash_input_tokens": "c429b48d608f4a3c", "hash_cont_tokens": "03445051a61ba689" }, "truncated": 0, "non_truncated": 311, "padded": 1244, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:prehistory|0": { "hashes": { "hash_examples": "d6a0ecbdbb670e9c", "hash_full_prompts": "d6a0ecbdbb670e9c", "hash_input_tokens": "36fc61dcf61f7dfd", "hash_cont_tokens": "3c33de1f03e37f23" }, "truncated": 0, "non_truncated": 324, "padded": 1296, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:professional_accounting|0": { "hashes": { "hash_examples": "b4a95fe480b6540e", "hash_full_prompts": "b4a95fe480b6540e", "hash_input_tokens": "45c9bd7a35b94f44", "hash_cont_tokens": "8e6a1fea90b692b6" }, "truncated": 0, "non_truncated": 282, "padded": 1124, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:professional_law|0": { "hashes": { "hash_examples": "c2be9651cdbdde3b", "hash_full_prompts": "c2be9651cdbdde3b", "hash_input_tokens": "7685e3c1b4233feb", "hash_cont_tokens": "45c75a65f459e785" }, "truncated": 0, "non_truncated": 1534, "padded": 6132, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:professional_medicine|0": { "hashes": { "hash_examples": "26ce92416288f273", "hash_full_prompts": "26ce92416288f273", "hash_input_tokens": "386b75d627619e0b", "hash_cont_tokens": "2cf084b31713327a" }, "truncated": 0, "non_truncated": 272, "padded": 1088, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:professional_psychology|0": { "hashes": { "hash_examples": "71ea5f182ea9a641", "hash_full_prompts": "71ea5f182ea9a641", "hash_input_tokens": "8798ba2d0bb8a2d2", "hash_cont_tokens": "67a1ec0b9dc64487" }, "truncated": 0, "non_truncated": 612, "padded": 2448, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:public_relations|0": { "hashes": { "hash_examples": "125adc21f91f8d77", "hash_full_prompts": "125adc21f91f8d77", "hash_input_tokens": "e4a879fbe1591d71", "hash_cont_tokens": "604b9fcdab9c1bf7" }, "truncated": 0, "non_truncated": 110, "padded": 440, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:security_studies|0": { "hashes": { "hash_examples": "3c18b216c099fb26", "hash_full_prompts": "3c18b216c099fb26", "hash_input_tokens": "d15378256599a695", "hash_cont_tokens": "195bcea7f180ce88" }, "truncated": 0, "non_truncated": 245, "padded": 976, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:sociology|0": { "hashes": { "hash_examples": "3f2a9634cef7417d", "hash_full_prompts": "3f2a9634cef7417d", "hash_input_tokens": "96830e4978160419", "hash_cont_tokens": "aeb7c6ca96da4cde" }, "truncated": 0, "non_truncated": 201, "padded": 804, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:us_foreign_policy|0": { "hashes": { "hash_examples": "22249da54056475e", "hash_full_prompts": "22249da54056475e", "hash_input_tokens": "64a4d1f8583be500", "hash_cont_tokens": "4e81a5650d054535" }, "truncated": 0, "non_truncated": 100, "padded": 400, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:virology|0": { "hashes": { "hash_examples": "9d194b9471dc624e", "hash_full_prompts": "9d194b9471dc624e", "hash_input_tokens": "45b4b4c5ec31f8ca", "hash_cont_tokens": "34ec21d703ad122c" }, "truncated": 0, "non_truncated": 166, "padded": 664, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:world_religions|0": { "hashes": { "hash_examples": "229e5fe50082b064", "hash_full_prompts": "229e5fe50082b064", "hash_input_tokens": "4fd2a573d9c1897e", "hash_cont_tokens": "2d5fef82a3ab1e3d" }, "truncated": 0, "non_truncated": 171, "padded": 683, "non_padded": 1, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arc_challenge_okapi_ar|0": { "hashes": { "hash_examples": "ab893807673bc355", "hash_full_prompts": "ab893807673bc355", "hash_input_tokens": "6697aff32effa929", "hash_cont_tokens": "38fc41f1c6a6aa66" }, "truncated": 0, "non_truncated": 1160, "padded": 4557, "non_padded": 83, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arc_easy_ar|0": { "hashes": { "hash_examples": "acb688624acc3d04", "hash_full_prompts": "acb688624acc3d04", "hash_input_tokens": "a38e1c2bc322ccee", "hash_cont_tokens": "f0ea0f73d2dad503" }, "truncated": 0, "non_truncated": 2364, "padded": 9253, "non_padded": 203, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|boolq_ar|0": { "hashes": { "hash_examples": "48355a67867e0c32", "hash_full_prompts": "48355a67867e0c32", "hash_input_tokens": "62baaca9f1207831", "hash_cont_tokens": "62371a141f3b36ed" }, "truncated": 0, "non_truncated": 3260, "padded": 6486, "non_padded": 34, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|copa_ext_ar|0": { "hashes": { "hash_examples": "9bb83301bb72eecf", "hash_full_prompts": "9bb83301bb72eecf", "hash_input_tokens": "e6931628b1bc4e5c", "hash_cont_tokens": "68c1c3753a7311cc" }, "truncated": 0, "non_truncated": 90, "padded": 179, "non_padded": 1, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|hellaswag_okapi_ar|0": { "hashes": { "hash_examples": "6e8cf57a322dfadd", "hash_full_prompts": "6e8cf57a322dfadd", "hash_input_tokens": "856eb8ed2d5a86cd", "hash_cont_tokens": "282da749e2c8c3e4" }, "truncated": 0, "non_truncated": 9171, "padded": 36555, "non_padded": 129, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|openbook_qa_ext_ar|0": { "hashes": { "hash_examples": "923d41eb0aca93eb", "hash_full_prompts": "923d41eb0aca93eb", "hash_input_tokens": "f869e9ab2625572b", "hash_cont_tokens": "b2aa6bc962a65ac4" }, "truncated": 0, "non_truncated": 495, "padded": 1929, "non_padded": 51, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|piqa_ar|0": { "hashes": { "hash_examples": "94bc205a520d3ea0", "hash_full_prompts": "94bc205a520d3ea0", "hash_input_tokens": "abe3e59d5e940785", "hash_cont_tokens": "b9aca1a3effc4c6a" }, "truncated": 0, "non_truncated": 1833, "padded": 3582, "non_padded": 84, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|race_ar|0": { "hashes": { "hash_examples": "de65130bae647516", "hash_full_prompts": "de65130bae647516", "hash_input_tokens": "b572a5be0bf6fb80", "hash_cont_tokens": "086b72e2ee87f4e6" }, "truncated": 0, "non_truncated": 4929, "padded": 19691, "non_padded": 25, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|sciq_ar|0": { "hashes": { "hash_examples": "524f93e7a59e7de9", "hash_full_prompts": "524f93e7a59e7de9", "hash_input_tokens": "6475b674de627b7d", "hash_cont_tokens": "ad9570eea379ad7f" }, "truncated": 0, "non_truncated": 995, "padded": 3909, "non_padded": 71, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|toxigen_ar|0": { "hashes": { "hash_examples": "1e139513004a9a2e", "hash_full_prompts": "1e139513004a9a2e", "hash_input_tokens": "36b3ec35bbe79d47", "hash_cont_tokens": "f0c35800980df007" }, "truncated": 0, "non_truncated": 935, "padded": 1820, "non_padded": 50, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "lighteval|xstory_cloze:ar|0": { "hashes": { "hash_examples": "865426a22c787481", "hash_full_prompts": "865426a22c787481", "hash_input_tokens": "b2d78abe67d1c313", "hash_cont_tokens": "2bfcaff66ac5c6c2" }, "truncated": 0, "non_truncated": 1511, "padded": 2955, "non_padded": 67, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 } }, "summary_general": { "hashes": { "hash_examples": "47212d2435b2edc2", "hash_full_prompts": "47212d2435b2edc2", "hash_input_tokens": "3f7af6d7bc705c1d", "hash_cont_tokens": "037ec4dcc7ab4958" }, "truncated": 0, "non_truncated": 72964, "padded": 234632, "non_padded": 991, "num_truncated_few_shots": 0 } }