Datasets:
OALL
/

Modalities:
Text
Formats:
json
Size:
< 1K
Libraries:
Datasets
Dask
results / Qwen /Qwen2-7B /results_2024-05-24T17-30-53.204906.json
Ali-C137's picture
Upload 10 files
318f91c verified
raw
history blame
194 kB
{
"config_general": {
"lighteval_sha": "?",
"num_fewshot_seeds": 1,
"override_batch_size": 1,
"max_samples": null,
"job_id": "",
"start_time": 477.239928401,
"end_time": 5351.0833479,
"total_evaluation_time_secondes": "4873.843419499",
"model_name": "Qwen/Qwen2-7B",
"model_sha": "84f0d5b4241701847980f1710b61ee7c5c3820c5",
"model_dtype": "torch.bfloat16",
"model_size": "15.94 GB",
"config": null
},
"results": {
"community|acva:Algeria|0": {
"acc_norm": 0.5230769230769231,
"acc_norm_stderr": 0.0358596530894741
},
"community|acva:Ancient_Egypt|0": {
"acc_norm": 0.050793650793650794,
"acc_norm_stderr": 0.01239139518482262
},
"community|acva:Arab_Empire|0": {
"acc_norm": 0.30943396226415093,
"acc_norm_stderr": 0.028450154794118627
},
"community|acva:Arabic_Architecture|0": {
"acc_norm": 0.4564102564102564,
"acc_norm_stderr": 0.035761230969912135
},
"community|acva:Arabic_Art|0": {
"acc_norm": 0.3641025641025641,
"acc_norm_stderr": 0.03454653867786389
},
"community|acva:Arabic_Astronomy|0": {
"acc_norm": 0.4666666666666667,
"acc_norm_stderr": 0.03581804596782233
},
"community|acva:Arabic_Calligraphy|0": {
"acc_norm": 0.47843137254901963,
"acc_norm_stderr": 0.0313435870640056
},
"community|acva:Arabic_Ceremony|0": {
"acc_norm": 0.518918918918919,
"acc_norm_stderr": 0.036834092970087065
},
"community|acva:Arabic_Clothing|0": {
"acc_norm": 0.5128205128205128,
"acc_norm_stderr": 0.03588610523192215
},
"community|acva:Arabic_Culture|0": {
"acc_norm": 0.23076923076923078,
"acc_norm_stderr": 0.0302493752938313
},
"community|acva:Arabic_Food|0": {
"acc_norm": 0.441025641025641,
"acc_norm_stderr": 0.0356473293185358
},
"community|acva:Arabic_Funeral|0": {
"acc_norm": 0.4,
"acc_norm_stderr": 0.050529115263991134
},
"community|acva:Arabic_Geography|0": {
"acc_norm": 0.6068965517241379,
"acc_norm_stderr": 0.040703290137070705
},
"community|acva:Arabic_History|0": {
"acc_norm": 0.30256410256410254,
"acc_norm_stderr": 0.03298070870085619
},
"community|acva:Arabic_Language_Origin|0": {
"acc_norm": 0.5473684210526316,
"acc_norm_stderr": 0.051339113773544845
},
"community|acva:Arabic_Literature|0": {
"acc_norm": 0.4689655172413793,
"acc_norm_stderr": 0.04158632762097828
},
"community|acva:Arabic_Math|0": {
"acc_norm": 0.30256410256410254,
"acc_norm_stderr": 0.03298070870085618
},
"community|acva:Arabic_Medicine|0": {
"acc_norm": 0.46206896551724136,
"acc_norm_stderr": 0.041546596717075474
},
"community|acva:Arabic_Music|0": {
"acc_norm": 0.23741007194244604,
"acc_norm_stderr": 0.036220593237998276
},
"community|acva:Arabic_Ornament|0": {
"acc_norm": 0.4717948717948718,
"acc_norm_stderr": 0.035840746749208334
},
"community|acva:Arabic_Philosophy|0": {
"acc_norm": 0.5793103448275863,
"acc_norm_stderr": 0.0411391498118926
},
"community|acva:Arabic_Physics_and_Chemistry|0": {
"acc_norm": 0.5333333333333333,
"acc_norm_stderr": 0.03581804596782232
},
"community|acva:Arabic_Wedding|0": {
"acc_norm": 0.41025641025641024,
"acc_norm_stderr": 0.03531493712326671
},
"community|acva:Bahrain|0": {
"acc_norm": 0.3111111111111111,
"acc_norm_stderr": 0.06979205927323111
},
"community|acva:Comoros|0": {
"acc_norm": 0.37777777777777777,
"acc_norm_stderr": 0.07309112127323451
},
"community|acva:Egypt_modern|0": {
"acc_norm": 0.3157894736842105,
"acc_norm_stderr": 0.04794350420740798
},
"community|acva:InfluenceFromAncientEgypt|0": {
"acc_norm": 0.6051282051282051,
"acc_norm_stderr": 0.03509545602262038
},
"community|acva:InfluenceFromByzantium|0": {
"acc_norm": 0.7172413793103448,
"acc_norm_stderr": 0.03752833958003337
},
"community|acva:InfluenceFromChina|0": {
"acc_norm": 0.26666666666666666,
"acc_norm_stderr": 0.0317493043641267
},
"community|acva:InfluenceFromGreece|0": {
"acc_norm": 0.6307692307692307,
"acc_norm_stderr": 0.034648411418637566
},
"community|acva:InfluenceFromIslam|0": {
"acc_norm": 0.296551724137931,
"acc_norm_stderr": 0.03806142687309993
},
"community|acva:InfluenceFromPersia|0": {
"acc_norm": 0.6971428571428572,
"acc_norm_stderr": 0.03483414676585986
},
"community|acva:InfluenceFromRome|0": {
"acc_norm": 0.5743589743589743,
"acc_norm_stderr": 0.03549871080367708
},
"community|acva:Iraq|0": {
"acc_norm": 0.5176470588235295,
"acc_norm_stderr": 0.05452048340661895
},
"community|acva:Islam_Education|0": {
"acc_norm": 0.4512820512820513,
"acc_norm_stderr": 0.03572709860318392
},
"community|acva:Islam_branches_and_schools|0": {
"acc_norm": 0.4342857142857143,
"acc_norm_stderr": 0.037576101528126626
},
"community|acva:Islamic_law_system|0": {
"acc_norm": 0.4256410256410256,
"acc_norm_stderr": 0.035498710803677086
},
"community|acva:Jordan|0": {
"acc_norm": 0.3333333333333333,
"acc_norm_stderr": 0.07106690545187012
},
"community|acva:Kuwait|0": {
"acc_norm": 0.26666666666666666,
"acc_norm_stderr": 0.06666666666666667
},
"community|acva:Lebanon|0": {
"acc_norm": 0.17777777777777778,
"acc_norm_stderr": 0.05763774795025094
},
"community|acva:Libya|0": {
"acc_norm": 0.4444444444444444,
"acc_norm_stderr": 0.07491109582924914
},
"community|acva:Mauritania|0": {
"acc_norm": 0.4222222222222222,
"acc_norm_stderr": 0.07446027270295805
},
"community|acva:Mesopotamia_civilization|0": {
"acc_norm": 0.5225806451612903,
"acc_norm_stderr": 0.0402500394824441
},
"community|acva:Morocco|0": {
"acc_norm": 0.2222222222222222,
"acc_norm_stderr": 0.06267511942419628
},
"community|acva:Oman|0": {
"acc_norm": 0.17777777777777778,
"acc_norm_stderr": 0.05763774795025094
},
"community|acva:Palestine|0": {
"acc_norm": 0.24705882352941178,
"acc_norm_stderr": 0.047058823529411785
},
"community|acva:Qatar|0": {
"acc_norm": 0.4,
"acc_norm_stderr": 0.07385489458759964
},
"community|acva:Saudi_Arabia|0": {
"acc_norm": 0.3282051282051282,
"acc_norm_stderr": 0.03371243782413707
},
"community|acva:Somalia|0": {
"acc_norm": 0.35555555555555557,
"acc_norm_stderr": 0.07216392363431012
},
"community|acva:Sudan|0": {
"acc_norm": 0.35555555555555557,
"acc_norm_stderr": 0.07216392363431012
},
"community|acva:Syria|0": {
"acc_norm": 0.3333333333333333,
"acc_norm_stderr": 0.07106690545187012
},
"community|acva:Tunisia|0": {
"acc_norm": 0.3111111111111111,
"acc_norm_stderr": 0.06979205927323111
},
"community|acva:United_Arab_Emirates|0": {
"acc_norm": 0.23529411764705882,
"acc_norm_stderr": 0.04628210543937907
},
"community|acva:Yemen|0": {
"acc_norm": 0.2,
"acc_norm_stderr": 0.13333333333333333
},
"community|acva:communication|0": {
"acc_norm": 0.42857142857142855,
"acc_norm_stderr": 0.025974025974025955
},
"community|acva:computer_and_phone|0": {
"acc_norm": 0.45084745762711864,
"acc_norm_stderr": 0.02901934773187137
},
"community|acva:daily_life|0": {
"acc_norm": 0.18694362017804153,
"acc_norm_stderr": 0.021268948348414647
},
"community|acva:entertainment|0": {
"acc_norm": 0.23389830508474577,
"acc_norm_stderr": 0.024687839412166384
},
"community|alghafa:mcq_exams_test_ar|0": {
"acc_norm": 0.30700179533213645,
"acc_norm_stderr": 0.019561363189897877
},
"community|alghafa:meta_ar_dialects|0": {
"acc_norm": 0.3230769230769231,
"acc_norm_stderr": 0.006367473671626477
},
"community|alghafa:meta_ar_msa|0": {
"acc_norm": 0.37094972067039106,
"acc_norm_stderr": 0.016155910721341767
},
"community|alghafa:multiple_choice_facts_truefalse_balanced_task|0": {
"acc_norm": 0.52,
"acc_norm_stderr": 0.05807730170189531
},
"community|alghafa:multiple_choice_grounded_statement_soqal_task|0": {
"acc_norm": 0.5466666666666666,
"acc_norm_stderr": 0.040782795278808064
},
"community|alghafa:multiple_choice_grounded_statement_xglue_mlqa_task|0": {
"acc_norm": 0.36666666666666664,
"acc_norm_stderr": 0.039478328284971595
},
"community|alghafa:multiple_choice_rating_sentiment_no_neutral_task|0": {
"acc_norm": 0.8172607879924954,
"acc_norm_stderr": 0.0043222922978817715
},
"community|alghafa:multiple_choice_rating_sentiment_task|0": {
"acc_norm": 0.5562969140950792,
"acc_norm_stderr": 0.006417135072062418
},
"community|alghafa:multiple_choice_sentiment_task|0": {
"acc_norm": 0.3715116279069767,
"acc_norm_stderr": 0.011654587070710052
},
"community|arabic_exams|0": {
"acc_norm": 0.36685288640595903,
"acc_norm_stderr": 0.020816892095910654
},
"community|arabic_mmlu:abstract_algebra|0": {
"acc_norm": 0.28,
"acc_norm_stderr": 0.045126085985421276
},
"community|arabic_mmlu:anatomy|0": {
"acc_norm": 0.4074074074074074,
"acc_norm_stderr": 0.04244633238353228
},
"community|arabic_mmlu:astronomy|0": {
"acc_norm": 0.4934210526315789,
"acc_norm_stderr": 0.040685900502249704
},
"community|arabic_mmlu:business_ethics|0": {
"acc_norm": 0.42,
"acc_norm_stderr": 0.049604496374885836
},
"community|arabic_mmlu:clinical_knowledge|0": {
"acc_norm": 0.41132075471698115,
"acc_norm_stderr": 0.030285009259009805
},
"community|arabic_mmlu:college_biology|0": {
"acc_norm": 0.3472222222222222,
"acc_norm_stderr": 0.039812405437178615
},
"community|arabic_mmlu:college_chemistry|0": {
"acc_norm": 0.27,
"acc_norm_stderr": 0.04461960433384741
},
"community|arabic_mmlu:college_computer_science|0": {
"acc_norm": 0.32,
"acc_norm_stderr": 0.046882617226215034
},
"community|arabic_mmlu:college_mathematics|0": {
"acc_norm": 0.25,
"acc_norm_stderr": 0.04351941398892446
},
"community|arabic_mmlu:college_medicine|0": {
"acc_norm": 0.3815028901734104,
"acc_norm_stderr": 0.03703851193099521
},
"community|arabic_mmlu:college_physics|0": {
"acc_norm": 0.2647058823529412,
"acc_norm_stderr": 0.043898699568087785
},
"community|arabic_mmlu:computer_security|0": {
"acc_norm": 0.49,
"acc_norm_stderr": 0.05024183937956911
},
"community|arabic_mmlu:conceptual_physics|0": {
"acc_norm": 0.3404255319148936,
"acc_norm_stderr": 0.03097669299853442
},
"community|arabic_mmlu:econometrics|0": {
"acc_norm": 0.30701754385964913,
"acc_norm_stderr": 0.04339138322579861
},
"community|arabic_mmlu:electrical_engineering|0": {
"acc_norm": 0.3793103448275862,
"acc_norm_stderr": 0.04043461861916747
},
"community|arabic_mmlu:elementary_mathematics|0": {
"acc_norm": 0.3941798941798942,
"acc_norm_stderr": 0.02516798233389414
},
"community|arabic_mmlu:formal_logic|0": {
"acc_norm": 0.23809523809523808,
"acc_norm_stderr": 0.03809523809523811
},
"community|arabic_mmlu:global_facts|0": {
"acc_norm": 0.35,
"acc_norm_stderr": 0.047937248544110196
},
"community|arabic_mmlu:high_school_biology|0": {
"acc_norm": 0.4096774193548387,
"acc_norm_stderr": 0.02797605491534737
},
"community|arabic_mmlu:high_school_chemistry|0": {
"acc_norm": 0.35960591133004927,
"acc_norm_stderr": 0.03376458246509567
},
"community|arabic_mmlu:high_school_computer_science|0": {
"acc_norm": 0.54,
"acc_norm_stderr": 0.05009082659620333
},
"community|arabic_mmlu:high_school_european_history|0": {
"acc_norm": 0.296969696969697,
"acc_norm_stderr": 0.035679697722680474
},
"community|arabic_mmlu:high_school_geography|0": {
"acc_norm": 0.46464646464646464,
"acc_norm_stderr": 0.035534363688280626
},
"community|arabic_mmlu:high_school_government_and_politics|0": {
"acc_norm": 0.38860103626943004,
"acc_norm_stderr": 0.03517739796373132
},
"community|arabic_mmlu:high_school_macroeconomics|0": {
"acc_norm": 0.45384615384615384,
"acc_norm_stderr": 0.025242770987126184
},
"community|arabic_mmlu:high_school_mathematics|0": {
"acc_norm": 0.32222222222222224,
"acc_norm_stderr": 0.028493465091028597
},
"community|arabic_mmlu:high_school_microeconomics|0": {
"acc_norm": 0.37815126050420167,
"acc_norm_stderr": 0.031499305777849054
},
"community|arabic_mmlu:high_school_physics|0": {
"acc_norm": 0.2980132450331126,
"acc_norm_stderr": 0.037345356767871984
},
"community|arabic_mmlu:high_school_psychology|0": {
"acc_norm": 0.3926605504587156,
"acc_norm_stderr": 0.020937505161201093
},
"community|arabic_mmlu:high_school_statistics|0": {
"acc_norm": 0.25925925925925924,
"acc_norm_stderr": 0.029886910547626957
},
"community|arabic_mmlu:high_school_us_history|0": {
"acc_norm": 0.29901960784313725,
"acc_norm_stderr": 0.03213325717373616
},
"community|arabic_mmlu:high_school_world_history|0": {
"acc_norm": 0.3206751054852321,
"acc_norm_stderr": 0.030381931949990407
},
"community|arabic_mmlu:human_aging|0": {
"acc_norm": 0.39461883408071746,
"acc_norm_stderr": 0.03280400504755291
},
"community|arabic_mmlu:human_sexuality|0": {
"acc_norm": 0.3893129770992366,
"acc_norm_stderr": 0.042764865428145914
},
"community|arabic_mmlu:international_law|0": {
"acc_norm": 0.6033057851239669,
"acc_norm_stderr": 0.044658697805310094
},
"community|arabic_mmlu:jurisprudence|0": {
"acc_norm": 0.3888888888888889,
"acc_norm_stderr": 0.0471282125742677
},
"community|arabic_mmlu:logical_fallacies|0": {
"acc_norm": 0.3803680981595092,
"acc_norm_stderr": 0.038142698932618374
},
"community|arabic_mmlu:machine_learning|0": {
"acc_norm": 0.32142857142857145,
"acc_norm_stderr": 0.04432804055291518
},
"community|arabic_mmlu:management|0": {
"acc_norm": 0.4174757281553398,
"acc_norm_stderr": 0.04882840548212238
},
"community|arabic_mmlu:marketing|0": {
"acc_norm": 0.5683760683760684,
"acc_norm_stderr": 0.0324483553531149
},
"community|arabic_mmlu:medical_genetics|0": {
"acc_norm": 0.35,
"acc_norm_stderr": 0.0479372485441102
},
"community|arabic_mmlu:miscellaneous|0": {
"acc_norm": 0.508301404853129,
"acc_norm_stderr": 0.017877498991072008
},
"community|arabic_mmlu:moral_disputes|0": {
"acc_norm": 0.41040462427745666,
"acc_norm_stderr": 0.026483392042098177
},
"community|arabic_mmlu:moral_scenarios|0": {
"acc_norm": 0.25139664804469275,
"acc_norm_stderr": 0.014508979453553969
},
"community|arabic_mmlu:nutrition|0": {
"acc_norm": 0.43137254901960786,
"acc_norm_stderr": 0.028358956313423552
},
"community|arabic_mmlu:philosophy|0": {
"acc_norm": 0.5112540192926045,
"acc_norm_stderr": 0.028390897396863533
},
"community|arabic_mmlu:prehistory|0": {
"acc_norm": 0.404320987654321,
"acc_norm_stderr": 0.027306625297327688
},
"community|arabic_mmlu:professional_accounting|0": {
"acc_norm": 0.32978723404255317,
"acc_norm_stderr": 0.028045946942042398
},
"community|arabic_mmlu:professional_law|0": {
"acc_norm": 0.29921773142112124,
"acc_norm_stderr": 0.011695374630696037
},
"community|arabic_mmlu:professional_medicine|0": {
"acc_norm": 0.1801470588235294,
"acc_norm_stderr": 0.02334516361654486
},
"community|arabic_mmlu:professional_psychology|0": {
"acc_norm": 0.35947712418300654,
"acc_norm_stderr": 0.01941253924203216
},
"community|arabic_mmlu:public_relations|0": {
"acc_norm": 0.39090909090909093,
"acc_norm_stderr": 0.04673752333670237
},
"community|arabic_mmlu:security_studies|0": {
"acc_norm": 0.4204081632653061,
"acc_norm_stderr": 0.03160106993449604
},
"community|arabic_mmlu:sociology|0": {
"acc_norm": 0.46766169154228854,
"acc_norm_stderr": 0.03528131472933607
},
"community|arabic_mmlu:us_foreign_policy|0": {
"acc_norm": 0.58,
"acc_norm_stderr": 0.049604496374885836
},
"community|arabic_mmlu:virology|0": {
"acc_norm": 0.28313253012048195,
"acc_norm_stderr": 0.03507295431370519
},
"community|arabic_mmlu:world_religions|0": {
"acc_norm": 0.4269005847953216,
"acc_norm_stderr": 0.03793620616529916
},
"community|arc_challenge_okapi_ar|0": {
"acc_norm": 0.4025862068965517,
"acc_norm_stderr": 0.014405400941453846
},
"community|arc_easy_ar|0": {
"acc_norm": 0.3934010152284264,
"acc_norm_stderr": 0.010049320954602599
},
"community|boolq_ar|0": {
"acc_norm": 0.6211656441717791,
"acc_norm_stderr": 0.008497402932896662
},
"community|copa_ext_ar|0": {
"acc_norm": 0.4777777777777778,
"acc_norm_stderr": 0.05294752255076824
},
"community|hellaswag_okapi_ar|0": {
"acc_norm": 0.27685094319049175,
"acc_norm_stderr": 0.004672531608215482
},
"community|openbook_qa_ext_ar|0": {
"acc_norm": 0.45252525252525255,
"acc_norm_stderr": 0.02239442836434113
},
"community|piqa_ar|0": {
"acc_norm": 0.6044735406437534,
"acc_norm_stderr": 0.011423881351056209
},
"community|race_ar|0": {
"acc_norm": 0.4023128423615338,
"acc_norm_stderr": 0.006985275841226595
},
"community|sciq_ar|0": {
"acc_norm": 0.4663316582914573,
"acc_norm_stderr": 0.015823041891329388
},
"community|toxigen_ar|0": {
"acc_norm": 0.4320855614973262,
"acc_norm_stderr": 0.01620887578524445
},
"lighteval|xstory_cloze:ar|0": {
"acc": 0.5962938451356717,
"acc_stderr": 0.01262624973524659
},
"community|acva:_average|0": {
"acc_norm": 0.3953409511782693,
"acc_norm_stderr": 0.04579372208486963
},
"community|alghafa:_average|0": {
"acc_norm": 0.464381233600815,
"acc_norm_stderr": 0.022535243032132817
},
"community|arabic_mmlu:_average|0": {
"acc_norm": 0.37888461559931796,
"acc_norm_stderr": 0.035701876763064314
},
"all": {
"acc_norm": 0.3970517976955435,
"acc_norm_stderr": 0.03761544161498773,
"acc": 0.5962938451356717,
"acc_stderr": 0.01262624973524659
}
},
"versions": {
"community|acva:Algeria|0": 0,
"community|acva:Ancient_Egypt|0": 0,
"community|acva:Arab_Empire|0": 0,
"community|acva:Arabic_Architecture|0": 0,
"community|acva:Arabic_Art|0": 0,
"community|acva:Arabic_Astronomy|0": 0,
"community|acva:Arabic_Calligraphy|0": 0,
"community|acva:Arabic_Ceremony|0": 0,
"community|acva:Arabic_Clothing|0": 0,
"community|acva:Arabic_Culture|0": 0,
"community|acva:Arabic_Food|0": 0,
"community|acva:Arabic_Funeral|0": 0,
"community|acva:Arabic_Geography|0": 0,
"community|acva:Arabic_History|0": 0,
"community|acva:Arabic_Language_Origin|0": 0,
"community|acva:Arabic_Literature|0": 0,
"community|acva:Arabic_Math|0": 0,
"community|acva:Arabic_Medicine|0": 0,
"community|acva:Arabic_Music|0": 0,
"community|acva:Arabic_Ornament|0": 0,
"community|acva:Arabic_Philosophy|0": 0,
"community|acva:Arabic_Physics_and_Chemistry|0": 0,
"community|acva:Arabic_Wedding|0": 0,
"community|acva:Bahrain|0": 0,
"community|acva:Comoros|0": 0,
"community|acva:Egypt_modern|0": 0,
"community|acva:InfluenceFromAncientEgypt|0": 0,
"community|acva:InfluenceFromByzantium|0": 0,
"community|acva:InfluenceFromChina|0": 0,
"community|acva:InfluenceFromGreece|0": 0,
"community|acva:InfluenceFromIslam|0": 0,
"community|acva:InfluenceFromPersia|0": 0,
"community|acva:InfluenceFromRome|0": 0,
"community|acva:Iraq|0": 0,
"community|acva:Islam_Education|0": 0,
"community|acva:Islam_branches_and_schools|0": 0,
"community|acva:Islamic_law_system|0": 0,
"community|acva:Jordan|0": 0,
"community|acva:Kuwait|0": 0,
"community|acva:Lebanon|0": 0,
"community|acva:Libya|0": 0,
"community|acva:Mauritania|0": 0,
"community|acva:Mesopotamia_civilization|0": 0,
"community|acva:Morocco|0": 0,
"community|acva:Oman|0": 0,
"community|acva:Palestine|0": 0,
"community|acva:Qatar|0": 0,
"community|acva:Saudi_Arabia|0": 0,
"community|acva:Somalia|0": 0,
"community|acva:Sudan|0": 0,
"community|acva:Syria|0": 0,
"community|acva:Tunisia|0": 0,
"community|acva:United_Arab_Emirates|0": 0,
"community|acva:Yemen|0": 0,
"community|acva:communication|0": 0,
"community|acva:computer_and_phone|0": 0,
"community|acva:daily_life|0": 0,
"community|acva:entertainment|0": 0,
"community|alghafa:mcq_exams_test_ar|0": 0,
"community|alghafa:meta_ar_dialects|0": 0,
"community|alghafa:meta_ar_msa|0": 0,
"community|alghafa:multiple_choice_facts_truefalse_balanced_task|0": 0,
"community|alghafa:multiple_choice_grounded_statement_soqal_task|0": 0,
"community|alghafa:multiple_choice_grounded_statement_xglue_mlqa_task|0": 0,
"community|alghafa:multiple_choice_rating_sentiment_no_neutral_task|0": 0,
"community|alghafa:multiple_choice_rating_sentiment_task|0": 0,
"community|alghafa:multiple_choice_sentiment_task|0": 0,
"community|arabic_exams|0": 0,
"community|arabic_mmlu:abstract_algebra|0": 0,
"community|arabic_mmlu:anatomy|0": 0,
"community|arabic_mmlu:astronomy|0": 0,
"community|arabic_mmlu:business_ethics|0": 0,
"community|arabic_mmlu:clinical_knowledge|0": 0,
"community|arabic_mmlu:college_biology|0": 0,
"community|arabic_mmlu:college_chemistry|0": 0,
"community|arabic_mmlu:college_computer_science|0": 0,
"community|arabic_mmlu:college_mathematics|0": 0,
"community|arabic_mmlu:college_medicine|0": 0,
"community|arabic_mmlu:college_physics|0": 0,
"community|arabic_mmlu:computer_security|0": 0,
"community|arabic_mmlu:conceptual_physics|0": 0,
"community|arabic_mmlu:econometrics|0": 0,
"community|arabic_mmlu:electrical_engineering|0": 0,
"community|arabic_mmlu:elementary_mathematics|0": 0,
"community|arabic_mmlu:formal_logic|0": 0,
"community|arabic_mmlu:global_facts|0": 0,
"community|arabic_mmlu:high_school_biology|0": 0,
"community|arabic_mmlu:high_school_chemistry|0": 0,
"community|arabic_mmlu:high_school_computer_science|0": 0,
"community|arabic_mmlu:high_school_european_history|0": 0,
"community|arabic_mmlu:high_school_geography|0": 0,
"community|arabic_mmlu:high_school_government_and_politics|0": 0,
"community|arabic_mmlu:high_school_macroeconomics|0": 0,
"community|arabic_mmlu:high_school_mathematics|0": 0,
"community|arabic_mmlu:high_school_microeconomics|0": 0,
"community|arabic_mmlu:high_school_physics|0": 0,
"community|arabic_mmlu:high_school_psychology|0": 0,
"community|arabic_mmlu:high_school_statistics|0": 0,
"community|arabic_mmlu:high_school_us_history|0": 0,
"community|arabic_mmlu:high_school_world_history|0": 0,
"community|arabic_mmlu:human_aging|0": 0,
"community|arabic_mmlu:human_sexuality|0": 0,
"community|arabic_mmlu:international_law|0": 0,
"community|arabic_mmlu:jurisprudence|0": 0,
"community|arabic_mmlu:logical_fallacies|0": 0,
"community|arabic_mmlu:machine_learning|0": 0,
"community|arabic_mmlu:management|0": 0,
"community|arabic_mmlu:marketing|0": 0,
"community|arabic_mmlu:medical_genetics|0": 0,
"community|arabic_mmlu:miscellaneous|0": 0,
"community|arabic_mmlu:moral_disputes|0": 0,
"community|arabic_mmlu:moral_scenarios|0": 0,
"community|arabic_mmlu:nutrition|0": 0,
"community|arabic_mmlu:philosophy|0": 0,
"community|arabic_mmlu:prehistory|0": 0,
"community|arabic_mmlu:professional_accounting|0": 0,
"community|arabic_mmlu:professional_law|0": 0,
"community|arabic_mmlu:professional_medicine|0": 0,
"community|arabic_mmlu:professional_psychology|0": 0,
"community|arabic_mmlu:public_relations|0": 0,
"community|arabic_mmlu:security_studies|0": 0,
"community|arabic_mmlu:sociology|0": 0,
"community|arabic_mmlu:us_foreign_policy|0": 0,
"community|arabic_mmlu:virology|0": 0,
"community|arabic_mmlu:world_religions|0": 0,
"community|arc_challenge_okapi_ar|0": 0,
"community|arc_easy_ar|0": 0,
"community|boolq_ar|0": 0,
"community|copa_ext_ar|0": 0,
"community|hellaswag_okapi_ar|0": 0,
"community|openbook_qa_ext_ar|0": 0,
"community|piqa_ar|0": 0,
"community|race_ar|0": 0,
"community|sciq_ar|0": 0,
"community|toxigen_ar|0": 0,
"lighteval|xstory_cloze:ar|0": 0
},
"config_tasks": {
"community|acva:Algeria": {
"name": "acva:Algeria",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Algeria",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 195,
"effective_num_docs": 195,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Ancient_Egypt": {
"name": "acva:Ancient_Egypt",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Ancient_Egypt",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 315,
"effective_num_docs": 315,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Arab_Empire": {
"name": "acva:Arab_Empire",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Arab_Empire",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 265,
"effective_num_docs": 265,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Arabic_Architecture": {
"name": "acva:Arabic_Architecture",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Arabic_Architecture",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 195,
"effective_num_docs": 195,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Arabic_Art": {
"name": "acva:Arabic_Art",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Arabic_Art",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 195,
"effective_num_docs": 195,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Arabic_Astronomy": {
"name": "acva:Arabic_Astronomy",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Arabic_Astronomy",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 195,
"effective_num_docs": 195,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Arabic_Calligraphy": {
"name": "acva:Arabic_Calligraphy",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Arabic_Calligraphy",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 255,
"effective_num_docs": 255,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Arabic_Ceremony": {
"name": "acva:Arabic_Ceremony",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Arabic_Ceremony",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 185,
"effective_num_docs": 185,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Arabic_Clothing": {
"name": "acva:Arabic_Clothing",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Arabic_Clothing",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 195,
"effective_num_docs": 195,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Arabic_Culture": {
"name": "acva:Arabic_Culture",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Arabic_Culture",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 195,
"effective_num_docs": 195,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Arabic_Food": {
"name": "acva:Arabic_Food",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Arabic_Food",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 195,
"effective_num_docs": 195,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Arabic_Funeral": {
"name": "acva:Arabic_Funeral",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Arabic_Funeral",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 95,
"effective_num_docs": 95,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Arabic_Geography": {
"name": "acva:Arabic_Geography",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Arabic_Geography",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 145,
"effective_num_docs": 145,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Arabic_History": {
"name": "acva:Arabic_History",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Arabic_History",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 195,
"effective_num_docs": 195,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Arabic_Language_Origin": {
"name": "acva:Arabic_Language_Origin",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Arabic_Language_Origin",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 95,
"effective_num_docs": 95,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Arabic_Literature": {
"name": "acva:Arabic_Literature",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Arabic_Literature",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 145,
"effective_num_docs": 145,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Arabic_Math": {
"name": "acva:Arabic_Math",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Arabic_Math",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 195,
"effective_num_docs": 195,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Arabic_Medicine": {
"name": "acva:Arabic_Medicine",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Arabic_Medicine",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 145,
"effective_num_docs": 145,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Arabic_Music": {
"name": "acva:Arabic_Music",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Arabic_Music",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 139,
"effective_num_docs": 139,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Arabic_Ornament": {
"name": "acva:Arabic_Ornament",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Arabic_Ornament",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 195,
"effective_num_docs": 195,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Arabic_Philosophy": {
"name": "acva:Arabic_Philosophy",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Arabic_Philosophy",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 145,
"effective_num_docs": 145,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Arabic_Physics_and_Chemistry": {
"name": "acva:Arabic_Physics_and_Chemistry",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Arabic_Physics_and_Chemistry",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 195,
"effective_num_docs": 195,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Arabic_Wedding": {
"name": "acva:Arabic_Wedding",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Arabic_Wedding",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 195,
"effective_num_docs": 195,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Bahrain": {
"name": "acva:Bahrain",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Bahrain",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 45,
"effective_num_docs": 45,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Comoros": {
"name": "acva:Comoros",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Comoros",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 45,
"effective_num_docs": 45,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Egypt_modern": {
"name": "acva:Egypt_modern",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Egypt_modern",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 95,
"effective_num_docs": 95,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:InfluenceFromAncientEgypt": {
"name": "acva:InfluenceFromAncientEgypt",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "InfluenceFromAncientEgypt",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 195,
"effective_num_docs": 195,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:InfluenceFromByzantium": {
"name": "acva:InfluenceFromByzantium",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "InfluenceFromByzantium",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 145,
"effective_num_docs": 145,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:InfluenceFromChina": {
"name": "acva:InfluenceFromChina",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "InfluenceFromChina",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 195,
"effective_num_docs": 195,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:InfluenceFromGreece": {
"name": "acva:InfluenceFromGreece",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "InfluenceFromGreece",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 195,
"effective_num_docs": 195,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:InfluenceFromIslam": {
"name": "acva:InfluenceFromIslam",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "InfluenceFromIslam",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 145,
"effective_num_docs": 145,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:InfluenceFromPersia": {
"name": "acva:InfluenceFromPersia",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "InfluenceFromPersia",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 175,
"effective_num_docs": 175,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:InfluenceFromRome": {
"name": "acva:InfluenceFromRome",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "InfluenceFromRome",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 195,
"effective_num_docs": 195,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Iraq": {
"name": "acva:Iraq",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Iraq",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 85,
"effective_num_docs": 85,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Islam_Education": {
"name": "acva:Islam_Education",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Islam_Education",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 195,
"effective_num_docs": 195,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Islam_branches_and_schools": {
"name": "acva:Islam_branches_and_schools",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Islam_branches_and_schools",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 175,
"effective_num_docs": 175,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Islamic_law_system": {
"name": "acva:Islamic_law_system",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Islamic_law_system",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 195,
"effective_num_docs": 195,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Jordan": {
"name": "acva:Jordan",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Jordan",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 45,
"effective_num_docs": 45,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Kuwait": {
"name": "acva:Kuwait",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Kuwait",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 45,
"effective_num_docs": 45,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Lebanon": {
"name": "acva:Lebanon",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Lebanon",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 45,
"effective_num_docs": 45,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Libya": {
"name": "acva:Libya",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Libya",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 45,
"effective_num_docs": 45,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Mauritania": {
"name": "acva:Mauritania",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Mauritania",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 45,
"effective_num_docs": 45,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Mesopotamia_civilization": {
"name": "acva:Mesopotamia_civilization",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Mesopotamia_civilization",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 155,
"effective_num_docs": 155,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Morocco": {
"name": "acva:Morocco",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Morocco",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 45,
"effective_num_docs": 45,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Oman": {
"name": "acva:Oman",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Oman",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 45,
"effective_num_docs": 45,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Palestine": {
"name": "acva:Palestine",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Palestine",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 85,
"effective_num_docs": 85,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Qatar": {
"name": "acva:Qatar",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Qatar",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 45,
"effective_num_docs": 45,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Saudi_Arabia": {
"name": "acva:Saudi_Arabia",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Saudi_Arabia",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 195,
"effective_num_docs": 195,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Somalia": {
"name": "acva:Somalia",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Somalia",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 45,
"effective_num_docs": 45,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Sudan": {
"name": "acva:Sudan",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Sudan",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 45,
"effective_num_docs": 45,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Syria": {
"name": "acva:Syria",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Syria",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 45,
"effective_num_docs": 45,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Tunisia": {
"name": "acva:Tunisia",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Tunisia",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 45,
"effective_num_docs": 45,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:United_Arab_Emirates": {
"name": "acva:United_Arab_Emirates",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "United_Arab_Emirates",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 85,
"effective_num_docs": 85,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:Yemen": {
"name": "acva:Yemen",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "Yemen",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 10,
"effective_num_docs": 10,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:communication": {
"name": "acva:communication",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "communication",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 364,
"effective_num_docs": 364,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:computer_and_phone": {
"name": "acva:computer_and_phone",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "computer_and_phone",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 295,
"effective_num_docs": 295,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:daily_life": {
"name": "acva:daily_life",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "daily_life",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 337,
"effective_num_docs": 337,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|acva:entertainment": {
"name": "acva:entertainment",
"prompt_function": "acva",
"hf_repo": "OALL/ACVA",
"hf_subset": "entertainment",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 295,
"effective_num_docs": 295,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|alghafa:mcq_exams_test_ar": {
"name": "alghafa:mcq_exams_test_ar",
"prompt_function": "alghafa_prompt",
"hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native",
"hf_subset": "mcq_exams_test_ar",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 557,
"effective_num_docs": 557,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|alghafa:meta_ar_dialects": {
"name": "alghafa:meta_ar_dialects",
"prompt_function": "alghafa_prompt",
"hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native",
"hf_subset": "meta_ar_dialects",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 5395,
"effective_num_docs": 5395,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|alghafa:meta_ar_msa": {
"name": "alghafa:meta_ar_msa",
"prompt_function": "alghafa_prompt",
"hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native",
"hf_subset": "meta_ar_msa",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 895,
"effective_num_docs": 895,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|alghafa:multiple_choice_facts_truefalse_balanced_task": {
"name": "alghafa:multiple_choice_facts_truefalse_balanced_task",
"prompt_function": "alghafa_prompt",
"hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native",
"hf_subset": "multiple_choice_facts_truefalse_balanced_task",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 75,
"effective_num_docs": 75,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|alghafa:multiple_choice_grounded_statement_soqal_task": {
"name": "alghafa:multiple_choice_grounded_statement_soqal_task",
"prompt_function": "alghafa_prompt",
"hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native",
"hf_subset": "multiple_choice_grounded_statement_soqal_task",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 150,
"effective_num_docs": 150,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|alghafa:multiple_choice_grounded_statement_xglue_mlqa_task": {
"name": "alghafa:multiple_choice_grounded_statement_xglue_mlqa_task",
"prompt_function": "alghafa_prompt",
"hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native",
"hf_subset": "multiple_choice_grounded_statement_xglue_mlqa_task",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 150,
"effective_num_docs": 150,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|alghafa:multiple_choice_rating_sentiment_no_neutral_task": {
"name": "alghafa:multiple_choice_rating_sentiment_no_neutral_task",
"prompt_function": "alghafa_prompt",
"hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native",
"hf_subset": "multiple_choice_rating_sentiment_no_neutral_task",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 7995,
"effective_num_docs": 7995,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|alghafa:multiple_choice_rating_sentiment_task": {
"name": "alghafa:multiple_choice_rating_sentiment_task",
"prompt_function": "alghafa_prompt",
"hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native",
"hf_subset": "multiple_choice_rating_sentiment_task",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 5995,
"effective_num_docs": 5995,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|alghafa:multiple_choice_sentiment_task": {
"name": "alghafa:multiple_choice_sentiment_task",
"prompt_function": "alghafa_prompt",
"hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native",
"hf_subset": "multiple_choice_sentiment_task",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 1720,
"effective_num_docs": 1720,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_exams": {
"name": "arabic_exams",
"prompt_function": "arabic_exams",
"hf_repo": "OALL/Arabic_EXAMS",
"hf_subset": "default",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": null,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 537,
"effective_num_docs": 537,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:abstract_algebra": {
"name": "arabic_mmlu:abstract_algebra",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "abstract_algebra",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 100,
"effective_num_docs": 100,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:anatomy": {
"name": "arabic_mmlu:anatomy",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "anatomy",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 135,
"effective_num_docs": 135,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:astronomy": {
"name": "arabic_mmlu:astronomy",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "astronomy",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 152,
"effective_num_docs": 152,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:business_ethics": {
"name": "arabic_mmlu:business_ethics",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "business_ethics",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 100,
"effective_num_docs": 100,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:clinical_knowledge": {
"name": "arabic_mmlu:clinical_knowledge",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "clinical_knowledge",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 265,
"effective_num_docs": 265,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:college_biology": {
"name": "arabic_mmlu:college_biology",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "college_biology",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 144,
"effective_num_docs": 144,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:college_chemistry": {
"name": "arabic_mmlu:college_chemistry",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "college_chemistry",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 100,
"effective_num_docs": 100,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:college_computer_science": {
"name": "arabic_mmlu:college_computer_science",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "college_computer_science",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 100,
"effective_num_docs": 100,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:college_mathematics": {
"name": "arabic_mmlu:college_mathematics",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "college_mathematics",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 100,
"effective_num_docs": 100,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:college_medicine": {
"name": "arabic_mmlu:college_medicine",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "college_medicine",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 173,
"effective_num_docs": 173,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:college_physics": {
"name": "arabic_mmlu:college_physics",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "college_physics",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 102,
"effective_num_docs": 102,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:computer_security": {
"name": "arabic_mmlu:computer_security",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "computer_security",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 100,
"effective_num_docs": 100,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:conceptual_physics": {
"name": "arabic_mmlu:conceptual_physics",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "conceptual_physics",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 235,
"effective_num_docs": 235,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:econometrics": {
"name": "arabic_mmlu:econometrics",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "econometrics",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 114,
"effective_num_docs": 114,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:electrical_engineering": {
"name": "arabic_mmlu:electrical_engineering",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "electrical_engineering",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 145,
"effective_num_docs": 145,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:elementary_mathematics": {
"name": "arabic_mmlu:elementary_mathematics",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "elementary_mathematics",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 378,
"effective_num_docs": 378,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:formal_logic": {
"name": "arabic_mmlu:formal_logic",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "formal_logic",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 126,
"effective_num_docs": 126,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:global_facts": {
"name": "arabic_mmlu:global_facts",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "global_facts",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 100,
"effective_num_docs": 100,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:high_school_biology": {
"name": "arabic_mmlu:high_school_biology",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "high_school_biology",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 310,
"effective_num_docs": 310,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:high_school_chemistry": {
"name": "arabic_mmlu:high_school_chemistry",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "high_school_chemistry",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 203,
"effective_num_docs": 203,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:high_school_computer_science": {
"name": "arabic_mmlu:high_school_computer_science",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "high_school_computer_science",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 100,
"effective_num_docs": 100,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:high_school_european_history": {
"name": "arabic_mmlu:high_school_european_history",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "high_school_european_history",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 165,
"effective_num_docs": 165,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:high_school_geography": {
"name": "arabic_mmlu:high_school_geography",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "high_school_geography",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 198,
"effective_num_docs": 198,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:high_school_government_and_politics": {
"name": "arabic_mmlu:high_school_government_and_politics",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "high_school_government_and_politics",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 193,
"effective_num_docs": 193,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:high_school_macroeconomics": {
"name": "arabic_mmlu:high_school_macroeconomics",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "high_school_macroeconomics",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 390,
"effective_num_docs": 390,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:high_school_mathematics": {
"name": "arabic_mmlu:high_school_mathematics",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "high_school_mathematics",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 270,
"effective_num_docs": 270,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:high_school_microeconomics": {
"name": "arabic_mmlu:high_school_microeconomics",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "high_school_microeconomics",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 238,
"effective_num_docs": 238,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:high_school_physics": {
"name": "arabic_mmlu:high_school_physics",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "high_school_physics",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 151,
"effective_num_docs": 151,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:high_school_psychology": {
"name": "arabic_mmlu:high_school_psychology",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "high_school_psychology",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 545,
"effective_num_docs": 545,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:high_school_statistics": {
"name": "arabic_mmlu:high_school_statistics",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "high_school_statistics",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 216,
"effective_num_docs": 216,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:high_school_us_history": {
"name": "arabic_mmlu:high_school_us_history",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "high_school_us_history",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 204,
"effective_num_docs": 204,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:high_school_world_history": {
"name": "arabic_mmlu:high_school_world_history",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "high_school_world_history",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 237,
"effective_num_docs": 237,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:human_aging": {
"name": "arabic_mmlu:human_aging",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "human_aging",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 223,
"effective_num_docs": 223,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:human_sexuality": {
"name": "arabic_mmlu:human_sexuality",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "human_sexuality",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 131,
"effective_num_docs": 131,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:international_law": {
"name": "arabic_mmlu:international_law",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "international_law",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 121,
"effective_num_docs": 121,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:jurisprudence": {
"name": "arabic_mmlu:jurisprudence",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "jurisprudence",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 108,
"effective_num_docs": 108,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:logical_fallacies": {
"name": "arabic_mmlu:logical_fallacies",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "logical_fallacies",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 163,
"effective_num_docs": 163,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:machine_learning": {
"name": "arabic_mmlu:machine_learning",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "machine_learning",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 112,
"effective_num_docs": 112,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:management": {
"name": "arabic_mmlu:management",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "management",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 103,
"effective_num_docs": 103,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:marketing": {
"name": "arabic_mmlu:marketing",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "marketing",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 234,
"effective_num_docs": 234,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:medical_genetics": {
"name": "arabic_mmlu:medical_genetics",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "medical_genetics",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 100,
"effective_num_docs": 100,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:miscellaneous": {
"name": "arabic_mmlu:miscellaneous",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "miscellaneous",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 783,
"effective_num_docs": 783,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:moral_disputes": {
"name": "arabic_mmlu:moral_disputes",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "moral_disputes",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 346,
"effective_num_docs": 346,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:moral_scenarios": {
"name": "arabic_mmlu:moral_scenarios",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "moral_scenarios",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 895,
"effective_num_docs": 895,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:nutrition": {
"name": "arabic_mmlu:nutrition",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "nutrition",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 306,
"effective_num_docs": 306,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:philosophy": {
"name": "arabic_mmlu:philosophy",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "philosophy",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 311,
"effective_num_docs": 311,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:prehistory": {
"name": "arabic_mmlu:prehistory",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "prehistory",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 324,
"effective_num_docs": 324,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:professional_accounting": {
"name": "arabic_mmlu:professional_accounting",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "professional_accounting",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 282,
"effective_num_docs": 282,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:professional_law": {
"name": "arabic_mmlu:professional_law",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "professional_law",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 1534,
"effective_num_docs": 1534,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:professional_medicine": {
"name": "arabic_mmlu:professional_medicine",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "professional_medicine",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 272,
"effective_num_docs": 272,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:professional_psychology": {
"name": "arabic_mmlu:professional_psychology",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "professional_psychology",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 612,
"effective_num_docs": 612,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:public_relations": {
"name": "arabic_mmlu:public_relations",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "public_relations",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 110,
"effective_num_docs": 110,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:security_studies": {
"name": "arabic_mmlu:security_studies",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "security_studies",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 245,
"effective_num_docs": 245,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:sociology": {
"name": "arabic_mmlu:sociology",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "sociology",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 201,
"effective_num_docs": 201,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:us_foreign_policy": {
"name": "arabic_mmlu:us_foreign_policy",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "us_foreign_policy",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 100,
"effective_num_docs": 100,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:virology": {
"name": "arabic_mmlu:virology",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "virology",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 166,
"effective_num_docs": 166,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arabic_mmlu:world_religions": {
"name": "arabic_mmlu:world_religions",
"prompt_function": "mmlu_arabic",
"hf_repo": "OALL/Arabic_MMLU",
"hf_subset": "world_religions",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"dev"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": "sequential",
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 171,
"effective_num_docs": 171,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arc_challenge_okapi_ar": {
"name": "arc_challenge_okapi_ar",
"prompt_function": "alghafa_prompt",
"hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Translated",
"hf_subset": "arc_challenge_okapi_ar",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": null,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 1160,
"effective_num_docs": 1160,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|arc_easy_ar": {
"name": "arc_easy_ar",
"prompt_function": "alghafa_prompt",
"hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Translated",
"hf_subset": "arc_easy_ar",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": null,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 2364,
"effective_num_docs": 2364,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|boolq_ar": {
"name": "boolq_ar",
"prompt_function": "boolq_prompt_arabic",
"hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Translated",
"hf_subset": "boolq_ar",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": null,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 3260,
"effective_num_docs": 3260,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|copa_ext_ar": {
"name": "copa_ext_ar",
"prompt_function": "copa_prompt_arabic",
"hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Translated",
"hf_subset": "copa_ext_ar",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": null,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 90,
"effective_num_docs": 90,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|hellaswag_okapi_ar": {
"name": "hellaswag_okapi_ar",
"prompt_function": "hellaswag_prompt_arabic",
"hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Translated",
"hf_subset": "hellaswag_okapi_ar",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": null,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 9171,
"effective_num_docs": 9171,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|openbook_qa_ext_ar": {
"name": "openbook_qa_ext_ar",
"prompt_function": "alghafa_prompt",
"hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Translated",
"hf_subset": "openbook_qa_ext_ar",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": null,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 495,
"effective_num_docs": 495,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|piqa_ar": {
"name": "piqa_ar",
"prompt_function": "alghafa_prompt",
"hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Translated",
"hf_subset": "piqa_ar",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": null,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 1833,
"effective_num_docs": 1833,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|race_ar": {
"name": "race_ar",
"prompt_function": "alghafa_prompt",
"hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Translated",
"hf_subset": "race_ar",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": null,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 4929,
"effective_num_docs": 4929,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|sciq_ar": {
"name": "sciq_ar",
"prompt_function": "sciq_prompt_arabic",
"hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Translated",
"hf_subset": "sciq_ar",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": null,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 995,
"effective_num_docs": 995,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"community|toxigen_ar": {
"name": "toxigen_ar",
"prompt_function": "toxigen_prompt_arabic",
"hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Translated",
"hf_subset": "toxigen_ar",
"metric": [
"loglikelihood_acc_norm"
],
"hf_avail_splits": [
"test",
"validation"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": "sequential",
"generation_size": null,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"community"
],
"original_num_docs": 935,
"effective_num_docs": 935,
"trust_dataset": null,
"must_remove_duplicate_docs": null,
"version": 0
},
"lighteval|xstory_cloze:ar": {
"name": "xstory_cloze:ar",
"prompt_function": "storycloze",
"hf_repo": "juletxara/xstory_cloze",
"hf_subset": "ar",
"metric": [
"loglikelihood_acc"
],
"hf_avail_splits": [
"training",
"eval"
],
"evaluation_splits": [
"eval"
],
"few_shots_split": null,
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": [
"\n"
],
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"lighteval"
],
"original_num_docs": 1511,
"effective_num_docs": 1511,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
}
},
"summary_tasks": {
"community|acva:Algeria|0": {
"hashes": {
"hash_examples": "da5a3003cd46f6f9",
"hash_full_prompts": "da5a3003cd46f6f9",
"hash_input_tokens": "5e5502efd841c7d2",
"hash_cont_tokens": "ebddcaf492db5bb8"
},
"truncated": 0,
"non_truncated": 195,
"padded": 390,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Ancient_Egypt|0": {
"hashes": {
"hash_examples": "52d6f767fede195b",
"hash_full_prompts": "52d6f767fede195b",
"hash_input_tokens": "b8ac457ad647341e",
"hash_cont_tokens": "02a204d955f29ed4"
},
"truncated": 0,
"non_truncated": 315,
"padded": 630,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Arab_Empire|0": {
"hashes": {
"hash_examples": "8dacff6a79804a75",
"hash_full_prompts": "8dacff6a79804a75",
"hash_input_tokens": "93b427d5ed1d49fb",
"hash_cont_tokens": "0be121aeaa740bc8"
},
"truncated": 0,
"non_truncated": 265,
"padded": 530,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Arabic_Architecture|0": {
"hashes": {
"hash_examples": "df286cd862d9f6bb",
"hash_full_prompts": "df286cd862d9f6bb",
"hash_input_tokens": "7c6bca181e962a03",
"hash_cont_tokens": "ebddcaf492db5bb8"
},
"truncated": 0,
"non_truncated": 195,
"padded": 390,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Arabic_Art|0": {
"hashes": {
"hash_examples": "112883d764118a49",
"hash_full_prompts": "112883d764118a49",
"hash_input_tokens": "4aa2e2de78e5f719",
"hash_cont_tokens": "ebddcaf492db5bb8"
},
"truncated": 0,
"non_truncated": 195,
"padded": 390,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Arabic_Astronomy|0": {
"hashes": {
"hash_examples": "20dcdf2454bf8671",
"hash_full_prompts": "20dcdf2454bf8671",
"hash_input_tokens": "01b513d145b07902",
"hash_cont_tokens": "ebddcaf492db5bb8"
},
"truncated": 0,
"non_truncated": 195,
"padded": 390,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Arabic_Calligraphy|0": {
"hashes": {
"hash_examples": "3a9f9d1ebe868a15",
"hash_full_prompts": "3a9f9d1ebe868a15",
"hash_input_tokens": "20c10f13ffe11c17",
"hash_cont_tokens": "3a362560f15a8d81"
},
"truncated": 0,
"non_truncated": 255,
"padded": 510,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Arabic_Ceremony|0": {
"hashes": {
"hash_examples": "c927630f8d2f44da",
"hash_full_prompts": "c927630f8d2f44da",
"hash_input_tokens": "b10f27e0d2601a1b",
"hash_cont_tokens": "219de3ed588d7bf7"
},
"truncated": 0,
"non_truncated": 185,
"padded": 370,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Arabic_Clothing|0": {
"hashes": {
"hash_examples": "6ad0740c2ac6ac92",
"hash_full_prompts": "6ad0740c2ac6ac92",
"hash_input_tokens": "578d4eceacdf7e47",
"hash_cont_tokens": "ebddcaf492db5bb8"
},
"truncated": 0,
"non_truncated": 195,
"padded": 390,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Arabic_Culture|0": {
"hashes": {
"hash_examples": "2177bd857ad872ae",
"hash_full_prompts": "2177bd857ad872ae",
"hash_input_tokens": "2959d51ecb3a16e1",
"hash_cont_tokens": "ebddcaf492db5bb8"
},
"truncated": 0,
"non_truncated": 195,
"padded": 390,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Arabic_Food|0": {
"hashes": {
"hash_examples": "a6ada65b71d7c9c5",
"hash_full_prompts": "a6ada65b71d7c9c5",
"hash_input_tokens": "1db6ea42e71abb5e",
"hash_cont_tokens": "ebddcaf492db5bb8"
},
"truncated": 0,
"non_truncated": 195,
"padded": 390,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Arabic_Funeral|0": {
"hashes": {
"hash_examples": "fcee39dc29eaae91",
"hash_full_prompts": "fcee39dc29eaae91",
"hash_input_tokens": "5b5c82b733f08c7f",
"hash_cont_tokens": "c36c7371f1293511"
},
"truncated": 0,
"non_truncated": 95,
"padded": 190,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Arabic_Geography|0": {
"hashes": {
"hash_examples": "d36eda7c89231c02",
"hash_full_prompts": "d36eda7c89231c02",
"hash_input_tokens": "42438125d51d9bac",
"hash_cont_tokens": "625e58e7a01dba13"
},
"truncated": 0,
"non_truncated": 145,
"padded": 290,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Arabic_History|0": {
"hashes": {
"hash_examples": "6354ac0d6db6a5fc",
"hash_full_prompts": "6354ac0d6db6a5fc",
"hash_input_tokens": "455f8fdea0d50b72",
"hash_cont_tokens": "ebddcaf492db5bb8"
},
"truncated": 0,
"non_truncated": 195,
"padded": 390,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Arabic_Language_Origin|0": {
"hashes": {
"hash_examples": "ddc967c8aca34402",
"hash_full_prompts": "ddc967c8aca34402",
"hash_input_tokens": "4f27fe063a4781d7",
"hash_cont_tokens": "c36c7371f1293511"
},
"truncated": 0,
"non_truncated": 95,
"padded": 190,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Arabic_Literature|0": {
"hashes": {
"hash_examples": "4305379fd46be5d8",
"hash_full_prompts": "4305379fd46be5d8",
"hash_input_tokens": "752a60c1510ab7e1",
"hash_cont_tokens": "625e58e7a01dba13"
},
"truncated": 0,
"non_truncated": 145,
"padded": 290,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Arabic_Math|0": {
"hashes": {
"hash_examples": "dec621144f4d28be",
"hash_full_prompts": "dec621144f4d28be",
"hash_input_tokens": "03066d8c133c1025",
"hash_cont_tokens": "ebddcaf492db5bb8"
},
"truncated": 0,
"non_truncated": 195,
"padded": 390,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Arabic_Medicine|0": {
"hashes": {
"hash_examples": "2b344cdae9495ff2",
"hash_full_prompts": "2b344cdae9495ff2",
"hash_input_tokens": "7882af2f09acdb21",
"hash_cont_tokens": "625e58e7a01dba13"
},
"truncated": 0,
"non_truncated": 145,
"padded": 290,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Arabic_Music|0": {
"hashes": {
"hash_examples": "0c54624d881944ce",
"hash_full_prompts": "0c54624d881944ce",
"hash_input_tokens": "2c1a4d248e1db3f5",
"hash_cont_tokens": "4ac287553cdf8021"
},
"truncated": 0,
"non_truncated": 139,
"padded": 278,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Arabic_Ornament|0": {
"hashes": {
"hash_examples": "251a4a84289d8bc1",
"hash_full_prompts": "251a4a84289d8bc1",
"hash_input_tokens": "7387a2e247f8f14f",
"hash_cont_tokens": "ebddcaf492db5bb8"
},
"truncated": 0,
"non_truncated": 195,
"padded": 390,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Arabic_Philosophy|0": {
"hashes": {
"hash_examples": "3f86fb9c94c13d22",
"hash_full_prompts": "3f86fb9c94c13d22",
"hash_input_tokens": "9eacb3e2b284a169",
"hash_cont_tokens": "625e58e7a01dba13"
},
"truncated": 0,
"non_truncated": 145,
"padded": 290,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Arabic_Physics_and_Chemistry|0": {
"hashes": {
"hash_examples": "8fec65af3695b62a",
"hash_full_prompts": "8fec65af3695b62a",
"hash_input_tokens": "f4d7d84454b40520",
"hash_cont_tokens": "ebddcaf492db5bb8"
},
"truncated": 0,
"non_truncated": 195,
"padded": 390,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Arabic_Wedding|0": {
"hashes": {
"hash_examples": "9cc3477184d7a4b8",
"hash_full_prompts": "9cc3477184d7a4b8",
"hash_input_tokens": "3967bd0c965df759",
"hash_cont_tokens": "ebddcaf492db5bb8"
},
"truncated": 0,
"non_truncated": 195,
"padded": 390,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Bahrain|0": {
"hashes": {
"hash_examples": "c92e803a0fa8b9e2",
"hash_full_prompts": "c92e803a0fa8b9e2",
"hash_input_tokens": "205d732bc1dc23a6",
"hash_cont_tokens": "9ad7f58ff8a11e98"
},
"truncated": 0,
"non_truncated": 45,
"padded": 90,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Comoros|0": {
"hashes": {
"hash_examples": "06e5d4bba8e54cae",
"hash_full_prompts": "06e5d4bba8e54cae",
"hash_input_tokens": "f4ebc061fbb409fe",
"hash_cont_tokens": "9ad7f58ff8a11e98"
},
"truncated": 0,
"non_truncated": 45,
"padded": 90,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Egypt_modern|0": {
"hashes": {
"hash_examples": "c6ec369164f93446",
"hash_full_prompts": "c6ec369164f93446",
"hash_input_tokens": "12f22beebab110fa",
"hash_cont_tokens": "c36c7371f1293511"
},
"truncated": 0,
"non_truncated": 95,
"padded": 190,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:InfluenceFromAncientEgypt|0": {
"hashes": {
"hash_examples": "b9d56d74818b9bd4",
"hash_full_prompts": "b9d56d74818b9bd4",
"hash_input_tokens": "2fc941d815bb9ad8",
"hash_cont_tokens": "ebddcaf492db5bb8"
},
"truncated": 0,
"non_truncated": 195,
"padded": 390,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:InfluenceFromByzantium|0": {
"hashes": {
"hash_examples": "5316c9624e7e59b8",
"hash_full_prompts": "5316c9624e7e59b8",
"hash_input_tokens": "1dbcfca70847b6b1",
"hash_cont_tokens": "625e58e7a01dba13"
},
"truncated": 0,
"non_truncated": 145,
"padded": 290,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:InfluenceFromChina|0": {
"hashes": {
"hash_examples": "87894bce95a56411",
"hash_full_prompts": "87894bce95a56411",
"hash_input_tokens": "6ac0c524bbf2b8ed",
"hash_cont_tokens": "ebddcaf492db5bb8"
},
"truncated": 0,
"non_truncated": 195,
"padded": 390,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:InfluenceFromGreece|0": {
"hashes": {
"hash_examples": "0baa78a27e469312",
"hash_full_prompts": "0baa78a27e469312",
"hash_input_tokens": "b906cef6e3cb8c6c",
"hash_cont_tokens": "ebddcaf492db5bb8"
},
"truncated": 0,
"non_truncated": 195,
"padded": 390,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:InfluenceFromIslam|0": {
"hashes": {
"hash_examples": "0c2532cde6541ff2",
"hash_full_prompts": "0c2532cde6541ff2",
"hash_input_tokens": "85da877775f205da",
"hash_cont_tokens": "625e58e7a01dba13"
},
"truncated": 0,
"non_truncated": 145,
"padded": 290,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:InfluenceFromPersia|0": {
"hashes": {
"hash_examples": "efcd8112dc53c6e5",
"hash_full_prompts": "efcd8112dc53c6e5",
"hash_input_tokens": "893353ab110326ab",
"hash_cont_tokens": "0060d8f35205c778"
},
"truncated": 0,
"non_truncated": 175,
"padded": 350,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:InfluenceFromRome|0": {
"hashes": {
"hash_examples": "9db61480e2e85fd3",
"hash_full_prompts": "9db61480e2e85fd3",
"hash_input_tokens": "9bedbde7584d6dc2",
"hash_cont_tokens": "ebddcaf492db5bb8"
},
"truncated": 0,
"non_truncated": 195,
"padded": 390,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Iraq|0": {
"hashes": {
"hash_examples": "96dac3dfa8d2f41f",
"hash_full_prompts": "96dac3dfa8d2f41f",
"hash_input_tokens": "04c5585fe16fd21c",
"hash_cont_tokens": "174ee430e070c2fa"
},
"truncated": 0,
"non_truncated": 85,
"padded": 170,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Islam_Education|0": {
"hashes": {
"hash_examples": "0d80355f6a4cb51b",
"hash_full_prompts": "0d80355f6a4cb51b",
"hash_input_tokens": "cb51bd08db7bcea1",
"hash_cont_tokens": "ebddcaf492db5bb8"
},
"truncated": 0,
"non_truncated": 195,
"padded": 390,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Islam_branches_and_schools|0": {
"hashes": {
"hash_examples": "5cedce1be2c3ad50",
"hash_full_prompts": "5cedce1be2c3ad50",
"hash_input_tokens": "cd751ff1e9dcbc87",
"hash_cont_tokens": "0060d8f35205c778"
},
"truncated": 0,
"non_truncated": 175,
"padded": 350,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Islamic_law_system|0": {
"hashes": {
"hash_examples": "c0e6db8bc84e105e",
"hash_full_prompts": "c0e6db8bc84e105e",
"hash_input_tokens": "c744f917fdc9b04a",
"hash_cont_tokens": "ebddcaf492db5bb8"
},
"truncated": 0,
"non_truncated": 195,
"padded": 390,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Jordan|0": {
"hashes": {
"hash_examples": "33deb5b4e5ddd6a1",
"hash_full_prompts": "33deb5b4e5ddd6a1",
"hash_input_tokens": "0ce1fceb935aa165",
"hash_cont_tokens": "9ad7f58ff8a11e98"
},
"truncated": 0,
"non_truncated": 45,
"padded": 90,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Kuwait|0": {
"hashes": {
"hash_examples": "eb41773346d7c46c",
"hash_full_prompts": "eb41773346d7c46c",
"hash_input_tokens": "9e9579cf2b2ebc64",
"hash_cont_tokens": "9ad7f58ff8a11e98"
},
"truncated": 0,
"non_truncated": 45,
"padded": 90,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Lebanon|0": {
"hashes": {
"hash_examples": "25932dbf4c13d34f",
"hash_full_prompts": "25932dbf4c13d34f",
"hash_input_tokens": "b79e84bbedee5490",
"hash_cont_tokens": "9ad7f58ff8a11e98"
},
"truncated": 0,
"non_truncated": 45,
"padded": 90,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Libya|0": {
"hashes": {
"hash_examples": "f2c4db63cd402926",
"hash_full_prompts": "f2c4db63cd402926",
"hash_input_tokens": "2cb43bcd49941332",
"hash_cont_tokens": "9ad7f58ff8a11e98"
},
"truncated": 0,
"non_truncated": 45,
"padded": 90,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Mauritania|0": {
"hashes": {
"hash_examples": "8723ab5fdf286b54",
"hash_full_prompts": "8723ab5fdf286b54",
"hash_input_tokens": "8a827b2325549900",
"hash_cont_tokens": "9ad7f58ff8a11e98"
},
"truncated": 0,
"non_truncated": 45,
"padded": 90,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Mesopotamia_civilization|0": {
"hashes": {
"hash_examples": "c33f5502a6130ca9",
"hash_full_prompts": "c33f5502a6130ca9",
"hash_input_tokens": "d3c7dfce230b99c3",
"hash_cont_tokens": "ac62599297c498fd"
},
"truncated": 0,
"non_truncated": 155,
"padded": 310,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Morocco|0": {
"hashes": {
"hash_examples": "588a5ed27904b1ae",
"hash_full_prompts": "588a5ed27904b1ae",
"hash_input_tokens": "3601bbc792720f39",
"hash_cont_tokens": "9ad7f58ff8a11e98"
},
"truncated": 0,
"non_truncated": 45,
"padded": 90,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Oman|0": {
"hashes": {
"hash_examples": "d447c52b94248b69",
"hash_full_prompts": "d447c52b94248b69",
"hash_input_tokens": "fa818744026bbd5a",
"hash_cont_tokens": "9ad7f58ff8a11e98"
},
"truncated": 0,
"non_truncated": 45,
"padded": 90,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Palestine|0": {
"hashes": {
"hash_examples": "19197e076ad14ff5",
"hash_full_prompts": "19197e076ad14ff5",
"hash_input_tokens": "9e403929d0fbd5bc",
"hash_cont_tokens": "174ee430e070c2fa"
},
"truncated": 0,
"non_truncated": 85,
"padded": 170,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Qatar|0": {
"hashes": {
"hash_examples": "cf0736fa185b28f6",
"hash_full_prompts": "cf0736fa185b28f6",
"hash_input_tokens": "38f74abfff7d01b1",
"hash_cont_tokens": "9ad7f58ff8a11e98"
},
"truncated": 0,
"non_truncated": 45,
"padded": 90,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Saudi_Arabia|0": {
"hashes": {
"hash_examples": "69beda6e1b85a08d",
"hash_full_prompts": "69beda6e1b85a08d",
"hash_input_tokens": "ec2bd2ea47723d3a",
"hash_cont_tokens": "ebddcaf492db5bb8"
},
"truncated": 0,
"non_truncated": 195,
"padded": 390,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Somalia|0": {
"hashes": {
"hash_examples": "b387940c65784fbf",
"hash_full_prompts": "b387940c65784fbf",
"hash_input_tokens": "b019481d4975f72f",
"hash_cont_tokens": "9ad7f58ff8a11e98"
},
"truncated": 0,
"non_truncated": 45,
"padded": 90,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Sudan|0": {
"hashes": {
"hash_examples": "e02c32b9d2dd0c3f",
"hash_full_prompts": "e02c32b9d2dd0c3f",
"hash_input_tokens": "dbaefc650494d268",
"hash_cont_tokens": "9ad7f58ff8a11e98"
},
"truncated": 0,
"non_truncated": 45,
"padded": 90,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Syria|0": {
"hashes": {
"hash_examples": "60a6f8fe73bda4bb",
"hash_full_prompts": "60a6f8fe73bda4bb",
"hash_input_tokens": "74fbcf3742e9aa6f",
"hash_cont_tokens": "9ad7f58ff8a11e98"
},
"truncated": 0,
"non_truncated": 45,
"padded": 90,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Tunisia|0": {
"hashes": {
"hash_examples": "34bb15d3830c5649",
"hash_full_prompts": "34bb15d3830c5649",
"hash_input_tokens": "bf3397eb31bcea7a",
"hash_cont_tokens": "9ad7f58ff8a11e98"
},
"truncated": 0,
"non_truncated": 45,
"padded": 90,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:United_Arab_Emirates|0": {
"hashes": {
"hash_examples": "98a0ba78172718ce",
"hash_full_prompts": "98a0ba78172718ce",
"hash_input_tokens": "e6c77b414bcfae6f",
"hash_cont_tokens": "174ee430e070c2fa"
},
"truncated": 0,
"non_truncated": 85,
"padded": 170,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:Yemen|0": {
"hashes": {
"hash_examples": "18e9bcccbb4ced7a",
"hash_full_prompts": "18e9bcccbb4ced7a",
"hash_input_tokens": "dd62fe18bf13fb56",
"hash_cont_tokens": "96702f2356f6107c"
},
"truncated": 0,
"non_truncated": 10,
"padded": 20,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:communication|0": {
"hashes": {
"hash_examples": "9ff28ab5eab5c97b",
"hash_full_prompts": "9ff28ab5eab5c97b",
"hash_input_tokens": "091627cf16e9588b",
"hash_cont_tokens": "b628a89bcecf356d"
},
"truncated": 0,
"non_truncated": 364,
"padded": 728,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:computer_and_phone|0": {
"hashes": {
"hash_examples": "37bac2f086aaf6c2",
"hash_full_prompts": "37bac2f086aaf6c2",
"hash_input_tokens": "82e83ba7597b0a5f",
"hash_cont_tokens": "d8ae57c62ca85f4b"
},
"truncated": 0,
"non_truncated": 295,
"padded": 590,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:daily_life|0": {
"hashes": {
"hash_examples": "bf07363c1c252e2f",
"hash_full_prompts": "bf07363c1c252e2f",
"hash_input_tokens": "ed258043197d08b9",
"hash_cont_tokens": "d598caf874354f48"
},
"truncated": 0,
"non_truncated": 337,
"padded": 674,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|acva:entertainment|0": {
"hashes": {
"hash_examples": "37077bc00f0ac56a",
"hash_full_prompts": "37077bc00f0ac56a",
"hash_input_tokens": "9497c570b31562af",
"hash_cont_tokens": "d8ae57c62ca85f4b"
},
"truncated": 0,
"non_truncated": 295,
"padded": 590,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|alghafa:mcq_exams_test_ar|0": {
"hashes": {
"hash_examples": "c07a5e78c5c0b8fe",
"hash_full_prompts": "c07a5e78c5c0b8fe",
"hash_input_tokens": "4812b6976e7cbee4",
"hash_cont_tokens": "478fa268e1fd1e4d"
},
"truncated": 0,
"non_truncated": 557,
"padded": 2228,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|alghafa:meta_ar_dialects|0": {
"hashes": {
"hash_examples": "c0b6081f83e14064",
"hash_full_prompts": "c0b6081f83e14064",
"hash_input_tokens": "753ce1027cb5a251",
"hash_cont_tokens": "3dea80a1ee9dc316"
},
"truncated": 0,
"non_truncated": 5395,
"padded": 21451,
"non_padded": 129,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|alghafa:meta_ar_msa|0": {
"hashes": {
"hash_examples": "64eb78a7c5b7484b",
"hash_full_prompts": "64eb78a7c5b7484b",
"hash_input_tokens": "c2e69e4759dfa85f",
"hash_cont_tokens": "b7b27059aa46f9f5"
},
"truncated": 0,
"non_truncated": 895,
"padded": 3554,
"non_padded": 26,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|alghafa:multiple_choice_facts_truefalse_balanced_task|0": {
"hashes": {
"hash_examples": "54fc3502c1c02c06",
"hash_full_prompts": "54fc3502c1c02c06",
"hash_input_tokens": "f314743540d19edf",
"hash_cont_tokens": "30ae320a6284bd96"
},
"truncated": 0,
"non_truncated": 75,
"padded": 150,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|alghafa:multiple_choice_grounded_statement_soqal_task|0": {
"hashes": {
"hash_examples": "46572d83696552ae",
"hash_full_prompts": "46572d83696552ae",
"hash_input_tokens": "cf890efdb961be06",
"hash_cont_tokens": "2e89ca40d66b31a0"
},
"truncated": 0,
"non_truncated": 150,
"padded": 743,
"non_padded": 7,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|alghafa:multiple_choice_grounded_statement_xglue_mlqa_task|0": {
"hashes": {
"hash_examples": "f430d97ff715bc1c",
"hash_full_prompts": "f430d97ff715bc1c",
"hash_input_tokens": "1fd6dfd6c4f3e840",
"hash_cont_tokens": "6e7f62230276d03d"
},
"truncated": 0,
"non_truncated": 150,
"padded": 746,
"non_padded": 4,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|alghafa:multiple_choice_rating_sentiment_no_neutral_task|0": {
"hashes": {
"hash_examples": "6b70a7416584f98c",
"hash_full_prompts": "6b70a7416584f98c",
"hash_input_tokens": "8ede0deff0ca4402",
"hash_cont_tokens": "cd589f2d0662aca9"
},
"truncated": 0,
"non_truncated": 7995,
"padded": 15990,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|alghafa:multiple_choice_rating_sentiment_task|0": {
"hashes": {
"hash_examples": "bc2005cc9d2f436e",
"hash_full_prompts": "bc2005cc9d2f436e",
"hash_input_tokens": "f82c929764d98b6c",
"hash_cont_tokens": "43c11b648549da66"
},
"truncated": 0,
"non_truncated": 5995,
"padded": 17843,
"non_padded": 142,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|alghafa:multiple_choice_sentiment_task|0": {
"hashes": {
"hash_examples": "6fb0e254ea5945d8",
"hash_full_prompts": "6fb0e254ea5945d8",
"hash_input_tokens": "1520cda76edf4ddf",
"hash_cont_tokens": "040bf50c610943db"
},
"truncated": 0,
"non_truncated": 1720,
"padded": 5061,
"non_padded": 99,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_exams|0": {
"hashes": {
"hash_examples": "6d721df351722656",
"hash_full_prompts": "6d721df351722656",
"hash_input_tokens": "455dfc728ef96014",
"hash_cont_tokens": "f6dc9d4abb83a50a"
},
"truncated": 0,
"non_truncated": 537,
"padded": 2116,
"non_padded": 32,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:abstract_algebra|0": {
"hashes": {
"hash_examples": "f2ddca8f45c0a511",
"hash_full_prompts": "f2ddca8f45c0a511",
"hash_input_tokens": "7858cfffc6425214",
"hash_cont_tokens": "67c9ff842b18298a"
},
"truncated": 0,
"non_truncated": 100,
"padded": 400,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:anatomy|0": {
"hashes": {
"hash_examples": "dfdbc1b83107668d",
"hash_full_prompts": "dfdbc1b83107668d",
"hash_input_tokens": "4b1111d0946349b3",
"hash_cont_tokens": "b5c1c612518185a5"
},
"truncated": 0,
"non_truncated": 135,
"padded": 532,
"non_padded": 8,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:astronomy|0": {
"hashes": {
"hash_examples": "9736a606002a848e",
"hash_full_prompts": "9736a606002a848e",
"hash_input_tokens": "657e8992cfef92af",
"hash_cont_tokens": "9a91066ba51b2074"
},
"truncated": 0,
"non_truncated": 152,
"padded": 608,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:business_ethics|0": {
"hashes": {
"hash_examples": "735e452fbb6dc63d",
"hash_full_prompts": "735e452fbb6dc63d",
"hash_input_tokens": "b2a236a7ab3d205c",
"hash_cont_tokens": "67c9ff842b18298a"
},
"truncated": 0,
"non_truncated": 100,
"padded": 396,
"non_padded": 4,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:clinical_knowledge|0": {
"hashes": {
"hash_examples": "6ab0ca4da98aedcf",
"hash_full_prompts": "6ab0ca4da98aedcf",
"hash_input_tokens": "bc1a859d2002fda7",
"hash_cont_tokens": "4a838d5fa832139f"
},
"truncated": 0,
"non_truncated": 265,
"padded": 1044,
"non_padded": 16,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:college_biology|0": {
"hashes": {
"hash_examples": "17e4e390848018a4",
"hash_full_prompts": "17e4e390848018a4",
"hash_input_tokens": "ee810c1450628676",
"hash_cont_tokens": "d95798e261e3ecf2"
},
"truncated": 0,
"non_truncated": 144,
"padded": 568,
"non_padded": 8,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:college_chemistry|0": {
"hashes": {
"hash_examples": "4abb169f6dfd234b",
"hash_full_prompts": "4abb169f6dfd234b",
"hash_input_tokens": "62b5efc891e1db95",
"hash_cont_tokens": "67c9ff842b18298a"
},
"truncated": 0,
"non_truncated": 100,
"padded": 392,
"non_padded": 8,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:college_computer_science|0": {
"hashes": {
"hash_examples": "a369e2e941358a1e",
"hash_full_prompts": "a369e2e941358a1e",
"hash_input_tokens": "b0e6c3e2f44b99c2",
"hash_cont_tokens": "67c9ff842b18298a"
},
"truncated": 0,
"non_truncated": 100,
"padded": 400,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:college_mathematics|0": {
"hashes": {
"hash_examples": "d7be03b8b6020bff",
"hash_full_prompts": "d7be03b8b6020bff",
"hash_input_tokens": "00a1737df2431aad",
"hash_cont_tokens": "67c9ff842b18298a"
},
"truncated": 0,
"non_truncated": 100,
"padded": 392,
"non_padded": 8,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:college_medicine|0": {
"hashes": {
"hash_examples": "0518a00f097346bf",
"hash_full_prompts": "0518a00f097346bf",
"hash_input_tokens": "807f8db1bb06c3c5",
"hash_cont_tokens": "0a01d731701f68e5"
},
"truncated": 0,
"non_truncated": 173,
"padded": 680,
"non_padded": 12,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:college_physics|0": {
"hashes": {
"hash_examples": "5d842cd49bc70e12",
"hash_full_prompts": "5d842cd49bc70e12",
"hash_input_tokens": "8d07b5bb7da46fa6",
"hash_cont_tokens": "77e73d9510077678"
},
"truncated": 0,
"non_truncated": 102,
"padded": 404,
"non_padded": 4,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:computer_security|0": {
"hashes": {
"hash_examples": "8e85d9f85be9b32f",
"hash_full_prompts": "8e85d9f85be9b32f",
"hash_input_tokens": "f40a4d9292aaaeb2",
"hash_cont_tokens": "67c9ff842b18298a"
},
"truncated": 0,
"non_truncated": 100,
"padded": 392,
"non_padded": 8,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:conceptual_physics|0": {
"hashes": {
"hash_examples": "7964b55a0a49502b",
"hash_full_prompts": "7964b55a0a49502b",
"hash_input_tokens": "49a11009d47c09af",
"hash_cont_tokens": "01df071ebfaec74d"
},
"truncated": 0,
"non_truncated": 235,
"padded": 896,
"non_padded": 44,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:econometrics|0": {
"hashes": {
"hash_examples": "1e192eae38347257",
"hash_full_prompts": "1e192eae38347257",
"hash_input_tokens": "3eeb17363808c24e",
"hash_cont_tokens": "e3fc2917921a9eaf"
},
"truncated": 0,
"non_truncated": 114,
"padded": 432,
"non_padded": 24,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:electrical_engineering|0": {
"hashes": {
"hash_examples": "cf97671d5c441da1",
"hash_full_prompts": "cf97671d5c441da1",
"hash_input_tokens": "03cb887116deb2ea",
"hash_cont_tokens": "c8a5aa308b735a83"
},
"truncated": 0,
"non_truncated": 145,
"padded": 560,
"non_padded": 20,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:elementary_mathematics|0": {
"hashes": {
"hash_examples": "6f49107ed43c40c5",
"hash_full_prompts": "6f49107ed43c40c5",
"hash_input_tokens": "fb131224a2771748",
"hash_cont_tokens": "df4fdf22c42e07a2"
},
"truncated": 0,
"non_truncated": 378,
"padded": 1488,
"non_padded": 24,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:formal_logic|0": {
"hashes": {
"hash_examples": "7922c376008ba77b",
"hash_full_prompts": "7922c376008ba77b",
"hash_input_tokens": "1d931041ffe8d2da",
"hash_cont_tokens": "3a65271847fd1f2e"
},
"truncated": 0,
"non_truncated": 126,
"padded": 496,
"non_padded": 8,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:global_facts|0": {
"hashes": {
"hash_examples": "11f9813185047d5b",
"hash_full_prompts": "11f9813185047d5b",
"hash_input_tokens": "cd6087dfa6ca0f41",
"hash_cont_tokens": "67c9ff842b18298a"
},
"truncated": 0,
"non_truncated": 100,
"padded": 380,
"non_padded": 20,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:high_school_biology|0": {
"hashes": {
"hash_examples": "2a804b1d90cbe66e",
"hash_full_prompts": "2a804b1d90cbe66e",
"hash_input_tokens": "3e9dd78feacde5ab",
"hash_cont_tokens": "a6b83eb9e3d31488"
},
"truncated": 0,
"non_truncated": 310,
"padded": 1212,
"non_padded": 28,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:high_school_chemistry|0": {
"hashes": {
"hash_examples": "0032168adabc53b4",
"hash_full_prompts": "0032168adabc53b4",
"hash_input_tokens": "4fd13f56dca097e4",
"hash_cont_tokens": "8a704d95abde0656"
},
"truncated": 0,
"non_truncated": 203,
"padded": 796,
"non_padded": 16,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:high_school_computer_science|0": {
"hashes": {
"hash_examples": "f2fb8740f9df980f",
"hash_full_prompts": "f2fb8740f9df980f",
"hash_input_tokens": "a57eef8134d049e0",
"hash_cont_tokens": "67c9ff842b18298a"
},
"truncated": 0,
"non_truncated": 100,
"padded": 392,
"non_padded": 8,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:high_school_european_history|0": {
"hashes": {
"hash_examples": "73509021e7e66435",
"hash_full_prompts": "73509021e7e66435",
"hash_input_tokens": "da584f9c9bb7979e",
"hash_cont_tokens": "bc6082a11551e6de"
},
"truncated": 0,
"non_truncated": 165,
"padded": 576,
"non_padded": 84,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:high_school_geography|0": {
"hashes": {
"hash_examples": "9e08d1894940ff42",
"hash_full_prompts": "9e08d1894940ff42",
"hash_input_tokens": "abe7e7df3659e3f5",
"hash_cont_tokens": "17f69684d1b7da75"
},
"truncated": 0,
"non_truncated": 198,
"padded": 772,
"non_padded": 20,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:high_school_government_and_politics|0": {
"hashes": {
"hash_examples": "64b7e97817ca6c76",
"hash_full_prompts": "64b7e97817ca6c76",
"hash_input_tokens": "ad4386ff272c8c6a",
"hash_cont_tokens": "bfdbe0094ac4e89d"
},
"truncated": 0,
"non_truncated": 193,
"padded": 764,
"non_padded": 8,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:high_school_macroeconomics|0": {
"hashes": {
"hash_examples": "9f582da8534bd2ef",
"hash_full_prompts": "9f582da8534bd2ef",
"hash_input_tokens": "12003b73793b6c1c",
"hash_cont_tokens": "a61cb28cbed86a76"
},
"truncated": 0,
"non_truncated": 390,
"padded": 1532,
"non_padded": 28,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:high_school_mathematics|0": {
"hashes": {
"hash_examples": "fd54f1c10d423c51",
"hash_full_prompts": "fd54f1c10d423c51",
"hash_input_tokens": "dcee8c64fa2bea63",
"hash_cont_tokens": "ede1813083303def"
},
"truncated": 0,
"non_truncated": 270,
"padded": 1068,
"non_padded": 12,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:high_school_microeconomics|0": {
"hashes": {
"hash_examples": "7037896925aaf42f",
"hash_full_prompts": "7037896925aaf42f",
"hash_input_tokens": "39faab0baaafb922",
"hash_cont_tokens": "2ce9a16a87758ab3"
},
"truncated": 0,
"non_truncated": 238,
"padded": 920,
"non_padded": 32,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:high_school_physics|0": {
"hashes": {
"hash_examples": "60c3776215167dae",
"hash_full_prompts": "60c3776215167dae",
"hash_input_tokens": "342071ea073c01bf",
"hash_cont_tokens": "34e90e2adee42b92"
},
"truncated": 0,
"non_truncated": 151,
"padded": 580,
"non_padded": 24,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:high_school_psychology|0": {
"hashes": {
"hash_examples": "61176bfd5da1298f",
"hash_full_prompts": "61176bfd5da1298f",
"hash_input_tokens": "6dd2d293b2bc5e82",
"hash_cont_tokens": "2d376df2c1814495"
},
"truncated": 0,
"non_truncated": 545,
"padded": 2140,
"non_padded": 40,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:high_school_statistics|0": {
"hashes": {
"hash_examples": "40dfeebd1ea10f76",
"hash_full_prompts": "40dfeebd1ea10f76",
"hash_input_tokens": "ba20aa6141d96617",
"hash_cont_tokens": "4bd9660a3b058b49"
},
"truncated": 0,
"non_truncated": 216,
"padded": 856,
"non_padded": 8,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:high_school_us_history|0": {
"hashes": {
"hash_examples": "03daa510ba917f4d",
"hash_full_prompts": "03daa510ba917f4d",
"hash_input_tokens": "9e89dd0527da9fe6",
"hash_cont_tokens": "5c36ec463d3a0755"
},
"truncated": 0,
"non_truncated": 204,
"padded": 788,
"non_padded": 28,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:high_school_world_history|0": {
"hashes": {
"hash_examples": "be075ffd579f43c2",
"hash_full_prompts": "be075ffd579f43c2",
"hash_input_tokens": "7677f982ce5bf4e2",
"hash_cont_tokens": "f21688a7138caf58"
},
"truncated": 0,
"non_truncated": 237,
"padded": 872,
"non_padded": 76,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:human_aging|0": {
"hashes": {
"hash_examples": "caa5b69f640bd1ef",
"hash_full_prompts": "caa5b69f640bd1ef",
"hash_input_tokens": "67c6460b4e389bed",
"hash_cont_tokens": "22b919059cbabb52"
},
"truncated": 0,
"non_truncated": 223,
"padded": 868,
"non_padded": 24,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:human_sexuality|0": {
"hashes": {
"hash_examples": "5ed2e38fb25a3767",
"hash_full_prompts": "5ed2e38fb25a3767",
"hash_input_tokens": "9a8027a4c32c679c",
"hash_cont_tokens": "7033845bac78be24"
},
"truncated": 0,
"non_truncated": 131,
"padded": 524,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:international_law|0": {
"hashes": {
"hash_examples": "4e3e9e28d1b96484",
"hash_full_prompts": "4e3e9e28d1b96484",
"hash_input_tokens": "b2d48ab2870e7c34",
"hash_cont_tokens": "29048e59854cec5c"
},
"truncated": 0,
"non_truncated": 121,
"padded": 476,
"non_padded": 8,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:jurisprudence|0": {
"hashes": {
"hash_examples": "e264b755366310b3",
"hash_full_prompts": "e264b755366310b3",
"hash_input_tokens": "80cd61ca4914a391",
"hash_cont_tokens": "fae0419b42375fd2"
},
"truncated": 0,
"non_truncated": 108,
"padded": 420,
"non_padded": 12,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:logical_fallacies|0": {
"hashes": {
"hash_examples": "a4ab6965a3e38071",
"hash_full_prompts": "a4ab6965a3e38071",
"hash_input_tokens": "eeeafa883dba0880",
"hash_cont_tokens": "93c4e716b01bd87c"
},
"truncated": 0,
"non_truncated": 163,
"padded": 636,
"non_padded": 16,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:machine_learning|0": {
"hashes": {
"hash_examples": "b92320efa6636b40",
"hash_full_prompts": "b92320efa6636b40",
"hash_input_tokens": "e880d2d1ae98c564",
"hash_cont_tokens": "2e4467ffdab3254d"
},
"truncated": 0,
"non_truncated": 112,
"padded": 432,
"non_padded": 16,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:management|0": {
"hashes": {
"hash_examples": "c9ee4872a850fe20",
"hash_full_prompts": "c9ee4872a850fe20",
"hash_input_tokens": "5080848bd10b3e5b",
"hash_cont_tokens": "b0936addbab0c265"
},
"truncated": 0,
"non_truncated": 103,
"padded": 408,
"non_padded": 4,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:marketing|0": {
"hashes": {
"hash_examples": "0c151b70f6a047e3",
"hash_full_prompts": "0c151b70f6a047e3",
"hash_input_tokens": "9ee648fb81a8a3ea",
"hash_cont_tokens": "54e2619fba846f17"
},
"truncated": 0,
"non_truncated": 234,
"padded": 920,
"non_padded": 16,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:medical_genetics|0": {
"hashes": {
"hash_examples": "513f6cb8fca3a24e",
"hash_full_prompts": "513f6cb8fca3a24e",
"hash_input_tokens": "8339b42d523bf66b",
"hash_cont_tokens": "67c9ff842b18298a"
},
"truncated": 0,
"non_truncated": 100,
"padded": 388,
"non_padded": 12,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:miscellaneous|0": {
"hashes": {
"hash_examples": "259a190d635331db",
"hash_full_prompts": "259a190d635331db",
"hash_input_tokens": "35a748a0356b70d4",
"hash_cont_tokens": "b68c3a07a4a75876"
},
"truncated": 0,
"non_truncated": 783,
"padded": 3088,
"non_padded": 44,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:moral_disputes|0": {
"hashes": {
"hash_examples": "b85052c48a0b7bc3",
"hash_full_prompts": "b85052c48a0b7bc3",
"hash_input_tokens": "2c001a64fc71d2f6",
"hash_cont_tokens": "f54406d1e4cf99f8"
},
"truncated": 0,
"non_truncated": 346,
"padded": 1348,
"non_padded": 36,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:moral_scenarios|0": {
"hashes": {
"hash_examples": "28d0b069ef00dd00",
"hash_full_prompts": "28d0b069ef00dd00",
"hash_input_tokens": "2abc1b6e7e5f9c01",
"hash_cont_tokens": "1a38cd5b4241444e"
},
"truncated": 0,
"non_truncated": 895,
"padded": 3580,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:nutrition|0": {
"hashes": {
"hash_examples": "00c9bc5f1d305b2f",
"hash_full_prompts": "00c9bc5f1d305b2f",
"hash_input_tokens": "727553c742bb2c70",
"hash_cont_tokens": "9d443ff23bc12c12"
},
"truncated": 0,
"non_truncated": 306,
"padded": 1192,
"non_padded": 32,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:philosophy|0": {
"hashes": {
"hash_examples": "a458c08454a3fd5f",
"hash_full_prompts": "a458c08454a3fd5f",
"hash_input_tokens": "57add157acb7e940",
"hash_cont_tokens": "09e5454b7258b0b7"
},
"truncated": 0,
"non_truncated": 311,
"padded": 1204,
"non_padded": 40,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:prehistory|0": {
"hashes": {
"hash_examples": "d6a0ecbdbb670e9c",
"hash_full_prompts": "d6a0ecbdbb670e9c",
"hash_input_tokens": "127750b7ad1469d6",
"hash_cont_tokens": "f20b5fcd2df4488d"
},
"truncated": 0,
"non_truncated": 324,
"padded": 1272,
"non_padded": 24,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:professional_accounting|0": {
"hashes": {
"hash_examples": "b4a95fe480b6540e",
"hash_full_prompts": "b4a95fe480b6540e",
"hash_input_tokens": "948721299e30d644",
"hash_cont_tokens": "a3a4b0df5e20638f"
},
"truncated": 0,
"non_truncated": 282,
"padded": 1112,
"non_padded": 16,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:professional_law|0": {
"hashes": {
"hash_examples": "c2be9651cdbdde3b",
"hash_full_prompts": "c2be9651cdbdde3b",
"hash_input_tokens": "39761a4442ef3036",
"hash_cont_tokens": "eb156ebe8faf1aaf"
},
"truncated": 0,
"non_truncated": 1534,
"padded": 6080,
"non_padded": 56,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:professional_medicine|0": {
"hashes": {
"hash_examples": "26ce92416288f273",
"hash_full_prompts": "26ce92416288f273",
"hash_input_tokens": "bf9adf3ed824bf11",
"hash_cont_tokens": "0218217c2fa604a0"
},
"truncated": 0,
"non_truncated": 272,
"padded": 1076,
"non_padded": 12,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:professional_psychology|0": {
"hashes": {
"hash_examples": "71ea5f182ea9a641",
"hash_full_prompts": "71ea5f182ea9a641",
"hash_input_tokens": "e6b8cc529a34bac6",
"hash_cont_tokens": "4fea6b16917c8330"
},
"truncated": 0,
"non_truncated": 612,
"padded": 2396,
"non_padded": 52,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:public_relations|0": {
"hashes": {
"hash_examples": "125adc21f91f8d77",
"hash_full_prompts": "125adc21f91f8d77",
"hash_input_tokens": "468b7ccadfc686a7",
"hash_cont_tokens": "fae64d3b41255dc8"
},
"truncated": 0,
"non_truncated": 110,
"padded": 420,
"non_padded": 20,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:security_studies|0": {
"hashes": {
"hash_examples": "3c18b216c099fb26",
"hash_full_prompts": "3c18b216c099fb26",
"hash_input_tokens": "3624bd1f4784c494",
"hash_cont_tokens": "f81c4b3cc61f9738"
},
"truncated": 0,
"non_truncated": 245,
"padded": 980,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:sociology|0": {
"hashes": {
"hash_examples": "3f2a9634cef7417d",
"hash_full_prompts": "3f2a9634cef7417d",
"hash_input_tokens": "43c816e3c9002b3a",
"hash_cont_tokens": "a68d71e598a7eb7f"
},
"truncated": 0,
"non_truncated": 201,
"padded": 768,
"non_padded": 36,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:us_foreign_policy|0": {
"hashes": {
"hash_examples": "22249da54056475e",
"hash_full_prompts": "22249da54056475e",
"hash_input_tokens": "3cbd6e9d0eb34d5b",
"hash_cont_tokens": "67c9ff842b18298a"
},
"truncated": 0,
"non_truncated": 100,
"padded": 380,
"non_padded": 20,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:virology|0": {
"hashes": {
"hash_examples": "9d194b9471dc624e",
"hash_full_prompts": "9d194b9471dc624e",
"hash_input_tokens": "23c589682c1232c7",
"hash_cont_tokens": "6c8625e5b2b8ffeb"
},
"truncated": 0,
"non_truncated": 166,
"padded": 664,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arabic_mmlu:world_religions|0": {
"hashes": {
"hash_examples": "229e5fe50082b064",
"hash_full_prompts": "229e5fe50082b064",
"hash_input_tokens": "8346238260c68093",
"hash_cont_tokens": "e0faaa109c671b8f"
},
"truncated": 0,
"non_truncated": 171,
"padded": 668,
"non_padded": 16,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arc_challenge_okapi_ar|0": {
"hashes": {
"hash_examples": "ab893807673bc355",
"hash_full_prompts": "ab893807673bc355",
"hash_input_tokens": "67bdcfb746f10cb1",
"hash_cont_tokens": "5ece5b0fdbfa8076"
},
"truncated": 0,
"non_truncated": 1160,
"padded": 4558,
"non_padded": 82,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|arc_easy_ar|0": {
"hashes": {
"hash_examples": "acb688624acc3d04",
"hash_full_prompts": "acb688624acc3d04",
"hash_input_tokens": "7d56c3da642cc95a",
"hash_cont_tokens": "f9e7ef6e6d49b466"
},
"truncated": 0,
"non_truncated": 2364,
"padded": 9254,
"non_padded": 202,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|boolq_ar|0": {
"hashes": {
"hash_examples": "48355a67867e0c32",
"hash_full_prompts": "48355a67867e0c32",
"hash_input_tokens": "e26c9f27ead9fb07",
"hash_cont_tokens": "6cc6329be7dcb0ef"
},
"truncated": 0,
"non_truncated": 3260,
"padded": 6474,
"non_padded": 46,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|copa_ext_ar|0": {
"hashes": {
"hash_examples": "9bb83301bb72eecf",
"hash_full_prompts": "9bb83301bb72eecf",
"hash_input_tokens": "b35c263eadac1729",
"hash_cont_tokens": "48c2323d601b2b0c"
},
"truncated": 0,
"non_truncated": 90,
"padded": 180,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|hellaswag_okapi_ar|0": {
"hashes": {
"hash_examples": "6e8cf57a322dfadd",
"hash_full_prompts": "6e8cf57a322dfadd",
"hash_input_tokens": "868c488359529ebe",
"hash_cont_tokens": "963f62478d9c5df7"
},
"truncated": 0,
"non_truncated": 9171,
"padded": 36560,
"non_padded": 124,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|openbook_qa_ext_ar|0": {
"hashes": {
"hash_examples": "923d41eb0aca93eb",
"hash_full_prompts": "923d41eb0aca93eb",
"hash_input_tokens": "a34ef18cc8920237",
"hash_cont_tokens": "e9d7e284d35d6b14"
},
"truncated": 0,
"non_truncated": 495,
"padded": 1949,
"non_padded": 31,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|piqa_ar|0": {
"hashes": {
"hash_examples": "94bc205a520d3ea0",
"hash_full_prompts": "94bc205a520d3ea0",
"hash_input_tokens": "885d72e00580f435",
"hash_cont_tokens": "6b782d2479e2c028"
},
"truncated": 0,
"non_truncated": 1833,
"padded": 3621,
"non_padded": 45,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|race_ar|0": {
"hashes": {
"hash_examples": "de65130bae647516",
"hash_full_prompts": "de65130bae647516",
"hash_input_tokens": "99ceeda1a94cf9be",
"hash_cont_tokens": "03d651c5338bf364"
},
"truncated": 0,
"non_truncated": 4929,
"padded": 19698,
"non_padded": 18,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|sciq_ar|0": {
"hashes": {
"hash_examples": "89327174c0761058",
"hash_full_prompts": "89327174c0761058",
"hash_input_tokens": "4d059ab1e6cc6c9e",
"hash_cont_tokens": "4bf272057f0dcde2"
},
"truncated": 0,
"non_truncated": 995,
"padded": 3961,
"non_padded": 19,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"community|toxigen_ar|0": {
"hashes": {
"hash_examples": "1e139513004a9a2e",
"hash_full_prompts": "1e139513004a9a2e",
"hash_input_tokens": "a6b4900440169d5f",
"hash_cont_tokens": "23c85267d1c209f3"
},
"truncated": 0,
"non_truncated": 935,
"padded": 1858,
"non_padded": 12,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"lighteval|xstory_cloze:ar|0": {
"hashes": {
"hash_examples": "865426a22c787481",
"hash_full_prompts": "865426a22c787481",
"hash_input_tokens": "b8ea36d9021ee8d6",
"hash_cont_tokens": "cb73d8896239c71e"
},
"truncated": 0,
"non_truncated": 1511,
"padded": 2980,
"non_padded": 42,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
}
},
"summary_general": {
"hashes": {
"hash_examples": "3785b6258eb7e042",
"hash_full_prompts": "3785b6258eb7e042",
"hash_input_tokens": "9abb83e9730a1b93",
"hash_cont_tokens": "89f5ee1f9ee7d57a"
},
"truncated": 0,
"non_truncated": 72964,
"padded": 233423,
"non_padded": 2200,
"num_truncated_few_shots": 0
}
}