hf-causal-experimental (pretrained=BEE-spoke-data/smol_llama-81M-tied,trust_remote_code=True,dtype=float), limit: None, provide_description: False, num_fewshot: 0, batch_size: 64 | Task |Version| Metric | Value | |Stderr| |--------------|------:|--------|------:|---|-----:| |arc_easy | 0|acc | 0.4162|± |0.0101| | | |acc_norm| 0.3885|± |0.0100| |boolq | 1|acc | 0.5832|± |0.0086| |lambada_openai| 0|ppl |79.4522|± |3.1355| | | |acc | 0.2523|± |0.0061| |openbookqa | 0|acc | 0.1540|± |0.0162| | | |acc_norm| 0.2780|± |0.0201| |piqa | 0|acc | 0.6050|± |0.0114| | | |acc_norm| 0.5898|± |0.0115| |winogrande | 0|acc | 0.5272|± |0.0140| hf-causal-experimental (pretrained=BEE-spoke-data/smol_llama-81M-tied,trust_remote_code=True,dtype=float), limit: None, provide_description: False, num_fewshot: 25, batch_size: 64 | Task |Version| Metric |Value | |Stderr| |-------------|------:|--------|-----:|---|-----:| |arc_challenge| 0|acc |0.1672|± |0.0109| | | |acc_norm|0.2218|± |0.0121| hf-causal-experimental (pretrained=BEE-spoke-data/smol_llama-81M-tied,trust_remote_code=True,dtype=float), limit: None, provide_description: False, num_fewshot: 10, batch_size: 64 | Task |Version| Metric |Value | |Stderr| |---------|------:|--------|-----:|---|-----:| |hellaswag| 0|acc |0.2769|± |0.0045| | | |acc_norm|0.2923|± |0.0045| hf-causal-experimental (pretrained=BEE-spoke-data/smol_llama-81M-tied,trust_remote_code=True,dtype=float), limit: None, provide_description: False, num_fewshot: 0, batch_size: 64 | Task |Version|Metric|Value | |Stderr| |-------------|------:|------|-----:|---|-----:| |truthfulqa_mc| 1|mc1 |0.2424|± |0.0150| | | |mc2 |0.4353|± |0.0152| hf-causal-experimental (pretrained=BEE-spoke-data/smol_llama-81M-tied,trust_remote_code=True,dtype=float), limit: None, provide_description: False, num_fewshot: 5, batch_size: 64 | Task |Version| Metric |Value | |Stderr| |-------------------------------------------------|------:|--------|-----:|---|-----:| |hendrycksTest-abstract_algebra | 1|acc |0.2200|± |0.0416| | | |acc_norm|0.2200|± |0.0416| |hendrycksTest-anatomy | 1|acc |0.2741|± |0.0385| | | |acc_norm|0.2741|± |0.0385| |hendrycksTest-astronomy | 1|acc |0.1776|± |0.0311| | | |acc_norm|0.1776|± |0.0311| |hendrycksTest-business_ethics | 1|acc |0.2100|± |0.0409| | | |acc_norm|0.2100|± |0.0409| |hendrycksTest-clinical_knowledge | 1|acc |0.2264|± |0.0258| | | |acc_norm|0.2264|± |0.0258| |hendrycksTest-college_biology | 1|acc |0.2361|± |0.0355| | | |acc_norm|0.2361|± |0.0355| |hendrycksTest-college_chemistry | 1|acc |0.1900|± |0.0394| | | |acc_norm|0.1900|± |0.0394| |hendrycksTest-college_computer_science | 1|acc |0.2100|± |0.0409| | | |acc_norm|0.2100|± |0.0409| |hendrycksTest-college_mathematics | 1|acc |0.1800|± |0.0386| | | |acc_norm|0.1800|± |0.0386| |hendrycksTest-college_medicine | 1|acc |0.2023|± |0.0306| | | |acc_norm|0.2023|± |0.0306| |hendrycksTest-college_physics | 1|acc |0.2157|± |0.0409| | | |acc_norm|0.2157|± |0.0409| |hendrycksTest-computer_security | 1|acc |0.2400|± |0.0429| | | |acc_norm|0.2400|± |0.0429| |hendrycksTest-conceptual_physics | 1|acc |0.2596|± |0.0287| | | |acc_norm|0.2596|± |0.0287| |hendrycksTest-econometrics | 1|acc |0.2544|± |0.0410| | | |acc_norm|0.2544|± |0.0410| |hendrycksTest-electrical_engineering | 1|acc |0.2207|± |0.0346| | | |acc_norm|0.2207|± |0.0346| |hendrycksTest-elementary_mathematics | 1|acc |0.2169|± |0.0212| | | |acc_norm|0.2169|± |0.0212| |hendrycksTest-formal_logic | 1|acc |0.1587|± |0.0327| | | |acc_norm|0.1587|± |0.0327| |hendrycksTest-global_facts | 1|acc |0.1900|± |0.0394| | | |acc_norm|0.1900|± |0.0394| |hendrycksTest-high_school_biology | 1|acc |0.3000|± |0.0261| | | |acc_norm|0.3000|± |0.0261| |hendrycksTest-high_school_chemistry | 1|acc |0.2808|± |0.0316| | | |acc_norm|0.2808|± |0.0316| |hendrycksTest-high_school_computer_science | 1|acc |0.2800|± |0.0451| | | |acc_norm|0.2800|± |0.0451| |hendrycksTest-high_school_european_history | 1|acc |0.2424|± |0.0335| | | |acc_norm|0.2424|± |0.0335| |hendrycksTest-high_school_geography | 1|acc |0.2576|± |0.0312| | | |acc_norm|0.2576|± |0.0312| |hendrycksTest-high_school_government_and_politics| 1|acc |0.2228|± |0.0300| | | |acc_norm|0.2228|± |0.0300| |hendrycksTest-high_school_macroeconomics | 1|acc |0.2231|± |0.0211| | | |acc_norm|0.2231|± |0.0211| |hendrycksTest-high_school_mathematics | 1|acc |0.2370|± |0.0259| | | |acc_norm|0.2370|± |0.0259| |hendrycksTest-high_school_microeconomics | 1|acc |0.2227|± |0.0270| | | |acc_norm|0.2227|± |0.0270| |hendrycksTest-high_school_physics | 1|acc |0.2053|± |0.0330| | | |acc_norm|0.2053|± |0.0330| |hendrycksTest-high_school_psychology | 1|acc |0.2110|± |0.0175| | | |acc_norm|0.2110|± |0.0175| |hendrycksTest-high_school_statistics | 1|acc |0.4120|± |0.0336| | | |acc_norm|0.4120|± |0.0336| |hendrycksTest-high_school_us_history | 1|acc |0.2990|± |0.0321| | | |acc_norm|0.2990|± |0.0321| |hendrycksTest-high_school_world_history | 1|acc |0.2658|± |0.0288| | | |acc_norm|0.2658|± |0.0288| |hendrycksTest-human_aging | 1|acc |0.2287|± |0.0282| | | |acc_norm|0.2287|± |0.0282| |hendrycksTest-human_sexuality | 1|acc |0.2595|± |0.0384| | | |acc_norm|0.2595|± |0.0384| |hendrycksTest-international_law | 1|acc |0.2975|± |0.0417| | | |acc_norm|0.2975|± |0.0417| |hendrycksTest-jurisprudence | 1|acc |0.2315|± |0.0408| | | |acc_norm|0.2315|± |0.0408| |hendrycksTest-logical_fallacies | 1|acc |0.2822|± |0.0354| | | |acc_norm|0.2822|± |0.0354| |hendrycksTest-machine_learning | 1|acc |0.2321|± |0.0401| | | |acc_norm|0.2321|± |0.0401| |hendrycksTest-management | 1|acc |0.1748|± |0.0376| | | |acc_norm|0.1748|± |0.0376| |hendrycksTest-marketing | 1|acc |0.2308|± |0.0276| | | |acc_norm|0.2308|± |0.0276| |hendrycksTest-medical_genetics | 1|acc |0.3000|± |0.0461| | | |acc_norm|0.3000|± |0.0461| |hendrycksTest-miscellaneous | 1|acc |0.2375|± |0.0152| | | |acc_norm|0.2375|± |0.0152| |hendrycksTest-moral_disputes | 1|acc |0.2486|± |0.0233| | | |acc_norm|0.2486|± |0.0233| |hendrycksTest-moral_scenarios | 1|acc |0.2425|± |0.0143| | | |acc_norm|0.2425|± |0.0143| |hendrycksTest-nutrition | 1|acc |0.2288|± |0.0241| | | |acc_norm|0.2288|± |0.0241| |hendrycksTest-philosophy | 1|acc |0.2090|± |0.0231| | | |acc_norm|0.2090|± |0.0231| |hendrycksTest-prehistory | 1|acc |0.2377|± |0.0237| | | |acc_norm|0.2377|± |0.0237| |hendrycksTest-professional_accounting | 1|acc |0.2234|± |0.0248| | | |acc_norm|0.2234|± |0.0248| |hendrycksTest-professional_law | 1|acc |0.2471|± |0.0110| | | |acc_norm|0.2471|± |0.0110| |hendrycksTest-professional_medicine | 1|acc |0.4081|± |0.0299| | | |acc_norm|0.4081|± |0.0299| |hendrycksTest-professional_psychology | 1|acc |0.2565|± |0.0177| | | |acc_norm|0.2565|± |0.0177| |hendrycksTest-public_relations | 1|acc |0.2182|± |0.0396| | | |acc_norm|0.2182|± |0.0396| |hendrycksTest-security_studies | 1|acc |0.2408|± |0.0274| | | |acc_norm|0.2408|± |0.0274| |hendrycksTest-sociology | 1|acc |0.2338|± |0.0299| | | |acc_norm|0.2338|± |0.0299| |hendrycksTest-us_foreign_policy | 1|acc |0.2500|± |0.0435| | | |acc_norm|0.2500|± |0.0435| |hendrycksTest-virology | 1|acc |0.2892|± |0.0353| | | |acc_norm|0.2892|± |0.0353| |hendrycksTest-world_religions | 1|acc |0.2105|± |0.0313| | | |acc_norm|0.2105|± |0.0313|