{ "name": "heavy", "uuid": "70f3f719-28b5-46f5-b58e-bd89765d1e40", "model": "/net/nfs.cirrascale/allennlp/davidw/checkpoints/moe-release/OLMoE-7B-A1B/step1220000-tokens5117B", "creation_date": "2024_08_06-17_05_21", "eval_metrics": { "icl": { "mmlu_zeroshot": 0.4273025756865217, "hellaswag_zeroshot": 0.7597092390060425, "jeopardy": 0.4871806979179382, "triviaqa_sm_sub": 0.5303333401679993, "gsm8k_cot": 0.07354056090116501, "agi_eval_sat_math_cot": 0.040909089148044586, "aqua_cot": 0.02857142873108387, "svamp_cot": 0.2800000011920929, "bigbench_qa_wikidata": 0.6884503960609436, "arc_easy": 0.7845118045806885, "arc_challenge": 0.5341296792030334, "mmlu_fewshot": 0.5194499257363772, "bigbench_misconceptions": 0.5525113940238953, "copa": 0.800000011920929, "siqa": 0.707267165184021, "commonsense_qa": 0.6699426770210266, "piqa": 0.8117519021034241, "openbook_qa": 0.4399999976158142, "bigbench_novel_concepts": 0.625, "bigbench_strange_stories": 0.6839080452919006, "bigbench_strategy_qa": 0.5810397267341614, "lambada_openai": 0.7219095826148987, "hellaswag": 0.7750447988510132, "winograd": 0.831501841545105, "winogrande": 0.6850828528404236, "bigbench_conlang_translation": 0.060975611209869385, "bigbench_language_identification": 0.34049999713897705, "bigbench_conceptual_combinations": 0.5048543810844421, "bigbench_elementary_math_qa": 0.27025681734085083, "bigbench_dyck_languages": 0.1589999943971634, "agi_eval_lsat_ar": 0.2652173936367035, "bigbench_cs_algorithms": 0.5015151500701904, "bigbench_logical_deduction": 0.25333333015441895, "bigbench_operators": 0.34285715222358704, "bigbench_repeat_copy_logic": 0.15625, "simple_arithmetic_nospaces": 0.1809999942779541, "simple_arithmetic_withspaces": 0.20600000023841858, "math_qa": 0.2708682417869568, "logi_qa": 0.3425499200820923, "pubmed_qa_labeled": 0.4659999907016754, "squad": 0.5244086980819702, "agi_eval_lsat_rc": 0.43656715750694275, "agi_eval_lsat_lr": 0.386274516582489, "coqa": 0.4436928331851959, "bigbench_understanding_fables": 0.4444444477558136, "boolq": 0.7281345725059509, "agi_eval_sat_en": 0.5485436916351318, "winogender_mc_female": 0.46666666865348816, "winogender_mc_male": 0.5833333134651184, "enterprise_pii_classification": 0.5372606515884399, "bbq": 0.48323566534302453, "gpqa_main": 0.2232142835855484, "gpqa_diamond": 0.21212121844291687 } }, "missing tasks": "[]", "aggregated_task_categories_centered": { "commonsense reasoning": 0.46081640452671535, "language understanding": 0.4762512398893946, "reading comprehension": 0.36885401178478144, "safety": 0.03524814952503552, "symbolic problem solving": 0.15957477013304083, "world knowledge": 0.34119598718414534 }, "aggregated_centered_results": 0.30931975984045923, "aggregated_results": 0.4599646118255447, "rw_small": 0.7152613600095113, "rw_small_centered": 0.5043107818441781, "95%_CI_above": 0.5589368432469957, "95%_CI_above_centered": 0.41426754302993024, "99%_CI_above": 0.5701544312031374, "99%_CI_above_centered": 0.45654352726393493, "low_variance_datasets": 0.5659450578418646, "low_variance_datasets_centered": 0.46520260353587645 }