run_id
large_stringlengths
64
64
timestamp
unknown
model_name_or_path
large_stringclasses
5 values
unitxt_recipe
large_stringlengths
326
371
quantization_type
large_stringclasses
1 value
quantization_bit_count
large_stringclasses
1 value
inference_runtime_s
float64
1.05
37.4
generation_args
large_stringclasses
1 value
model_args
large_stringclasses
5 values
inference_engine
large_stringclasses
1 value
packages_versions
large_stringclasses
1 value
scores
large_stringlengths
174
240
num_gpu
int64
1
1
device
large_stringclasses
1 value
f06d1e3627dd28d28a525fd5467b63bdadba95028b2299e7f601d0ec0e19333c
"2024-12-22T20:01:05.609000Z"
meta-llama/Llama-3.2-3B-Instruct
card=cards.mmlu.high_school_macroeconomics,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopicHelm.enumerator_numbers_choicesSeparator_pipe_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.967827
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-3B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.62, 'score': 0.62, 'score_name': 'accuracy', 'accuracy_ci_low': 0.53, 'accuracy_ci_high': 0.72, 'score_ci_low': 0.53, 'score_ci_high': 0.72}
1
a100_80gb
f51697baf05a9d07e493d2940653bfae2fab444a780918f13a68536763c1ae24
"2024-12-22T20:01:11.444000Z"
meta-llama/Llama-3.2-3B-Instruct
card=cards.mmlu.high_school_macroeconomics,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopicHelm.enumerator_numbers_choicesSeparator_OrCapital_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
4.639898
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-3B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.51, 'score': 0.51, 'score_name': 'accuracy', 'accuracy_ci_low': 0.42, 'accuracy_ci_high': 0.61, 'score_ci_low': 0.42, 'score_ci_high': 0.61}
1
a100_80gb
3778475bce0c323a1495e00234a15f5fd58afd059b14aa8fca538222c85b2ff9
"2024-12-22T20:01:17.788000Z"
meta-llama/Llama-3.2-3B-Instruct
card=cards.mmlu.high_school_macroeconomics,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopicHelm.enumerator_numbers_choicesSeparator_OrCapital_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
5.138048
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-3B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.57, 'score': 0.57, 'score_name': 'accuracy', 'accuracy_ci_low': 0.48, 'accuracy_ci_high': 0.67, 'score_ci_low': 0.48, 'score_ci_high': 0.67}
1
a100_80gb
774533ad71acb7cccaaca65f60561e96ce3cb2031140372ba0c88cd870215045
"2024-12-22T20:01:22.865000Z"
meta-llama/Llama-3.2-3B-Instruct
card=cards.mmlu.high_school_macroeconomics,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopicHelm.enumerator_numbers_choicesSeparator_orLower_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.901287
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-3B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.5, 'score': 0.5, 'score_name': 'accuracy', 'accuracy_ci_low': 0.4086616654817124, 'accuracy_ci_high': 0.6, 'score_ci_low': 0.4086616654817124, 'score_ci_high': 0.6}
1
a100_80gb
7a7f45cc7ed01ee968e82a95fa99ced6a28a55c86f4ec2280ae317dc2cb05203
"2024-12-22T20:01:28.031000Z"
meta-llama/Llama-3.2-3B-Instruct
card=cards.mmlu.high_school_macroeconomics,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopicHelm.enumerator_numbers_choicesSeparator_orLower_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.972339
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-3B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.59, 'score': 0.59, 'score_name': 'accuracy', 'accuracy_ci_low': 0.5, 'accuracy_ci_high': 0.69, 'score_ci_low': 0.5, 'score_ci_high': 0.69}
1
a100_80gb
d0e4895d6a6b57789d43337e352655ded7263c57ad3e1a6d8a76b550a7dd29a9
"2024-12-22T20:00:23.485000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.high_school_geography,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopicHelm.enumerator_numbers_choicesSeparator_OrCapital_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
8.696543
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.75, 'score': 0.75, 'score_name': 'accuracy', 'accuracy_ci_low': 0.6533793665648993, 'accuracy_ci_high': 0.83, 'score_ci_low': 0.6533793665648993, 'score_ci_high': 0.83}
1
a100_80gb
07303e05bd95c6e365b29cd782f881db91ec75ca909028700b9fc7eb7c3c1103
"2024-12-22T20:00:30.915000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.high_school_geography,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopicHelm.enumerator_numbers_choicesSeparator_orLower_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
6.117618
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.69, 'score': 0.69, 'score_name': 'accuracy', 'accuracy_ci_low': 0.6, 'accuracy_ci_high': 0.77, 'score_ci_low': 0.6, 'score_ci_high': 0.77}
1
a100_80gb
17cbb021cef9c4d7a7c1e2d3ff3de471e47ba6ee4a85e1593bcef876a98b2fb2
"2024-12-22T20:00:42.011000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.high_school_geography,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopicHelm.enumerator_numbers_choicesSeparator_orLower_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
8.388628
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.72, 'score': 0.72, 'score_name': 'accuracy', 'accuracy_ci_low': 0.62, 'accuracy_ci_high': 0.8, 'score_ci_low': 0.62, 'score_ci_high': 0.8}
1
a100_80gb
02baaae35b439224843d38fddf0df920650ccc65e157bc4cfe39592d7eb1e6e8
"2024-12-22T20:00:48.979000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.high_school_geography,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopicHelm.enumerator_roman_choicesSeparator_space_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
5.70242
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.7, 'score': 0.7, 'score_name': 'accuracy', 'accuracy_ci_low': 0.6, 'accuracy_ci_high': 0.78, 'score_ci_low': 0.6, 'score_ci_high': 0.78}
1
a100_80gb
8c775173e593a98642e0ae1603d66f4f76b24fdd18a46d8d92a3f27adfe06034
"2024-12-22T20:00:57.021000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.high_school_geography,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopicHelm.enumerator_roman_choicesSeparator_space_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
6.924687
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.69, 'score': 0.69, 'score_name': 'accuracy', 'accuracy_ci_low': 0.59, 'accuracy_ci_high': 0.78, 'score_ci_low': 0.59, 'score_ci_high': 0.78}
1
a100_80gb
0eace5a54e6365dd800a9b9466163319a42c2557e117bc3a297119b51ed6f715
"2024-12-22T20:01:04.522000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.high_school_geography,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopicHelm.enumerator_roman_choicesSeparator_newline_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
6.335705
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.74, 'score': 0.74, 'score_name': 'accuracy', 'accuracy_ci_low': 0.6426965928005162, 'accuracy_ci_high': 0.82, 'score_ci_low': 0.6426965928005162, 'score_ci_high': 0.82}
1
a100_80gb
8b369a3d0939c341b054ccf8365e61e466004d9a51c7cb1859e706b8b899d3b1
"2024-12-22T20:01:11.446000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.high_school_geography,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopicHelm.enumerator_roman_choicesSeparator_newline_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
5.773009
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.71, 'score': 0.71, 'score_name': 'accuracy', 'accuracy_ci_low': 0.61, 'accuracy_ci_high': 0.79, 'score_ci_low': 0.61, 'score_ci_high': 0.79}
1
a100_80gb
63cea0cc3f5777ec26697c20259d5f99cc6c25a8d55ff67473b22e09d2e81502
"2024-12-22T20:01:18.174000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.high_school_geography,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopicHelm.enumerator_roman_choicesSeparator_comma_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
5.601466
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.7, 'score': 0.7, 'score_name': 'accuracy', 'accuracy_ci_low': 0.61, 'accuracy_ci_high': 0.78, 'score_ci_low': 0.61, 'score_ci_high': 0.78}
1
a100_80gb
ba6be585c56b34c2520ece7448a1e34b5541c88884787039af543edda52bed38
"2024-12-22T20:01:25.530000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.high_school_geography,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopicHelm.enumerator_roman_choicesSeparator_comma_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
6.222353
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.72, 'score': 0.72, 'score_name': 'accuracy', 'accuracy_ci_low': 0.62, 'accuracy_ci_high': 0.8, 'score_ci_low': 0.62, 'score_ci_high': 0.8}
1
a100_80gb
8b59fbedbc5ab58fa15e3a734f755ed55913eb366845788420992a3766db2477
"2024-12-22T20:01:32.287000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.high_school_geography,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopicHelm.enumerator_roman_choicesSeparator_semicolon_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
5.623621
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.77, 'score': 0.77, 'score_name': 'accuracy', 'accuracy_ci_low': 0.68, 'accuracy_ci_high': 0.84, 'score_ci_low': 0.68, 'score_ci_high': 0.84}
1
a100_80gb
1e4a52133c6015a7efb4556ef0f08d8af045c5529aabef77edd8df8857d865b6
"2024-12-22T20:00:06.714000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.college_computer_science,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopicHelm.enumerator_roman_choicesSeparator_space_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
8.940969
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.53, 'score': 0.53, 'score_name': 'accuracy', 'accuracy_ci_low': 0.43, 'accuracy_ci_high': 0.62, 'score_ci_low': 0.43, 'score_ci_high': 0.62}
1
a100_80gb
17d8b357961b14c0d9607013c012006a9049307dd2f549764c70f207506164fc
"2024-12-22T20:00:18.336000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.college_computer_science,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopicHelm.enumerator_roman_choicesSeparator_newline_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
9.925794
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.48, 'score': 0.48, 'score_name': 'accuracy', 'accuracy_ci_low': 0.39, 'accuracy_ci_high': 0.57, 'score_ci_low': 0.39, 'score_ci_high': 0.57}
1
a100_80gb
021c4c59aa61b474bf7c88e3abccac580dcf29beb6f4817ed7ba42b91b3d6871
"2024-12-22T20:00:29.148000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.college_computer_science,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopicHelm.enumerator_roman_choicesSeparator_newline_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
9.090302
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.54, 'score': 0.54, 'score_name': 'accuracy', 'accuracy_ci_low': 0.45, 'accuracy_ci_high': 0.63, 'score_ci_low': 0.45, 'score_ci_high': 0.63}
1
a100_80gb
f7523fdca6c602aff36d370eff5815cb2c8a4f59430a22ca286bcc8c3e631973
"2024-12-22T20:00:40.654000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.college_computer_science,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopicHelm.enumerator_roman_choicesSeparator_comma_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
9.785626
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.49, 'score': 0.49, 'score_name': 'accuracy', 'accuracy_ci_low': 0.39, 'accuracy_ci_high': 0.5846383870584236, 'score_ci_low': 0.39, 'score_ci_high': 0.5846383870584236}
1
a100_80gb
3134eb82dc233521cbff8d6012d9e673e36bc9dc60b1b11279cbf15859b8ab72
"2024-12-22T20:00:52.264000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.college_computer_science,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopicHelm.enumerator_roman_choicesSeparator_comma_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
9.883312
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.53, 'score': 0.53, 'score_name': 'accuracy', 'accuracy_ci_low': 0.44, 'accuracy_ci_high': 0.63, 'score_ci_low': 0.44, 'score_ci_high': 0.63}
1
a100_80gb
4ee142b9c19291b06445d0c0a47a94b38e4bf15e56241e1034c933735df29750
"2024-12-22T20:00:10.034000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.high_school_world_history,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_newline_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
8.24325
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.73, 'score': 0.73, 'score_name': 'accuracy', 'accuracy_ci_low': 0.63, 'accuracy_ci_high': 0.81, 'score_ci_low': 0.63, 'score_ci_high': 0.81}
1
a100_80gb
c456af7bb9cbe02bdac2b120857600c6b408e49325d95373c701a5d62e941a19
"2024-12-22T20:00:21.216000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.high_school_world_history,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_comma_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
8.323369
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.75, 'score': 0.75, 'score_name': 'accuracy', 'accuracy_ci_low': 0.66, 'accuracy_ci_high': 0.83, 'score_ci_low': 0.66, 'score_ci_high': 0.83}
1
a100_80gb
e6df2e41e4c554df6ba85ab6b5032ccc6649cd2222e450335af569c248487458
"2024-12-22T20:00:05.532000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.professional_medicine,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_numbers_choicesSeparator_orLower_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
5.670691
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.46, 'score': 0.46, 'score_name': 'accuracy', 'accuracy_ci_low': 0.36, 'accuracy_ci_high': 0.56, 'score_ci_low': 0.36, 'score_ci_high': 0.56}
1
a100_80gb
0ceb8616b79a6c2fa1ee62d81eb3f052f2b0adfb316e495f618468ff975a3b75
"2024-12-22T20:00:14.336000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.professional_medicine,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_roman_choicesSeparator_space_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
6.634117
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.41, 'score': 0.41, 'score_name': 'accuracy', 'accuracy_ci_low': 0.31700161580845054, 'accuracy_ci_high': 0.51, 'score_ci_low': 0.31700161580845054, 'score_ci_high': 0.51}
1
a100_80gb
88b7108a0107e71d9a173fb59b85b5d4be22d42a1b0057be959bb9c5c300c33e
"2024-12-22T20:00:21.997000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.professional_medicine,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_roman_choicesSeparator_space_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
5.56502
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.48, 'score': 0.48, 'score_name': 'accuracy', 'accuracy_ci_low': 0.39, 'accuracy_ci_high': 0.58, 'score_ci_low': 0.39, 'score_ci_high': 0.58}
1
a100_80gb
545d1320ef4b1491b209833dd0f65bb97e0de3f13668ee04f4f19a597793bdfa
"2024-12-22T20:00:01.169000Z"
meta-llama/Llama-3.2-3B-Instruct
card=cards.mmlu.high_school_macroeconomics,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopicHelm.enumerator_lowercase_choicesSeparator_orLower_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
6.307394
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-3B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.65, 'score': 0.65, 'score_name': 'accuracy', 'accuracy_ci_low': 0.55, 'accuracy_ci_high': 0.74, 'score_ci_low': 0.55, 'score_ci_high': 0.74}
1
a100_80gb
1b72c1efef73e1aee2011a86544af95e75d88230b5d9220bc7483309bc07b837
"2024-12-22T20:00:06.223000Z"
meta-llama/Llama-3.2-3B-Instruct
card=cards.mmlu.high_school_macroeconomics,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopicHelm.enumerator_lowercase_choicesSeparator_orLower_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.868965
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-3B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.54, 'score': 0.54, 'score_name': 'accuracy', 'accuracy_ci_low': 0.43, 'accuracy_ci_high': 0.63, 'score_ci_low': 0.43, 'score_ci_high': 0.63}
1
a100_80gb
8eef8cb4308eb3b2c47640848ced4e026149ab26797463bdc2b3dbf0f5b8ee80
"2024-12-22T20:00:11.402000Z"
meta-llama/Llama-3.2-3B-Instruct
card=cards.mmlu.high_school_macroeconomics,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopicHelm.enumerator_numbers_choicesSeparator_space_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.970108
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-3B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.52, 'score': 0.52, 'score_name': 'accuracy', 'accuracy_ci_low': 0.42, 'accuracy_ci_high': 0.61, 'score_ci_low': 0.42, 'score_ci_high': 0.61}
1
a100_80gb
f4184d7c252056d1262c4d58bff1cb189f812142d8351c9794ce753144720b47
"2024-12-22T20:00:17.917000Z"
meta-llama/Llama-3.2-3B-Instruct
card=cards.mmlu.high_school_macroeconomics,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopicHelm.enumerator_numbers_choicesSeparator_space_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
5.330562
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-3B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.58, 'score': 0.58, 'score_name': 'accuracy', 'accuracy_ci_low': 0.49, 'accuracy_ci_high': 0.68, 'score_ci_low': 0.49, 'score_ci_high': 0.68}
1
a100_80gb
b7dc768de91908552b0b72ebbc5621cbebfe79b3164962a9b27bf34d01f4144f
"2024-12-22T20:00:23.533000Z"
meta-llama/Llama-3.2-3B-Instruct
card=cards.mmlu.high_school_macroeconomics,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopicHelm.enumerator_numbers_choicesSeparator_newline_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
4.439801
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-3B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.51, 'score': 0.51, 'score_name': 'accuracy', 'accuracy_ci_low': 0.41, 'accuracy_ci_high': 0.61, 'score_ci_low': 0.41, 'score_ci_high': 0.61}
1
a100_80gb
d46f525305a1b0a181cf29db16ca94fec27fb4dde2921ed6351023871117ff81
"2024-12-22T20:00:28.525000Z"
meta-llama/Llama-3.2-3B-Instruct
card=cards.mmlu.high_school_macroeconomics,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopicHelm.enumerator_numbers_choicesSeparator_newline_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.821913
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-3B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.6, 'score': 0.6, 'score_name': 'accuracy', 'accuracy_ci_low': 0.51, 'accuracy_ci_high': 0.69, 'score_ci_low': 0.51, 'score_ci_high': 0.69}
1
a100_80gb
e1dc68e34ddf39184b0339b0b869ad187eda539db99549d04c0a671391d42cf3
"2024-12-22T20:00:06.393000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.high_school_geography,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopicHelm.enumerator_numbers_choicesSeparator_pipe_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
6.680029
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.71, 'score': 0.71, 'score_name': 'accuracy', 'accuracy_ci_low': 0.61, 'accuracy_ci_high': 0.79, 'score_ci_low': 0.61, 'score_ci_high': 0.79}
1
a100_80gb
b5d9a0d8db1e37597ab33c971b0ce019512ba248b558c9b07da5225122e7ef25
"2024-12-22T20:00:13.603000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.high_school_geography,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopicHelm.enumerator_numbers_choicesSeparator_OrCapital_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
5.978386
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.74, 'score': 0.74, 'score_name': 'accuracy', 'accuracy_ci_low': 0.64, 'accuracy_ci_high': 0.82, 'score_ci_low': 0.64, 'score_ci_high': 0.82}
1
a100_80gb
6fedbea17fce0fd8ab39473aacb54e31e35fec0b73e3671a071bce3934d167ea
"2024-12-22T20:00:02.019000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.moral_disputes,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopicHelm.enumerator_lowercase_choicesSeparator_orLower_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
7.479195
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.62, 'score': 0.62, 'score_name': 'accuracy', 'accuracy_ci_low': 0.52, 'accuracy_ci_high': 0.71, 'score_ci_low': 0.52, 'score_ci_high': 0.71}
1
a100_80gb
cb32552f1bf74269869d187737b0299ee5f56606cec7659a5679ddc62e27eb82
"2024-12-22T20:00:02.394000Z"
meta-llama/Meta-Llama-3-8B-Instruct
card=cards.mmlu.college_computer_science,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopicHelm.enumerator_roman_choicesSeparator_space_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
10.064471
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Meta-Llama-3-8B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.47, 'score': 0.47, 'score_name': 'accuracy', 'accuracy_ci_low': 0.38, 'accuracy_ci_high': 0.58, 'score_ci_low': 0.38, 'score_ci_high': 0.58}
1
a100_80gb
c5dfbefa003b46e0e99016bb8800d91788fe68aef3df4e280ae8e7c9fefde9ee
"2024-12-22T23:49:30.450000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_numbers_choicesSeparator_orLower_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.661437
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.28, 'score': 0.28, 'score_name': 'accuracy', 'accuracy_ci_low': 0.19977778158344378, 'accuracy_ci_high': 0.37, 'score_ci_low': 0.19977778158344378, 'score_ci_high': 0.37}
1
a100_80gb
f5572276d33a874372ace893efd55e247c2141c18e9244976e85c314e494c3ce
"2024-12-22T23:49:34.397000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_numbers_choicesSeparator_orLower_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.944961
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.32, 'score': 0.32, 'score_name': 'accuracy', 'accuracy_ci_low': 0.23, 'accuracy_ci_high': 0.4166454078812933, 'score_ci_low': 0.23, 'score_ci_high': 0.4166454078812933}
1
a100_80gb
a9ae4836cd7529ccaf9ec053a32d0db6576fc69414de67d1d69c8a8df849d7ea
"2024-12-22T23:49:38.436000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_roman_choicesSeparator_space_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
1.957382
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.26, 'score': 0.26, 'score_name': 'accuracy', 'accuracy_ci_low': 0.18, 'accuracy_ci_high': 0.36, 'score_ci_low': 0.18, 'score_ci_high': 0.36}
1
a100_80gb
49968fffdc03f07b6bbfd17434dbc9057ccb5038c995ae073dc5dbd8b0c048d0
"2024-12-22T23:49:41.566000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_roman_choicesSeparator_space_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.165575
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.32, 'score': 0.32, 'score_name': 'accuracy', 'accuracy_ci_low': 0.23, 'accuracy_ci_high': 0.41, 'score_ci_low': 0.23, 'score_ci_high': 0.41}
1
a100_80gb
ed104ddf3e491363db409166f704e93da00c07419efc3285d78f139539ef3ad2
"2024-12-22T23:49:44.737000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_roman_choicesSeparator_newline_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.189664
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.4, 'score': 0.4, 'score_name': 'accuracy', 'accuracy_ci_low': 0.31, 'accuracy_ci_high': 0.5, 'score_ci_low': 0.31, 'score_ci_high': 0.5}
1
a100_80gb
13f3765b73f32f6155b52203888663d9f3ba9dd9d709a8abcf2f898f954ec823
"2024-12-22T23:49:47.906000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_roman_choicesSeparator_newline_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.185605
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.43, 'score': 0.43, 'score_name': 'accuracy', 'accuracy_ci_low': 0.33, 'accuracy_ci_high': 0.53, 'score_ci_low': 0.33, 'score_ci_high': 0.53}
1
a100_80gb
44cce59f9249885eb48dc79e29372220b95bb118a1c4c409015dff8308ffcc77
"2024-12-22T23:49:50.994000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_roman_choicesSeparator_comma_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.100002
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.34, 'score': 0.34, 'score_name': 'accuracy', 'accuracy_ci_low': 0.25, 'accuracy_ci_high': 0.44, 'score_ci_low': 0.25, 'score_ci_high': 0.44}
1
a100_80gb
0925e12df1b4c5a2c7a3b2d927e008e0c78f9c3daadf4e4b7b3cb0f32c775a2c
"2024-12-22T23:49:55.804000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_roman_choicesSeparator_comma_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.824273
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.35, 'score': 0.35, 'score_name': 'accuracy', 'accuracy_ci_low': 0.26, 'accuracy_ci_high': 0.45, 'score_ci_low': 0.26, 'score_ci_high': 0.45}
1
a100_80gb
bfad2cd9576acd13c69a1238f81c9f6d42b0ca1ed6d91c054b62d7c042f4525e
"2024-12-22T23:49:59.589000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_roman_choicesSeparator_semicolon_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.806399
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.34, 'score': 0.34, 'score_name': 'accuracy', 'accuracy_ci_low': 0.26, 'accuracy_ci_high': 0.44, 'score_ci_low': 0.26, 'score_ci_high': 0.44}
1
a100_80gb
ab590f6a4a1645d51f5320f773b0afe3a5574a33b9f308e1ade2699fee7182bb
"2024-12-22T23:50:04.453000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_roman_choicesSeparator_semicolon_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.869399
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.37, 'score': 0.37, 'score_name': 'accuracy', 'accuracy_ci_low': 0.28, 'accuracy_ci_high': 0.47, 'score_ci_low': 0.28, 'score_ci_high': 0.47}
1
a100_80gb
101906de23c6ab1f8a95ee47ef4f317460931e3bf4267881fc3526340381aace
"2024-12-22T23:48:50.893000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_numbers_choicesSeparator_newline_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.191558
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.39, 'score': 0.39, 'score_name': 'accuracy', 'accuracy_ci_low': 0.31, 'accuracy_ci_high': 0.49, 'score_ci_low': 0.31, 'score_ci_high': 0.49}
1
a100_80gb
3dc0fd861d98219004c4baaa887c296d749a4c5ed07d5d516884d87ad13850fe
"2024-12-22T23:48:54.133000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_numbers_choicesSeparator_newline_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.259706
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.38, 'score': 0.38, 'score_name': 'accuracy', 'accuracy_ci_low': 0.29, 'accuracy_ci_high': 0.48, 'score_ci_low': 0.29, 'score_ci_high': 0.48}
1
a100_80gb
e684546fe7362e01fbf5c9f601ca96928abaaaa2c54928e4172faae7176a0747
"2024-12-22T23:48:57.489000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_numbers_choicesSeparator_comma_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.35603
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.35, 'score': 0.35, 'score_name': 'accuracy', 'accuracy_ci_low': 0.26, 'accuracy_ci_high': 0.45, 'score_ci_low': 0.26, 'score_ci_high': 0.45}
1
a100_80gb
00733aadcbfe514d673e19ce0039a4a8cae6ca59b2605a4e3ade573e15442a96
"2024-12-22T23:49:01.517000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_numbers_choicesSeparator_comma_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.305571
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.39, 'score': 0.39, 'score_name': 'accuracy', 'accuracy_ci_low': 0.2909280008095247, 'accuracy_ci_high': 0.49, 'score_ci_low': 0.2909280008095247, 'score_ci_high': 0.49}
1
a100_80gb
d184c0b27abba27eae8bcda26cafec3b6d0475905fc32d2eef9954cff0863193
"2024-12-22T23:49:06.035000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_numbers_choicesSeparator_semicolon_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.5199
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.38, 'score': 0.38, 'score_name': 'accuracy', 'accuracy_ci_low': 0.28635263227638214, 'accuracy_ci_high': 0.47, 'score_ci_low': 0.28635263227638214, 'score_ci_high': 0.47}
1
a100_80gb
2edadc3d3a31cfb469f3cd9d8525a6e687466a79f39b9aea23d5b09634876363
"2024-12-22T23:49:10.858000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_numbers_choicesSeparator_semicolon_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.29807
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.43, 'score': 0.43, 'score_name': 'accuracy', 'accuracy_ci_low': 0.3344602494774245, 'accuracy_ci_high': 0.53, 'score_ci_low': 0.3344602494774245, 'score_ci_high': 0.53}
1
a100_80gb
999e75bf6a126d29eade8e3e557ab46c7ba0e358ba653af6ff2bd75529013795
"2024-12-22T23:49:14.154000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_numbers_choicesSeparator_pipe_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.300538
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.37, 'score': 0.37, 'score_name': 'accuracy', 'accuracy_ci_low': 0.28, 'accuracy_ci_high': 0.47, 'score_ci_low': 0.28, 'score_ci_high': 0.47}
1
a100_80gb
419e7499e642c1a30040309ac225de855ed9a6249eef423caf4e7228dce089e8
"2024-12-22T23:49:17.518000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_numbers_choicesSeparator_pipe_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.358036
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.4, 'score': 0.4, 'score_name': 'accuracy', 'accuracy_ci_low': 0.31, 'accuracy_ci_high': 0.5, 'score_ci_low': 0.31, 'score_ci_high': 0.5}
1
a100_80gb
a15063495caf93d6fe40c4e7994c483d1db77d114467f264f383e03e808d20c6
"2024-12-22T23:49:20.835000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_numbers_choicesSeparator_OrCapital_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.252075
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.32, 'score': 0.32, 'score_name': 'accuracy', 'accuracy_ci_low': 0.24, 'accuracy_ci_high': 0.41, 'score_ci_low': 0.24, 'score_ci_high': 0.41}
1
a100_80gb
b77e425c0d78ac0610357687081d8799f1394d8a1aba36ff993d3840c1a5940a
"2024-12-22T23:49:25.798000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_numbers_choicesSeparator_OrCapital_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.158651
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.33, 'score': 0.33, 'score_name': 'accuracy', 'accuracy_ci_low': 0.24, 'accuracy_ci_high': 0.43, 'score_ci_low': 0.24, 'score_ci_high': 0.43}
1
a100_80gb
b10b6aefda446e1fbf7235f77d15f9ce78dfccb745ffdabe2d0c08fab1e412c1
"2024-12-22T23:48:14.868000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_lowercase_choicesSeparator_semicolon_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.650313
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.37, 'score': 0.37, 'score_name': 'accuracy', 'accuracy_ci_low': 0.29, 'accuracy_ci_high': 0.47, 'score_ci_low': 0.29, 'score_ci_high': 0.47}
1
a100_80gb
4120c28308ad32049649f78d00d895b877b839903c85cdc75ed9fcdaffe840b7
"2024-12-22T23:48:17.931000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_lowercase_choicesSeparator_semicolon_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.085725
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.45, 'score': 0.45, 'score_name': 'accuracy', 'accuracy_ci_low': 0.36, 'accuracy_ci_high': 0.56, 'score_ci_low': 0.36, 'score_ci_high': 0.56}
1
a100_80gb
42a49f84b70f3a2cc053855cb7cbe000748a916c07d418da5600211840e5b23e
"2024-12-22T23:48:20.979000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_lowercase_choicesSeparator_pipe_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.060158
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.37, 'score': 0.37, 'score_name': 'accuracy', 'accuracy_ci_low': 0.28, 'accuracy_ci_high': 0.46, 'score_ci_low': 0.28, 'score_ci_high': 0.46}
1
a100_80gb
0e8a87f7ef13ccdd7732c4d244b9d06836ba068634e85fbd6be8955fb81fc540
"2024-12-22T23:48:24.048000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_lowercase_choicesSeparator_pipe_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.083962
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.45, 'score': 0.45, 'score_name': 'accuracy', 'accuracy_ci_low': 0.36, 'accuracy_ci_high': 0.55, 'score_ci_low': 0.36, 'score_ci_high': 0.55}
1
a100_80gb
0bf0aa1569bb6dcf736b2896e9cc6c77774a6f01606d05fae0b7c3e33c3ccc78
"2024-12-22T23:48:27.233000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_lowercase_choicesSeparator_OrCapital_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.194401
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.29, 'score': 0.29, 'score_name': 'accuracy', 'accuracy_ci_low': 0.21, 'accuracy_ci_high': 0.38, 'score_ci_low': 0.21, 'score_ci_high': 0.38}
1
a100_80gb
fb403d0d66fce7deed84af0954098834bebdeda57c0c9e4d0ab6c7180b843821
"2024-12-22T23:48:30.321000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_lowercase_choicesSeparator_OrCapital_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.106329
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.39, 'score': 0.39, 'score_name': 'accuracy', 'accuracy_ci_low': 0.2934067372803742, 'accuracy_ci_high': 0.49, 'score_ci_low': 0.2934067372803742, 'score_ci_high': 0.49}
1
a100_80gb
2b670bc28a2e1206b4bfac0bf8bc2f301987a148202f55a66e22c5dd0a3938e5
"2024-12-22T23:48:35.011000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_lowercase_choicesSeparator_orLower_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.706397
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.34, 'score': 0.34, 'score_name': 'accuracy', 'accuracy_ci_low': 0.26, 'accuracy_ci_high': 0.44, 'score_ci_low': 0.26, 'score_ci_high': 0.44}
1
a100_80gb
7d9a4aa06ce5ccc6fd454c37e96df4fbdf6ab5eb46a5fb381ff736eeaf6a85de
"2024-12-22T23:48:40.282000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_lowercase_choicesSeparator_orLower_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
4.252146
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.36, 'score': 0.36, 'score_name': 'accuracy', 'accuracy_ci_low': 0.27, 'accuracy_ci_high': 0.47, 'score_ci_low': 0.27, 'score_ci_high': 0.47}
1
a100_80gb
072e55b2d6c241baa92cff4d568cb0d75024156e5b60315618863b54ebdbd543
"2024-12-22T23:48:43.445000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_numbers_choicesSeparator_space_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.172739
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.38, 'score': 0.38, 'score_name': 'accuracy', 'accuracy_ci_low': 0.29, 'accuracy_ci_high': 0.48, 'score_ci_low': 0.29, 'score_ci_high': 0.48}
1
a100_80gb
6435fcb651385af3df4b21d71ec3d0bfb47f8516be76852fce415d6112994b83
"2024-12-22T23:48:46.641000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_numbers_choicesSeparator_space_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.203881
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.4, 'score': 0.4, 'score_name': 'accuracy', 'accuracy_ci_low': 0.3, 'accuracy_ci_high': 0.5, 'score_ci_low': 0.3, 'score_ci_high': 0.5}
1
a100_80gb
45ae0ae9a15e51aba29e9cdee10ffce93d5f7e2f6d08bc934f5b821bf5b20c01
"2024-12-22T23:47:34.922000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_capitals_choicesSeparator_OrCapital_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.015188
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.22, 'score': 0.22, 'score_name': 'accuracy', 'accuracy_ci_low': 0.14, 'accuracy_ci_high': 0.31, 'score_ci_low': 0.14, 'score_ci_high': 0.31}
1
a100_80gb
282337bc58e92882168422b7d4e5c2f048697f4bdbc8e1f44c269b9d195d553d
"2024-12-22T23:47:38.190000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_capitals_choicesSeparator_OrCapital_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.21906
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.3, 'score': 0.3, 'score_name': 'accuracy', 'accuracy_ci_low': 0.22, 'accuracy_ci_high': 0.4, 'score_ci_low': 0.22, 'score_ci_high': 0.4}
1
a100_80gb
004b99f092a7ef93665fb9a99fcb2fc86b28809e75abceaa4462d8588e10214e
"2024-12-22T23:47:43.036000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_capitals_choicesSeparator_orLower_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.114097
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.28, 'score': 0.28, 'score_name': 'accuracy', 'accuracy_ci_low': 0.19, 'accuracy_ci_high': 0.37, 'score_ci_low': 0.19, 'score_ci_high': 0.37}
1
a100_80gb
06ab7b8c914c184e6482850c9cecdb506fb9b82a3ff20dada4cf64ef24360cc7
"2024-12-22T23:47:47.505000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_capitals_choicesSeparator_orLower_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.058259
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.31, 'score': 0.31, 'score_name': 'accuracy', 'accuracy_ci_low': 0.23, 'accuracy_ci_high': 0.41, 'score_ci_low': 0.23, 'score_ci_high': 0.41}
1
a100_80gb
bc36361d983028527baebbfabbf1c5507186bee4462c0b19555e05bd2ca7c231
"2024-12-22T23:47:50.486000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_lowercase_choicesSeparator_space_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.010846
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.25, 'score': 0.25, 'score_name': 'accuracy', 'accuracy_ci_low': 0.17, 'accuracy_ci_high': 0.34, 'score_ci_low': 0.17, 'score_ci_high': 0.34}
1
a100_80gb
54b9767e4bea0e7383e27c4b0ab1f8751282fd07ead95dbc8f7b68c580326c10
"2024-12-22T23:47:54.156000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_lowercase_choicesSeparator_space_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.704553
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.28, 'score': 0.28, 'score_name': 'accuracy', 'accuracy_ci_low': 0.2, 'accuracy_ci_high': 0.37, 'score_ci_low': 0.2, 'score_ci_high': 0.37}
1
a100_80gb
9f138bac2b41cac4decefb8079521aa2f5de6c2f69e479d78237522b5913b4e1
"2024-12-22T23:47:58.348000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_lowercase_choicesSeparator_newline_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.213005
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.43, 'score': 0.43, 'score_name': 'accuracy', 'accuracy_ci_low': 0.34, 'accuracy_ci_high': 0.53, 'score_ci_low': 0.34, 'score_ci_high': 0.53}
1
a100_80gb
ed8d1bb2c910b954ce804a88d93a683235977d6ca22445f9f2a90bd977a501dd
"2024-12-22T23:48:01.457000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_lowercase_choicesSeparator_newline_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.122167
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.43, 'score': 0.43, 'score_name': 'accuracy', 'accuracy_ci_low': 0.33, 'accuracy_ci_high': 0.54, 'score_ci_low': 0.33, 'score_ci_high': 0.54}
1
a100_80gb
1f72bbf684c621ae62503df36a33438a55f50c85e1c35e2c14e7e2bb498b1225
"2024-12-22T23:48:04.580000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_lowercase_choicesSeparator_comma_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.145734
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.4, 'score': 0.4, 'score_name': 'accuracy', 'accuracy_ci_low': 0.32, 'accuracy_ci_high': 0.5, 'score_ci_low': 0.32, 'score_ci_high': 0.5}
1
a100_80gb
986c1ec9f4e5055dc37041b53b92f4d3ecc72fca81030831fbb2d6785bc2d130
"2024-12-22T23:48:09.462000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_lowercase_choicesSeparator_comma_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.891199
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.42, 'score': 0.42, 'score_name': 'accuracy', 'accuracy_ci_low': 0.31, 'accuracy_ci_high': 0.52, 'score_ci_low': 0.31, 'score_ci_high': 0.52}
1
a100_80gb
412047b2868254c18f77def68089b4b4478025af43310792c6a0bd2fe5a734dd
"2024-12-22T23:46:59.655000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_capitals_choicesSeparator_space_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.016202
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.19, 'score': 0.19, 'score_name': 'accuracy', 'accuracy_ci_low': 0.12, 'accuracy_ci_high': 0.27, 'score_ci_low': 0.12, 'score_ci_high': 0.27}
1
a100_80gb
f72d155ce25500ce493cb0478ed65d62b6975720dfc5525a331f0d0f96f42da0
"2024-12-22T23:47:02.667000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_capitals_choicesSeparator_space_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.04635
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.28, 'score': 0.28, 'score_name': 'accuracy', 'accuracy_ci_low': 0.2, 'accuracy_ci_high': 0.3726521496509156, 'score_ci_low': 0.2, 'score_ci_high': 0.3726521496509156}
1
a100_80gb
2dfc1ff000f2791a8355076470cfd6e96723bfda6142355ba6ca61bd320ab014
"2024-12-22T23:47:07.007000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_capitals_choicesSeparator_newline_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.358681
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.31, 'score': 0.31, 'score_name': 'accuracy', 'accuracy_ci_low': 0.23, 'accuracy_ci_high': 0.41, 'score_ci_low': 0.23, 'score_ci_high': 0.41}
1
a100_80gb
3fda0a84ed24750c8b84eebd82e4470e54370aaa616724cf35ecd3c2269b1e89
"2024-12-22T23:47:10.198000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_capitals_choicesSeparator_newline_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.21486
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.4, 'score': 0.4, 'score_name': 'accuracy', 'accuracy_ci_low': 0.30180775878464533, 'accuracy_ci_high': 0.5, 'score_ci_low': 0.30180775878464533, 'score_ci_high': 0.5}
1
a100_80gb
ebdd34b60b2134ac63ff23373167a4bd7dfcb6571925732bbe98683129433898
"2024-12-22T23:47:13.219000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_capitals_choicesSeparator_comma_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.038143
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.24, 'score': 0.24, 'score_name': 'accuracy', 'accuracy_ci_low': 0.17, 'accuracy_ci_high': 0.34, 'score_ci_low': 0.17, 'score_ci_high': 0.34}
1
a100_80gb
91d157627899293c9b2b0353bb74bd475d0e809c31aec79f088c70e2a3f7a4b7
"2024-12-22T23:47:16.932000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_capitals_choicesSeparator_comma_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.733212
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.35, 'score': 0.35, 'score_name': 'accuracy', 'accuracy_ci_low': 0.26, 'accuracy_ci_high': 0.45, 'score_ci_low': 0.26, 'score_ci_high': 0.45}
1
a100_80gb
01551e5b7c1e7f5b66e714b63fb0976f232fc7a74879598a313a937e88a2263d
"2024-12-22T23:47:21.251000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_capitals_choicesSeparator_semicolon_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.340098
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.28, 'score': 0.28, 'score_name': 'accuracy', 'accuracy_ci_low': 0.2, 'accuracy_ci_high': 0.38, 'score_ci_low': 0.2, 'score_ci_high': 0.38}
1
a100_80gb
cc61ef55fe15a174e075d5e47694f7100afec6ba356f587bb2f99be1895df85c
"2024-12-22T23:47:25.946000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_capitals_choicesSeparator_semicolon_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.054962
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.38, 'score': 0.38, 'score_name': 'accuracy', 'accuracy_ci_low': 0.29, 'accuracy_ci_high': 0.48, 'score_ci_low': 0.29, 'score_ci_high': 0.48}
1
a100_80gb
4cd789ca8fdfc1f6110e6e6243828e7308946e90a24139f855a819ffa9fcfb8f
"2024-12-22T23:47:28.955000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_capitals_choicesSeparator_pipe_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.03671
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.31, 'score': 0.31, 'score_name': 'accuracy', 'accuracy_ci_low': 0.22847916254879963, 'accuracy_ci_high': 0.41, 'score_ci_low': 0.22847916254879963, 'score_ci_high': 0.41}
1
a100_80gb
2ac872e69ac95c49c8a94602b2975eb5eb5136ee22669644d4bea1fac29479cc
"2024-12-22T23:47:31.931000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_capitals_choicesSeparator_pipe_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.004322
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.41, 'score': 0.41, 'score_name': 'accuracy', 'accuracy_ci_low': 0.32, 'accuracy_ci_high': 0.52, 'score_ci_low': 0.32, 'score_ci_high': 0.52}
1
a100_80gb
dcc3bf1d6c8b3507349aedb2766b7033ca622285e25a156bfa11e5fa79ed8560
"2024-12-22T23:46:23.154000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_greek_choicesSeparator_comma_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
4.169602
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.24, 'score': 0.24, 'score_name': 'accuracy', 'accuracy_ci_low': 0.17, 'accuracy_ci_high': 0.3485957092557362, 'score_ci_low': 0.17, 'score_ci_high': 0.3485957092557362}
1
a100_80gb
58234f98e76391ae61d2f29930ac7a6b794d84f252a4f0d5b24f749388037afb
"2024-12-22T23:46:26.282000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_greek_choicesSeparator_comma_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.121241
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.27, 'score': 0.27, 'score_name': 'accuracy', 'accuracy_ci_low': 0.19, 'accuracy_ci_high': 0.36, 'score_ci_low': 0.19, 'score_ci_high': 0.36}
1
a100_80gb
1814f43169dcb7e862e00f8dfc88bf08f9ba4a84811cdc49062fd10d2353e0e9
"2024-12-22T23:46:30.525000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_greek_choicesSeparator_semicolon_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.241888
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.27, 'score': 0.27, 'score_name': 'accuracy', 'accuracy_ci_low': 0.19, 'accuracy_ci_high': 0.37, 'score_ci_low': 0.19, 'score_ci_high': 0.37}
1
a100_80gb
ca41fc9fa9be02a8ca7550f368ab736e43ac569dd4df0e31031078eaaca54d75
"2024-12-22T23:46:33.631000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_greek_choicesSeparator_semicolon_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.096562
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.29, 'score': 0.29, 'score_name': 'accuracy', 'accuracy_ci_low': 0.21, 'accuracy_ci_high': 0.39, 'score_ci_low': 0.21, 'score_ci_high': 0.39}
1
a100_80gb
92c44f46dcb68696f7988b4a9fe14b0404c91b7b54493fab98df502df6b81c4a
"2024-12-22T23:46:36.850000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_greek_choicesSeparator_pipe_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.147617
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.25, 'score': 0.25, 'score_name': 'accuracy', 'accuracy_ci_low': 0.17712379451679425, 'accuracy_ci_high': 0.35, 'score_ci_low': 0.17712379451679425, 'score_ci_high': 0.35}
1
a100_80gb
2e8c976c81827274d706e91eba14a18699182978f1c974ba0f47c77f21c6634a
"2024-12-22T23:46:39.963000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_greek_choicesSeparator_pipe_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.102285
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.33, 'score': 0.33, 'score_name': 'accuracy', 'accuracy_ci_low': 0.24007849969624384, 'accuracy_ci_high': 0.43, 'score_ci_low': 0.24007849969624384, 'score_ci_high': 0.43}
1
a100_80gb
5a5aace7b82bbcf540a2e6d07a2cc6e013758dd02aa4dda3b8bb2ce57ba06317
"2024-12-22T23:46:44.900000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_greek_choicesSeparator_OrCapital_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.929016
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.27, 'score': 0.27, 'score_name': 'accuracy', 'accuracy_ci_low': 0.19, 'accuracy_ci_high': 0.36, 'score_ci_low': 0.19, 'score_ci_high': 0.36}
1
a100_80gb
72640e868ccb7e6638b6a5e52f746d7083cd15825203384f93ae59f939b4e815
"2024-12-22T23:46:48.159000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_greek_choicesSeparator_OrCapital_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.226835
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.3, 'score': 0.3, 'score_name': 'accuracy', 'accuracy_ci_low': 0.22, 'accuracy_ci_high': 0.4, 'score_ci_low': 0.22, 'score_ci_high': 0.4}
1
a100_80gb
472431128eb19fe84a54a1630cdaca15ebc304cd514ab366d1ba2465efa903ca
"2024-12-22T23:46:53.495000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_greek_choicesSeparator_orLower_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
4.274025
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.22, 'score': 0.22, 'score_name': 'accuracy', 'accuracy_ci_low': 0.15, 'accuracy_ci_high': 0.31, 'score_ci_low': 0.15, 'score_ci_high': 0.31}
1
a100_80gb
7830ad0afa87a012eeabb7d9b984f1df0e7b0bc100c5e9f3a7bde5aa79dbd193
"2024-12-22T23:46:56.665000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_greek_choicesSeparator_orLower_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.151465
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.25, 'score': 0.25, 'score_name': 'accuracy', 'accuracy_ci_low': 0.17, 'accuracy_ci_high': 0.35, 'score_ci_low': 0.17, 'score_ci_high': 0.35}
1
a100_80gb
7d9e4f74bea1fed952d227c594c2687725b5a94fa4cca521052bc5ad150b972a
"2024-12-22T23:45:41.993000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_keyboard_choicesSeparator_pipe_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.520374
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.35, 'score': 0.35, 'score_name': 'accuracy', 'accuracy_ci_low': 0.26, 'accuracy_ci_high': 0.45, 'score_ci_low': 0.26, 'score_ci_high': 0.45}
1
a100_80gb
6f3c08b7d8fc067e454ec52c46f090204b725d41614e6cc7cb7603323955ece2
"2024-12-22T23:45:46.734000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_keyboard_choicesSeparator_pipe_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.739105
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.25, 'score': 0.25, 'score_name': 'accuracy', 'accuracy_ci_low': 0.17, 'accuracy_ci_high': 0.34, 'score_ci_low': 0.17, 'score_ci_high': 0.34}
1
a100_80gb
7455b84a0230154b0d8cbfaa110f8c1274fba76a71e38d5c24cb69588414face
"2024-12-22T23:45:50.010000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_keyboard_choicesSeparator_OrCapital_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.269456
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.32, 'score': 0.32, 'score_name': 'accuracy', 'accuracy_ci_low': 0.23, 'accuracy_ci_high': 0.42, 'score_ci_low': 0.23, 'score_ci_high': 0.42}
1
a100_80gb
e433bc72578b74fa5b37802efd6615a97e5ce55d7d8266ae6f56a81717eedbe1
"2024-12-22T23:45:53.135000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_keyboard_choicesSeparator_OrCapital_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.128859
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.29, 'score': 0.29, 'score_name': 'accuracy', 'accuracy_ci_low': 0.2, 'accuracy_ci_high': 0.39, 'score_ci_low': 0.2, 'score_ci_high': 0.39}
1
a100_80gb
7b776f775b37d706e095afe3bfb5f993eeabb00d6211614868152c01dc50d553
"2024-12-22T23:45:57.972000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.human_aging,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_keyboard_choicesSeparator_orLower_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.835338
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.32, 'score': 0.32, 'score_name': 'accuracy', 'accuracy_ci_low': 0.24, 'accuracy_ci_high': 0.42, 'score_ci_low': 0.24, 'score_ci_high': 0.42}
1
a100_80gb