model
stringlengths
4
89
revision
stringclasses
1 value
model_sha
stringlengths
0
40
results
dict
commit
stringlengths
40
40
date
unknown
score
float64
21.8
83
TheBloke/Kimiko-v2-13B-fp16
main
0fed305667508e50330e71a2d43e9cee5ea73783
{ "arc:challenge": 61, "hellaswag": 83.3, "hendrycksTest": 55.2, "truthfulqa:mc": 40.7 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
60
TheBloke/wizard-vicuna-13B-GPTQ
main
936a51c0219744d7a9598d0c65a7d18e01660601
{ "arc:challenge": 28.7, "hellaswag": 25.9, "hendrycksTest": 25.8, "truthfulqa:mc": 48.5 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
32.2
TheBloke/CodeLlama-34B-Python-fp16
main
875f9d97fb6c9619d8867887dd1d80918ff0f593
{ "arc:challenge": 38.1, "hellaswag": 34.8, "hendrycksTest": 32.9, "truthfulqa:mc": 43.6 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
37.4
TheBloke/manticore-13b-chat-pyg-GPTQ
main
923f27245d13058c9c1b3ab0eab6c6c93ffc162e
{ "arc:challenge": 57.8, "hellaswag": 81.1, "hendrycksTest": 47.6, "truthfulqa:mc": 47.8 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
58.6
TheBloke/Airoboros-L2-70B-2.1-GPTQ
main
23ed580cb77ebaee49ea11eb4538fd3ab3795b76
{ "arc:challenge": 70.4, "hellaswag": 86.5, "hendrycksTest": 68.9, "truthfulqa:mc": 55.5 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
70.3
TheBloke/Airoboros-L2-13B-2.1-GPTQ
main
d90d96e40b9359cb5c35e6b6c8f0eb24896e827b
{ "arc:challenge": 59, "hellaswag": 81.7, "hendrycksTest": 53.2, "truthfulqa:mc": 44.7 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
59.6
TheBloke/gpt4-x-vicuna-13B-HF
main
a247577c882940e0c6b040fe8239d760c0d10d40
{ "arc:challenge": 53.4, "hellaswag": 80.1, "hendrycksTest": 51.2, "truthfulqa:mc": 53.6 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
59.6
TheBloke/Wizard-Vicuna-30B-Superhot-8K-fp16
main
062fe5409861d7386279fb534b435be39c88ceaf
{ "arc:challenge": 26.2, "hellaswag": 33, "hendrycksTest": 23.5, "truthfulqa:mc": 47.5 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
32.6
TheBloke/Wizard-Vicuna-7B-Uncensored-HF
main
b802f1b4401d0b2242137160c20cc11b9ffd3a4c
{ "arc:challenge": 53.4, "hellaswag": 78.8, "hendrycksTest": 37.1, "truthfulqa:mc": 43.5 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
53.2
TheBloke/Wizard-Vicuna-13B-Uncensored-HF
main
fff9ac7f0e2e7b340f2301f5f089d989fc03be67
{ "arc:challenge": 59, "hellaswag": 81.9, "hendrycksTest": 47.9, "truthfulqa:mc": 51.7 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
60.1
TheBloke/Wizard-Vicuna-30B-Uncensored-GPTQ
main
56a82ece7a9309189561a590e8f4d2fe0d4be92b
{ "arc:challenge": 61.1, "hellaswag": 82.4, "hendrycksTest": 56.5, "truthfulqa:mc": 49.9 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
62.5
TheBloke/orca_mini_v3_7B-GPTQ
main
06ddd48cd904907e3c73d2dfe47d28626053598b
{ "arc:challenge": 30.1, "hellaswag": 26, "hendrycksTest": 24.3, "truthfulqa:mc": 48.4 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
32.2
TheBloke/OpenOrca-Platypus2-13B-GPTQ
main
0fa9a56066656fbc94e3ec088bc900fd1d4d38e8
{ "arc:challenge": 62.5, "hellaswag": 82.7, "hendrycksTest": 58.6, "truthfulqa:mc": 51.9 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
63.9
TheBloke/guanaco-33B-GPTQ
main
8e42e031bfc8be3bbf31dc546d7c51fb991ff6e0
{ "arc:challenge": 28.2, "hellaswag": 26.3, "hendrycksTest": 24.9, "truthfulqa:mc": 49 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
32.1
TheBloke/robin-33B-v2-GPTQ
main
4c2588d65302e9ca634548ed81e8650fb2975686
{ "arc:challenge": 27.7, "hellaswag": 26.3, "hendrycksTest": 23.5, "truthfulqa:mc": 49.5 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
31.8
TheBloke/WizardLM-33B-V1.0-Uncensored-GPTQ
main
1c65902c620fcdf6b9c8e36ce17f21360e186a1e
{ "arc:challenge": 27.4, "hellaswag": 26, "hendrycksTest": 25.8, "truthfulqa:mc": 48.9 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
32
TheBloke/wizard-vicuna-13B-HF
main
12dc8aacb474522ae2a83c18cb0fdf0907987f8f
{ "arc:challenge": 54.7, "hellaswag": 79.2, "hendrycksTest": 48.9, "truthfulqa:mc": 49.6 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
58.1
TheBloke/vicuna-13B-1.1-HF
main
8c71dbe9221e83d2ec72e4dc08beccfc78b563c0
{ "arc:challenge": 52.7, "hellaswag": 80.1, "hendrycksTest": 51.9, "truthfulqa:mc": 52.1 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
59.2
TheBloke/WizardLM-13B-V1-1-SuperHOT-8K-GPTQ
main
085eb5cd394f30d72bf5efcf83a580e87264b3e8
{ "arc:challenge": 57, "hellaswag": 80.3, "hendrycksTest": 47.1, "truthfulqa:mc": 53.5 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
59.5
TheBloke/Chinese-Alpaca-33B-SuperHOT-8K-fp16
main
a55ce761bace8be6d17c357c57ef927751afd40c
{ "arc:challenge": 26.8, "hellaswag": 29.6, "hendrycksTest": 24.1, "truthfulqa:mc": 47.7 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
32
TheBloke/gpt4-alpaca-lora-30b-HF
main
3c8007467a081dc72ae09b9d358416b056b38920
{ "arc:challenge": 64.8, "hellaswag": 85.7, "hendrycksTest": 58.5, "truthfulqa:mc": 52.2 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
65.3
TheBloke/tulu-30B-fp16
main
37c3655676c37662f60c68dacfce3f0e861be846
{ "arc:challenge": 60, "hellaswag": 83.4, "hendrycksTest": 56.1, "truthfulqa:mc": 45.1 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
61.2
TheBloke/airoboros-33B-gpt4-1-4-SuperHOT-8K-fp16
main
53fdac1cdb8a37647e5dbe4199bc3fb70e617fce
{ "arc:challenge": 26, "hellaswag": 30.7, "hendrycksTest": 23.6, "truthfulqa:mc": 47.9 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
32.1
TheBloke/EverythingLM-13B-16K-GPTQ
main
f14d3df05577f3e1ac35e2c4ec32ce0d39b97508
{ "arc:challenge": 29.3, "hellaswag": 26.2, "hendrycksTest": 25.4, "truthfulqa:mc": 48.6 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
32.4
TheBloke/Llama-2-7B-GPTQ
main
ecd7ab9f6adc36ecbe0d751eeea0d90ae1863c3b
{ "arc:challenge": 52, "hellaswag": 77.6, "hendrycksTest": 44, "truthfulqa:mc": 39.3 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
53.2
TheBloke/Kimiko-13B-fp16
main
27868769e2d6b1af46337f0997c71b0577952a3d
{ "arc:challenge": 59.2, "hellaswag": 82.4, "hendrycksTest": 55.8, "truthfulqa:mc": 39.6 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
59.3
TheBloke/CodeLlama-34B-Instruct-fp16
main
a4d0ce949de4d5b5f74691641efb5b70736a32a8
{ "arc:challenge": 40.8, "hellaswag": 35.7, "hendrycksTest": 39.7, "truthfulqa:mc": 44.3 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
40.1
TheBloke/WizardLM-70B-V1.0-GPTQ
main
c234d7c9c0fd26efb55757fdbfb604d549539fe0
{ "arc:challenge": 63.8, "hellaswag": 83.8, "hendrycksTest": 63.7, "truthfulqa:mc": 54.5 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
66.4
TheBloke/CodeLlama-13B-Python-fp16
main
442282f4207442b828953a72c51a919c332cba5c
{ "arc:challenge": 33.2, "hellaswag": 44.5, "hendrycksTest": 25.9, "truthfulqa:mc": 44 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
36.9
TheBloke/Manticore-13B-Chat-Pyg-Guanaco-SuperHOT-8K-GPTQ
main
bd3c66e626c81de4977f197e1534bd3dfa2f569d
{ "arc:challenge": 52.8, "hellaswag": 79.6, "hendrycksTest": 39.8, "truthfulqa:mc": 52.5 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
56.2
TheBloke/medalpaca-13B-GPTQ-4bit
main
12190f743a19e91dfe1f5c77abc0c1bf486073dd
{ "arc:challenge": 29.4, "hellaswag": 26.3, "hendrycksTest": 25.4, "truthfulqa:mc": 49.5 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
32.6
TheBloke/OpenOrcaxOpenChat-Preview2-13B-GPTQ
main
ec9eb4f471b5bb6a7e5e505369628586c0c72252
{ "arc:challenge": 61.3, "hellaswag": 82.1, "hendrycksTest": 57.8, "truthfulqa:mc": 50.2 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
62.8
TheBloke/llama-2-70b-Guanaco-QLoRA-fp16
main
54b0e39d5e9aee7b323f50b0a26db15295c3d5c9
{ "arc:challenge": 68.3, "hellaswag": 88.3, "hendrycksTest": 70.2, "truthfulqa:mc": 55.7 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
70.6
TheBloke/tulu-13B-fp16
main
532aeb363b0ceee155b3cf9479ef635b797cee7c
{ "arc:challenge": 53.9, "hellaswag": 80.7, "hendrycksTest": 53.2, "truthfulqa:mc": 43.8 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
57.9
TheBloke/Lemur-70B-Chat-v1-GPTQ
main
12499165df1785f50df3e95940406032776401ea
{ "arc:challenge": 65.3, "hellaswag": 84.4, "hendrycksTest": 64.7, "truthfulqa:mc": 57.1 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
67.9
TheBloke/wizardLM-7B-HF
main
a8e22531a48cece989e670f539eb18ebd2dbd0cf
{ "arc:challenge": 50.3, "hellaswag": 75.3, "hendrycksTest": 38.1, "truthfulqa:mc": 45.6 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
52.3
TheBloke/Llama-2-13B-GPTQ
main
b7db471d1789802a3a8e3b93cdd66a9f046f17c3
{ "arc:challenge": 57.3, "hellaswag": 81.6, "hendrycksTest": 54.8, "truthfulqa:mc": 36.6 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
57.6
TheBloke/dromedary-65b-lora-HF
main
3fa4546259d6bbd6b5d637484c325ab19181a73c
{ "arc:challenge": 61.6, "hellaswag": 82.5, "hendrycksTest": 63.1, "truthfulqa:mc": 38.8 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
61.5
TheBloke/landmark-attention-llama7b-fp16
main
bf8bdcb0c30cceb0ceda33cf5fde683807e39a58
{ "arc:challenge": 47.4, "hellaswag": 65.8, "hendrycksTest": 31.6, "truthfulqa:mc": 42.6 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
46.8
TheBloke/stable-vicuna-13B-HF
main
2b099b2be0dafb2606ae9808c0f6183fe4bff7bc
{ "arc:challenge": 53.3, "hellaswag": 78.5, "hendrycksTest": 50.3, "truthfulqa:mc": 48.4 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
57.6
TheBloke/WizardLM-13B-V1.1-GPTQ
main
9df807ac64034bc6e7387326689d6e39656ce5e0
{ "arc:challenge": 58.5, "hellaswag": 80.7, "hendrycksTest": 49.6, "truthfulqa:mc": 54.4 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
60.8
TheBloke/robin-13B-v2-fp16
main
f4dd8fc4440ed84fcf3ff1122f2b7f6024cca29d
{ "arc:challenge": 56.5, "hellaswag": 80.4, "hendrycksTest": 48.8, "truthfulqa:mc": 50.6 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
59.1
TheBloke/VicUnlocked-alpaca-65B-QLoRA-fp16
main
6cdacfda96970aa144e316b108ab9bc17c99a573
{ "arc:challenge": 65.6, "hellaswag": 85.2, "hendrycksTest": 63.1, "truthfulqa:mc": 52.5 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
66.6
TheBloke/koala-7B-HF
main
d102fe3b68f1a5a50d547e4fd1c8b33b783c993b
{ "arc:challenge": 47.1, "hellaswag": 73.6, "hendrycksTest": 25.5, "truthfulqa:mc": 46 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
48
TheBloke/Nous-Hermes-13B-SuperHOT-8K-fp16
main
b407c1ece029ad5693d38e6e0931e9482962ed15
{ "arc:challenge": 55.3, "hellaswag": 81.9, "hendrycksTest": 48.2, "truthfulqa:mc": 51.2 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
59.1
TheBloke/tulu-7B-fp16
main
8a026683f79119643f4007da4e9155c7849792cc
{ "arc:challenge": 50.2, "hellaswag": 77, "hendrycksTest": 47.6, "truthfulqa:mc": 41.6 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
54.1
TheBloke/Vicuna-33B-1-3-SuperHOT-8K-fp16
main
0b6484697d5cca5baa534b882dcad8101add8cda
{ "arc:challenge": 25.4, "hellaswag": 34.6, "hendrycksTest": 23.6, "truthfulqa:mc": 46.9 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
32.6
TheBloke/wizardLM-13B-1.0-fp16
main
b79733805e98e668ff9a459975c259881b1b8014
{ "arc:challenge": 57.3, "hellaswag": 80.9, "hendrycksTest": 52.9, "truthfulqa:mc": 50.5 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
60.4
TheBloke/Genz-70b-GPTQ
main
7d38987a43d2445b193db99a029a264b39dc6c8e
{ "arc:challenge": 71.1, "hellaswag": 87.6, "hendrycksTest": 70.3, "truthfulqa:mc": 62.3 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
72.8
TheBloke/Project-Baize-v2-7B-GPTQ
main
5dc039834e1ea42ac334458b2e3090fe3705cc59
{ "arc:challenge": 46, "hellaswag": 73.4, "hendrycksTest": 35.5, "truthfulqa:mc": 39.9 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
48.7
TheBloke/airoboros-13B-HF
main
9219b61a0e8bc880e4cd0f8bebc48a97ee0950c7
{ "arc:challenge": 58.3, "hellaswag": 81, "hendrycksTest": 50, "truthfulqa:mc": 51.6 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
60.2
TheBloke/Platypus2-70B-Instruct-GPTQ
main
4a44568aadd8a4babfa5549cf33e6e84cbae7ab8
{ "arc:challenge": 71.2, "hellaswag": 87.6, "hendrycksTest": 69.9, "truthfulqa:mc": 62.5 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
72.8
TheBloke/guanaco-7B-HF
main
293c24105fa15afa127a2ec3905fdc2a0a3a6dac
{ "arc:challenge": 53, "hellaswag": 80.1, "hendrycksTest": 35.3, "truthfulqa:mc": 39.2 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
51.9
TheBloke/koala-13B-HF
main
b20f96a0171ce4c0fa27d6048215ebe710521587
{ "arc:challenge": 53, "hellaswag": 77.6, "hendrycksTest": 45.3, "truthfulqa:mc": 50.2 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
56.5
chaoyi-wu/MedLLaMA_13B
main
893557ef32f98cd01deb1c5d063be6d640ffa657
{ "arc:challenge": 54.3, "hellaswag": 78.5, "hendrycksTest": 46.4, "truthfulqa:mc": 40.5 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
54.9
Voicelab/trurl-2-13b
main
c8b2bbc7a570a9ea67928674695a4e7dff017d66
{ "arc:challenge": 60.1, "hellaswag": 80.2, "hendrycksTest": 78.6, "truthfulqa:mc": 45.9 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
66.2
Voicelab/trurl-2-7b
main
e26ca5f157c60fc527170cc04db7fc0ea04ad26f
{ "arc:challenge": 53.4, "hellaswag": 75.3, "hendrycksTest": 50, "truthfulqa:mc": 45.4 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
56
THUDM/chatglm2-6b
main
162b620e3078b03eefff94eb5f762d4093425fb5
{ "arc:challenge": 38.8, "hellaswag": 59, "hendrycksTest": 46.7, "truthfulqa:mc": 48.1 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
48.2
Brillibits/Instruct_Llama70B_Dolly15k
main
45444ac60488594e0700e6c7313ff444b4468240
{ "arc:challenge": 68.3, "hellaswag": 87.2, "hendrycksTest": 69.5, "truthfulqa:mc": 46.5 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
67.9
joehuangx/spatial-vicuna-7b-v1.5-LoRA
main
dc71924cfb214b91461d35178e6ea6fef7946f13
{ "arc:challenge": 50.8, "hellaswag": 74.6, "hendrycksTest": 48.1, "truthfulqa:mc": 49.4 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
55.7
openthaigpt/openthaigpt-1.0.0-alpha-7b-chat-ckpt-hf
main
cdffb3488c5cb1a9aa5039a6b3bc72af24827db0
{ "arc:challenge": 50.9, "hellaswag": 74.9, "hendrycksTest": 40, "truthfulqa:mc": 47.2 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
53.2
breadlicker45/dough-instruct-base-001
main
3e1b0bf0a887feeb342982eee4f6d8041772a7dd
{ "arc:challenge": 23.9, "hellaswag": 24.8, "hendrycksTest": 23.1, "truthfulqa:mc": 53.4 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
31.3
Tincando/fiction_story_generator
main
377b080cf96e10d50289aa3e1fd79c330265f45a
{ "arc:challenge": 23.3, "hellaswag": 28.7, "hendrycksTest": 26.7, "truthfulqa:mc": 43.8 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
30.6
GOAT-AI/GOAT-7B-Community
main
a7073a0f5142ce04aaa1603b0812b358f62a8de8
{ "arc:challenge": 48.8, "hellaswag": 74.6, "hendrycksTest": 49.6, "truthfulqa:mc": 42.5 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
53.9
l3utterfly/open-llama-3b-v2-layla
main
465669ddafad25393ac3cfe94d3726cced112b30
{ "arc:challenge": 38.2, "hellaswag": 66.4, "hendrycksTest": 28.6, "truthfulqa:mc": 44.4 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
44.4
SaylorTwift/gpt2_test
main
ef61310a16ffda93bf8f6132e02658482ffc2bcc
{ "arc:challenge": 21.8, "hellaswag": 31.6, "hendrycksTest": 25.9, "truthfulqa:mc": 40.7 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
30
Lazycuber/L2-7b-Guanaco-Uncensored
main
9d49378c69c00113cf7f6e66d1ddb9d9b003dddc
{ "arc:challenge": 50.6, "hellaswag": 77, "hendrycksTest": 48.9, "truthfulqa:mc": 43.4 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
55
Lazycuber/pyg-instruct-wizardlm
main
f00ef7a7b0cc6f02af2a11ac764270dfd61b9e2f
{ "arc:challenge": 41, "hellaswag": 66.7, "hendrycksTest": 26.3, "truthfulqa:mc": 31.9 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
41.5
Lazycuber/Janemalion-6B
main
e72ae3ec110121115b1ae6c2e5fb3995997a2d96
{ "arc:challenge": 42.4, "hellaswag": 68.4, "hendrycksTest": 28.3, "truthfulqa:mc": 34.6 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
43.4
MrNJK/gpt2-xl-sft
main
53250831436460254b7ee9afc4014d4d3156b372
{ "arc:challenge": 30, "hellaswag": 49.2, "hendrycksTest": 25.6, "truthfulqa:mc": 38.8 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
35.9
malhajar/Platypus2-70B-instruct-4bit-gptq
main
2aa2f5646e496b3cd9b510681ba2c5081bde821f
{ "arc:challenge": 29, "hellaswag": 26, "hendrycksTest": 23.5, "truthfulqa:mc": 49.6 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
32
pillowtalks-ai/delta13b
main
83fa0860990df1db35550f973ba4306449e35412
{ "arc:challenge": 52.7, "hellaswag": 80.1, "hendrycksTest": 51.9, "truthfulqa:mc": 52.1 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
59.2
ethzanalytics/pythia-31m
main
8a3c2f1555de8a3c53d67d73b5d0d53a66a6c6c2
{ "arc:challenge": 20, "hellaswag": 26.3, "hendrycksTest": 24.3, "truthfulqa:mc": 50.1 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
30.2
Rachneet/gpt2-xl-alpaca
main
a1a19acc0ef161bfa35f460c15ed3015595714d8
{ "arc:challenge": 26.8, "hellaswag": 43.8, "hendrycksTest": 26.3, "truthfulqa:mc": 39.4 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
34.1
Panchovix/airoboros-33b-gpt4-1.2-SuperHOT-8k
main
47c14f699cbbc9bd24458edd86eb70d87552b623
{ "arc:challenge": 24.7, "hellaswag": 31.2, "hendrycksTest": 23.1, "truthfulqa:mc": 47.4 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
31.6
Panchovix/WizardLM-33B-V1.0-Uncensored-SuperHOT-8k
main
b6d0002b10d43ab48aa14e365d9e7b40655ec160
{ "arc:challenge": 25.4, "hellaswag": 32, "hendrycksTest": 23.4, "truthfulqa:mc": 47 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
32
HyperbeeAI/Tulpar-7b-v1
main
719d8e1eb4a820f01e0a92ef6220d041964bb472
{ "arc:challenge": 57, "hellaswag": 79.7, "hendrycksTest": 51.3, "truthfulqa:mc": 51.8 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
60
quantumaikr/QuantumLM
main
9058130b416355b37f5f78777748aa56d98a4da0
{ "arc:challenge": 55.8, "hellaswag": 79.7, "hendrycksTest": 54.2, "truthfulqa:mc": 46.7 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
59.1
quantumaikr/KoreanLM-hf
main
a7261e7ae6ee76c78e1ba1ac8c59bcc3e0868bf9
{ "arc:challenge": 51.5, "hellaswag": 76.8, "hendrycksTest": 40.6, "truthfulqa:mc": 44.3 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
53.3
quantumaikr/quantumairk-llama-2-70B-instruct
main
94ff2fcafd507b08e953f70806ec671ec3d17b15
{ "arc:challenge": 70.3, "hellaswag": 87.1, "hendrycksTest": 70.5, "truthfulqa:mc": 54.4 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
70.6
quantumaikr/llama-2-7b-hf-guanaco-1k
main
bdb57c5c992872ced47f48cb2177a5fa159f926a
{ "arc:challenge": 51.6, "hellaswag": 76.7, "hendrycksTest": 47.4, "truthfulqa:mc": 44.8 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
55.1
quantumaikr/QuantumLM-70B-hf
main
e13dd23ae5e611e959b6c8d5bc47bf4fd37cd9d7
{ "arc:challenge": 59.5, "hellaswag": 83, "hendrycksTest": 62.3, "truthfulqa:mc": 53.4 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
64.6
quantumaikr/llama-2-70b-fb16-guanaco-1k
main
c317af1b593a4f91b0e79c7142ca75f1e8d65278
{ "arc:challenge": 70.5, "hellaswag": 87.3, "hendrycksTest": 70.2, "truthfulqa:mc": 57.6 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
71.4
quantumaikr/llama-2-70b-fb16-korean
main
fd57855006c15c4121feccab1cbeee8107de5b5a
{ "arc:challenge": 67.2, "hellaswag": 86.8, "hendrycksTest": 69.3, "truthfulqa:mc": 56.5 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
70
quantumaikr/open_llama_7b_hf
main
41441cea58f963cfc4827da12ae5759e943151cb
{ "arc:challenge": 26.5, "hellaswag": 26.9, "hendrycksTest": 26.5, "truthfulqa:mc": 49.5 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
32.4
quantumaikr/llama-2-70B-chat
main
d242fdbf800e388e6ee456578064cab5e057f987
{ "arc:challenge": 67.6, "hellaswag": 86.9, "hendrycksTest": 69.2, "truthfulqa:mc": 57.3 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
70.2
quantumaikr/llama-2-70b-fb16-orca-chat-10k
main
697aaeb8eb9905c9b25bebb736d1905444c774a6
{ "arc:challenge": 68.1, "hellaswag": 87.1, "hendrycksTest": 69.2, "truthfulqa:mc": 61.6 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
71.5
quantumaikr/QuantumLM-llama2-70B-Korean-LoRA
main
ea21456e999f6ce35da1cd88b8f62bb5770b985a
{ "arc:challenge": 70.6, "hellaswag": 86.4, "hendrycksTest": 69.4, "truthfulqa:mc": 56.1 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
70.6
quantumaikr/QuantumLM-7B
main
f44998432fb90d88094ddf42e57ec458877a197f
{ "arc:challenge": 50.3, "hellaswag": 76.1, "hendrycksTest": 45.3, "truthfulqa:mc": 46.3 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
54.5
huggingtweets/jerma985
main
816206ad02a397161be78dcb70eeda67e0c53132
{ "arc:challenge": 21.7, "hellaswag": 30.9, "hendrycksTest": 26.6, "truthfulqa:mc": 44 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
30.8
huggingtweets/gladosystem
main
02a1bbcee7b584ace743b2fe4885cc0eaf2179ac
{ "arc:challenge": 24.4, "hellaswag": 29.7, "hendrycksTest": 23.2, "truthfulqa:mc": 41.8 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
29.8
oh-yeontaek/llama-2-7B-LoRA-assemble
main
72e866a96a2e9afc6527c8d757c69088c3a069c8
{ "arc:challenge": 57.3, "hellaswag": 78.8, "hendrycksTest": 50.8, "truthfulqa:mc": 53.2 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
60
oh-yeontaek/llama-2-70B-LoRA-assemble-v2
main
7feeb5b665ab1ecdfd9cc4fe45fadb86b7b91b5b
{ "arc:challenge": 71.8, "hellaswag": 86.9, "hendrycksTest": 69.4, "truthfulqa:mc": 64.8 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
73.2
oh-yeontaek/llama-2-70B-LoRA-assemble
main
91caffe08852dcbbdedd64786bd3b4ac0dcb2e96
{ "arc:challenge": 71.8, "hellaswag": 86.8, "hendrycksTest": 69.4, "truthfulqa:mc": 64.8 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
73.2
oh-yeontaek/llama-2-13B-LoRA-assemble
main
85bb49d333dba4a08b051418663d16853ce30cee
{ "arc:challenge": 63.6, "hellaswag": 83.5, "hendrycksTest": 59.8, "truthfulqa:mc": 56 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
65.7
teknium/OpenHermes-7B
main
74edb1ad58d3d517ef46c4e2a31081084ecbc473
{ "arc:challenge": 56.1, "hellaswag": 78.3, "hendrycksTest": 48.6, "truthfulqa:mc": 45 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
57
teknium/OpenHermes-13B
main
f09d0fe655ad57cce9179b7b40ea6f81e07db18c
{ "arc:challenge": 60.2, "hellaswag": 82.2, "hendrycksTest": 56.2, "truthfulqa:mc": 46 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
61.2
psmathur/model_007
main
0f5d81b13718a866cb078bd8762ab80a41972663
{ "arc:challenge": 71.1, "hellaswag": 87.7, "hendrycksTest": 69, "truthfulqa:mc": 63.1 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
72.7
psmathur/orca_mini_7b
main
6ed0dca683685cb5b9e7df599f87d311f00ba6db
{ "arc:challenge": 43.9, "hellaswag": 65.2, "hendrycksTest": 30, "truthfulqa:mc": 42 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
45.3
psmathur/orca_mini_v2_7b
main
165850882991d7fa4eabab577a03ed84e0713bfa
{ "arc:challenge": 50.8, "hellaswag": 76, "hendrycksTest": 39.5, "truthfulqa:mc": 43.9 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
52.6