model
stringlengths
4
89
revision
stringclasses
1 value
model_sha
stringlengths
0
40
results
dict
commit
stringlengths
40
40
date
unknown
score
float64
21.8
83
TaylorAI/Flash-Llama-3B
main
b4c7bb49171ff6955cfc1f7e33143383c57f7606
{ "arc:challenge": 40.1, "hellaswag": 71.6, "hendrycksTest": 26.9, "truthfulqa:mc": 34.7 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
43.3
TaylorAI/FLAN-Llama-7B-2_Llama2-7B-Flash_868_full_model
main
819f3f384e37f8906a62a8048556c9e58e495c02
{ "arc:challenge": 52.5, "hellaswag": 79.1, "hendrycksTest": 47.6, "truthfulqa:mc": 37.1 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
54.1
TaylorAI/Flash-Llama-13B
main
81b40096471a8980e3e1a8998f358bd363033783
{ "arc:challenge": 59.3, "hellaswag": 82.2, "hendrycksTest": 55.7, "truthfulqa:mc": 37.4 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
58.6
TaylorAI/Flash-Llama-7B
main
27c84ef23d850582453e1cc2dcea13de48da090f
{ "arc:challenge": 53.1, "hellaswag": 78.6, "hendrycksTest": 46.8, "truthfulqa:mc": 38.8 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
54.3
facebook/xglm-1.7B
main
d23a5e8e2164af31a84a26756b9b17f925143050
{ "arc:challenge": 25.9, "hellaswag": 45.7, "hendrycksTest": 25.1, "truthfulqa:mc": 37.2 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
33.5
facebook/xglm-564M
main
f3059f01b98ccc877c673149e0178c0e957660f9
{ "arc:challenge": 24.6, "hellaswag": 34.6, "hendrycksTest": 25.2, "truthfulqa:mc": 40.4 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
31.2
facebook/opt-iml-max-30b
main
291753b04817a31a742631053ee361874d6db8a4
{ "arc:challenge": 43.9, "hellaswag": 72.4, "hendrycksTest": 41.1, "truthfulqa:mc": 38.2 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
48.9
facebook/xglm-4.5B
main
dc6a67fac06c8bca7860b84656a0cb736293a7a8
{ "arc:challenge": 31.5, "hellaswag": 57.9, "hendrycksTest": 25.4, "truthfulqa:mc": 35.8 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
37.7
facebook/opt-13b
main
e515202d1e7750da62d245fbccb2723b9c1790f5
{ "arc:challenge": 39.9, "hellaswag": 71.2, "hendrycksTest": 24.9, "truthfulqa:mc": 34.1 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
42.5
facebook/galactica-30b
main
80bd55898b06c7c363c467dec877b8b32702a2c4
{ "arc:challenge": 47.4, "hellaswag": 61.2, "hendrycksTest": 47.6, "truthfulqa:mc": 38 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
48.6
facebook/opt-66b
main
7259969061237fe940036d22bea0fd349e4485e9
{ "arc:challenge": 46.3, "hellaswag": 76.2, "hendrycksTest": 27, "truthfulqa:mc": 35.4 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
46.2
facebook/galactica-1.3b
main
f711c69357d598defb703ddce93c5d7f7bc6e6da
{ "arc:challenge": 34.1, "hellaswag": 40.9, "hendrycksTest": 27.1, "truthfulqa:mc": 41.4 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
35.9
facebook/opt-1.3b
main
8c7b10754972749675d22364c25c428b29face51
{ "arc:challenge": 29.5, "hellaswag": 54.5, "hendrycksTest": 25, "truthfulqa:mc": 38.7 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
36.9
facebook/xglm-7.5B
main
732d59308a844004bd9a4def972cc7c3896a38e0
{ "arc:challenge": 34.1, "hellaswag": 60.8, "hendrycksTest": 27.8, "truthfulqa:mc": 36.7 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
39.8
facebook/opt-6.7b
main
a45aa65bbeb77c1558bc99bedc6779195462dab0
{ "arc:challenge": 39.2, "hellaswag": 68.7, "hendrycksTest": 24.6, "truthfulqa:mc": 35.1 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
41.9
facebook/opt-iml-max-1.3b
main
d60fa58f50def19751da2075791da359ca19d273
{ "arc:challenge": 30.7, "hellaswag": 53.8, "hendrycksTest": 27.6, "truthfulqa:mc": 38.3 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
37.6
facebook/opt-125m
main
3d2b5f275bdf882b8775f902e1bfdb790e2cfc32
{ "arc:challenge": 22.9, "hellaswag": 31.5, "hendrycksTest": 26, "truthfulqa:mc": 42.9 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
30.8
facebook/opt-2.7b
main
397f71a473a150c00f0fe3fc4a2f78ff3ccaf82d
{ "arc:challenge": 34, "hellaswag": 61.4, "hendrycksTest": 25.4, "truthfulqa:mc": 37.4 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
39.6
facebook/opt-30b
main
ceea0a90ac0f6fae7c2c34bcb40477438c152546
{ "arc:challenge": 43.2, "hellaswag": 74.1, "hendrycksTest": 26.6, "truthfulqa:mc": 35.2 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
44.8
facebook/opt-350m
main
cb32f77e905cccbca1d970436fb0f5e6b58ee3c5
{ "arc:challenge": 23.5, "hellaswag": 36.7, "hendrycksTest": 26, "truthfulqa:mc": 40.8 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
31.8
gpt2-medium
main
f65d4965d1221eff2bcf34f53a2ba12120e18f24
{ "arc:challenge": 27, "hellaswag": 40.2, "hendrycksTest": 26.6, "truthfulqa:mc": 40.8 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
33.7
ahnyeonchan/OpenOrca-AYT-13B
main
1357abceda30e8389007a023907824cc3a11e397
{ "arc:challenge": 27.2, "hellaswag": 26, "hendrycksTest": 25.1, "truthfulqa:mc": null }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
null
Taekyoon/llama2-ko-7b-test
main
1d9b52cc5832ae0ea37514330d38193b737e1d07
{ "arc:challenge": 37.8, "hellaswag": 63, "hendrycksTest": 29.6, "truthfulqa:mc": 36 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
41.6
Mikael110/llama-2-13b-guanaco-fp16
main
feb7ef47ceca6aec9548264a39622b63fdcb853c
{ "arc:challenge": 60.9, "hellaswag": 83.2, "hendrycksTest": 54.6, "truthfulqa:mc": 44 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
60.7
Mikael110/llama-2-7b-guanaco-fp16
main
f769fed10874af73ad12115efd044cb4a64506b0
{ "arc:challenge": 54.9, "hellaswag": 79.6, "hendrycksTest": 46.4, "truthfulqa:mc": 43.8 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
56.2
ehartford/dolphin-llama2-7b
main
85aa4f67191fd016ab7ea8c389fddb5d9e5a9a52
{ "arc:challenge": 46.6, "hellaswag": 67.5, "hendrycksTest": 48.4, "truthfulqa:mc": 49.7 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
53
ehartford/Wizard-Vicuna-30B-Uncensored
main
6374baef4cedd41f85c111b8eec3eb38ee24c4b9
{ "arc:challenge": 62.1, "hellaswag": 83.4, "hendrycksTest": 58.2, "truthfulqa:mc": 50.8 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
63.6
ehartford/CodeLlama-34b-Python-hf
main
45f38e53a579a2b39298cc57ab04078722bebec0
{ "arc:challenge": 38.1, "hellaswag": 34.8, "hendrycksTest": 33, "truthfulqa:mc": 43.6 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
37.4
ehartford/Samantha-1.11-7b
main
730cbd8f3077f3d24001aab714def991f1e4e7e8
{ "arc:challenge": 55, "hellaswag": 79.1, "hendrycksTest": 40.5, "truthfulqa:mc": 50.4 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
56.2
ehartford/WizardLM-1.0-Uncensored-Llama2-13b
main
134cea14627fd875f6f277cad92f988024855478
{ "arc:challenge": 55.7, "hellaswag": 80.3, "hendrycksTest": 55.4, "truthfulqa:mc": 51.4 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
60.7
ehartford/Samantha-1.11-70b
main
49e5b5ee0bed2864f0b38ba8bf9e01ccc5e0ba5f
{ "arc:challenge": 70.1, "hellaswag": 87.6, "hendrycksTest": 67.8, "truthfulqa:mc": 65 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
72.6
ehartford/Wizard-Vicuna-13B-Uncensored
main
95bfd1640a54e76b3e857c2462fd3a77eca0b275
{ "arc:challenge": 59, "hellaswag": 81.9, "hendrycksTest": 47.9, "truthfulqa:mc": 51.7 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
60.1
ehartford/WizardLM-13B-Uncensored
main
9025c5f96fef9525da9238369ad082961b0e9494
{ "arc:challenge": 50.9, "hellaswag": 76.6, "hendrycksTest": 44, "truthfulqa:mc": 46.7 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
54.6
ehartford/Wizard-Vicuna-7B-Uncensored
main
1097285acd9c48a1d09bc0a9844d365384732111
{ "arc:challenge": 53.4, "hellaswag": 78.8, "hendrycksTest": 37.1, "truthfulqa:mc": 43.5 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
53.2
ehartford/WizardLM-1.0-Uncensored-CodeLlama-34b
main
3e8df2cf4a4ee1c0b2d079cb7be70024d425ea8c
{ "arc:challenge": 56.4, "hellaswag": 75.5, "hendrycksTest": 54.5, "truthfulqa:mc": 43.1 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
57.4
ehartford/WizardLM-33B-V1.0-Uncensored
main
3eca9fdee0ce28d6a4a635a6f19d9a413caee3e7
{ "arc:challenge": 63.7, "hellaswag": 83.8, "hendrycksTest": 59.4, "truthfulqa:mc": 56.8 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
65.9
ehartford/WizardLM-30B-Uncensored
main
761783745fcb97831ad8035d3cbd5de484aca3ce
{ "arc:challenge": 60.2, "hellaswag": 82.9, "hendrycksTest": 56.8, "truthfulqa:mc": 51.6 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
62.9
ehartford/CodeLlama-34b-Instruct-hf
main
50ac374da09ab585b9cf7625a2ea3554ef97f18a
{ "arc:challenge": 40.8, "hellaswag": 35.7, "hendrycksTest": 39.7, "truthfulqa:mc": 44.3 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
40.1
ehartford/dolphin-llama-13b
main
b6d16c3e1cffef5e914863f41fd96152dafddd6f
{ "arc:challenge": 55.5, "hellaswag": 77.1, "hendrycksTest": 52.2, "truthfulqa:mc": 52.2 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
59.2
ehartford/minotaur-llama2-13b-qlora
main
22c83f7d68e547fb0b59acfa01c60b108c59fe55
{ "arc:challenge": 60.1, "hellaswag": 82.4, "hendrycksTest": 55.9, "truthfulqa:mc": 45.6 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
61
ehartford/Samantha-1.11-13b
main
e355ead3a939f471fe2586201156fb972fad0f4b
{ "arc:challenge": 60.8, "hellaswag": 83, "hendrycksTest": 56, "truthfulqa:mc": 47.7 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
61.9
ehartford/Samantha-1.1-70b
main
a3819d186f5b4d52ced7ddeb7fa16bf66e8a2ea7
{ "arc:challenge": 68.8, "hellaswag": 87.5, "hendrycksTest": 68.6, "truthfulqa:mc": 64.8 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
72.4
ehartford/WizardLM-7B-Uncensored
main
14c23f9fa775ab5ce49010418f00df06d92b0b13
{ "arc:challenge": 47.9, "hellaswag": 73.1, "hendrycksTest": 35.4, "truthfulqa:mc": 41.5 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
49.5
ehartford/based-30b
main
5818a6344f48dc5a324589b57cb288a9d54c0b79
{ "arc:challenge": 63.9, "hellaswag": 85.7, "hendrycksTest": 58.3, "truthfulqa:mc": 35.7 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
60.9
ehartford/Samantha-1.11-CodeLlama-34b
main
3fd110de9282e52f56f999bf1da1a76425f00e29
{ "arc:challenge": 56.6, "hellaswag": 75.5, "hendrycksTest": 53.5, "truthfulqa:mc": 50.5 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
59
junelee/wizard-vicuna-13b
main
419dc5acc391de54a60d0b041e94e767d1ef2032
{ "arc:challenge": 54.7, "hellaswag": 79.2, "hendrycksTest": 48.9, "truthfulqa:mc": 49.6 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
58.1
wenge-research/yayi-70b-llama2
main
2799b262292f78f7c3965a1410d0ad6211438603
{ "arc:challenge": 60.7, "hellaswag": 83.9, "hendrycksTest": 64.4, "truthfulqa:mc": 47.6 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
64.2
wenge-research/yayi-7b
main
00be6c9e41a8367a855c6f18ebfa08f5ecdb2cc4
{ "arc:challenge": 46.3, "hellaswag": 61.7, "hendrycksTest": 36.3, "truthfulqa:mc": 43.7 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
47
wenge-research/yayi-7b-llama2
main
18a4ed38285c732efc583a4bd883b3a681f8d005
{ "arc:challenge": 54.8, "hellaswag": 77.9, "hendrycksTest": 41.4, "truthfulqa:mc": 44 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
54.5
wenge-research/yayi-13b-llama2
main
9fc1bc4409b9e71f54213245a91c2742fbf7b3d0
{ "arc:challenge": 48.5, "hellaswag": 74.8, "hendrycksTest": 38.7, "truthfulqa:mc": 42.2 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
51
budecosystem/genz-13b-v2
main
98e0e2086df11b9f80e1571110540a657e52c2e8
{ "arc:challenge": 56, "hellaswag": 80, "hendrycksTest": 54.3, "truthfulqa:mc": 48.1 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
59.6
budecosystem/genz-70b
main
32110b4f33e5e80073ca1f47638482fdc0e19297
{ "arc:challenge": 71.4, "hellaswag": 88, "hendrycksTest": 70.8, "truthfulqa:mc": 62.7 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
73.2
Kiddyz/testlm-1
main
e00d8c50a007eb1da3fbfb4d5f5a73c1af3aa104
{ "arc:challenge": 53.5, "hellaswag": 75.8, "hendrycksTest": 51.2, "truthfulqa:mc": 48.4 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
57.2
Kiddyz/testlm
main
e00d8c50a007eb1da3fbfb4d5f5a73c1af3aa104
{ "arc:challenge": 53.5, "hellaswag": 75.8, "hendrycksTest": 51.2, "truthfulqa:mc": 48.4 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
57.2
Kiddyz/testlm2
main
9bffd9acfb12b5da1a1dd09825a633f804126dfa
{ "arc:challenge": 53, "hellaswag": 75.6, "hendrycksTest": 51.5, "truthfulqa:mc": 48.7 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
57.2
Kiddyz/testlm-1-1
main
e00d8c50a007eb1da3fbfb4d5f5a73c1af3aa104
{ "arc:challenge": 53.5, "hellaswag": 75.8, "hendrycksTest": 51.2, "truthfulqa:mc": 48.4 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
57.2
Kiddyz/testlm-3
main
6ba288ac39fc4145144e360a8f2641d6f5a6a33a
{ "arc:challenge": 53.6, "hellaswag": 78.5, "hendrycksTest": 51.8, "truthfulqa:mc": 46.4 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
57.6
baichuan-inc/Baichuan-7B
main
8baef65be8363f3b5670adfe9a0b9c0311962d90
{ "arc:challenge": 40.7, "hellaswag": 69, "hendrycksTest": 43.6, "truthfulqa:mc": 36.2 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
47.4
RobbeD/Orca-Platypus-3B
main
243f51d75ed6d425addde839740f6fd5bcc4630f
{ "arc:challenge": 43.1, "hellaswag": 65.3, "hendrycksTest": 26.8, "truthfulqa:mc": 41.9 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
44.3
RobbeD/OpenLlama-Platypus-3B
main
d3a0bf8e1181be02cc9c4c4cdfedaedacaefbfac
{ "arc:challenge": 41.2, "hellaswag": 71.7, "hendrycksTest": 29.9, "truthfulqa:mc": 36.5 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
44.8
behnamsh/gpt2_platypus-camel_physics
main
cd4d700d13b3bc9371bf45616ef74ac20d165c3d
{ "arc:challenge": 22.8, "hellaswag": 31.2, "hendrycksTest": 25.9, "truthfulqa:mc": 39 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
29.7
AlekseyKorshuk/pygmalion-6b-vicuna-chatml
main
ee3ada91a69a194cedfabbfeab98f1499b75cb44
{ "arc:challenge": 40.6, "hellaswag": 67.7, "hendrycksTest": 33.9, "truthfulqa:mc": 42.8 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
46.2
TheBloke/WizardLM-30B-fp16
main
465f87a243969963f25ae6cf8f8d2de6c0898bbe
{ "arc:challenge": 62.5, "hellaswag": 83.3, "hendrycksTest": 59, "truthfulqa:mc": 52.5 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
64.3
TheBloke/CodeLlama-13B-Instruct-fp16
main
521c208c7251ccd3e44ccd9500b6bed419bca565
{ "arc:challenge": 44.6, "hellaswag": 64.9, "hendrycksTest": 38.8, "truthfulqa:mc": 45.9 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
48.6
TheBloke/Planner-7B-fp16
main
afb4604a06c8541960fb51240259777764c4ce7e
{ "arc:challenge": 51, "hellaswag": 77.8, "hendrycksTest": 35.7, "truthfulqa:mc": 34.3 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
49.7
TheBloke/orca_mini_13B-GPTQ
main
8ec18e5c597da86fa123c08b6e6bef7da6ec7440
{ "arc:challenge": 27.3, "hellaswag": 25.9, "hendrycksTest": 25.3, "truthfulqa:mc": 48.1 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
31.6
TheBloke/WizardLM-30B-GPTQ
main
e2e97475a9775d2fe7afba098aee37e694b9220f
{ "arc:challenge": 28.8, "hellaswag": 26.1, "hendrycksTest": 24.6, "truthfulqa:mc": 49.1 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
32.2
TheBloke/Project-Baize-v2-13B-GPTQ
main
8dee7c7129aaad1ded245fce712ff5dbb2845258
{ "arc:challenge": 27.6, "hellaswag": 26.4, "hendrycksTest": 25.9, "truthfulqa:mc": 48.2 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
32
TheBloke/vicuna-13b-v1.3.0-GPTQ
main
6ef1f8d8638ea2d6681a8e3da73be57c501d847b
{ "arc:challenge": 54.4, "hellaswag": 79.5, "hendrycksTest": 52, "truthfulqa:mc": 50.9 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
59.2
TheBloke/Platypus-30B-SuperHOT-8K-fp16
main
e8ac508308911475125252dcf2677fe355dd3700
{ "arc:challenge": 25.7, "hellaswag": 30.8, "hendrycksTest": 23.6, "truthfulqa:mc": 47.1 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
31.8
TheBloke/chronos-wizardlm-uc-scot-st-13B-GPTQ
main
c4246e4b8d3fc77b9fe4ebb1ead61cda4b83575b
{ "arc:challenge": 28, "hellaswag": 26.1, "hendrycksTest": 25.7, "truthfulqa:mc": 49.7 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
32.4
TheBloke/Wizard-Vicuna-13B-Uncensored-GPTQ
main
d9b00ec47ae3546398432f0693fe2d5d92bf143b
{ "arc:challenge": 29.6, "hellaswag": 25.5, "hendrycksTest": 25.3, "truthfulqa:mc": 50.2 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
32.7
TheBloke/WizardLM-13B-V1-1-SuperHOT-8K-fp16
main
83905656ca3e63877b8d9f3a74118da0c9bc6939
{ "arc:challenge": 58.6, "hellaswag": 81.1, "hendrycksTest": 48.3, "truthfulqa:mc": 54.2 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
60.6
TheBloke/wizard-mega-13B-GPTQ
main
848bf2514f804799dd28c188e5428d497dc983fb
{ "arc:challenge": 27.7, "hellaswag": 26, "hendrycksTest": 25, "truthfulqa:mc": 48.7 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
31.8
TheBloke/Llama-2-70B-fp16
main
b25061ef1b440e970d15d4ac99bc42937cd442a2
{ "arc:challenge": 67.3, "hellaswag": 87.3, "hendrycksTest": 69.8, "truthfulqa:mc": 44.9 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
67.3
TheBloke/WizardLM-7B-uncensored-GPTQ
main
cc30c031fd795ee3d3a50312ab4549415bfbdb46
{ "arc:challenge": 28.5, "hellaswag": 25.4, "hendrycksTest": 24.9, "truthfulqa:mc": 50.9 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
32.4
TheBloke/OpenAssistant-SFT-7-Llama-30B-HF
main
a7a2306b9a63de2c545f35b24735f4540baf5903
{ "arc:challenge": 60.6, "hellaswag": 82.2, "hendrycksTest": 57.9, "truthfulqa:mc": 46.9 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
61.9
TheBloke/openchat_v2_openorca_preview-GPTQ
main
5a4c2ea612b71d7c00118f796db7189bc1a0c930
{ "arc:challenge": 28, "hellaswag": 26.1, "hendrycksTest": 24.2, "truthfulqa:mc": 50.1 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
32.1
TheBloke/Llama-2-70B-chat-GPTQ
main
054fbf6f65e7ab7691ec07ec9ad366acf2dd90bf
{ "arc:challenge": 62.6, "hellaswag": 84.8, "hendrycksTest": 62.7, "truthfulqa:mc": 51 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
65.3
TheBloke/VicUnlocked-30B-LoRA-HF
main
3259cb3c2a10cfb429fb51c4a76fffa049f4c44d
{ "arc:challenge": 59.7, "hellaswag": 84, "hendrycksTest": 57.8, "truthfulqa:mc": 48.5 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
62.5
TheBloke/h2ogpt-oasst1-512-30B-HF
main
3dc93836e4b08b7b2ee43e69c1e590a36fd24687
{ "arc:challenge": 57.3, "hellaswag": 81.4, "hendrycksTest": 48.1, "truthfulqa:mc": 45.5 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
58.1
TheBloke/gpt4-alpaca-lora-13B-HF
main
49678a2dd15fb4e1f1b99616ccc1ffd269912833
{ "arc:challenge": 59.6, "hellaswag": 82.1, "hendrycksTest": 47.5, "truthfulqa:mc": 49 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
59.6
TheBloke/alpaca-lora-65B-HF
main
113b61b37a2862b950ada68620e57acafbcefe13
{ "arc:challenge": 64.8, "hellaswag": 85.6, "hendrycksTest": 63.1, "truthfulqa:mc": 45.1 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
64.6
TheBloke/WizardLM-30B-Uncensored-GPTQ
main
43c701ddbe0bceac26c860307e06763cc5203500
{ "arc:challenge": 29.4, "hellaswag": 26.5, "hendrycksTest": 24.4, "truthfulqa:mc": 49.1 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
32.4
TheBloke/WizardLM-Uncensored-SuperCOT-StoryTelling-30B-GPTQ
main
cd07cc7c55b46524f61214012653c25226d24c0d
{ "arc:challenge": 28.4, "hellaswag": 26.1, "hendrycksTest": 24.7, "truthfulqa:mc": 49.5 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
32.2
TheBloke/robin-65b-v2-fp16
main
40edb31ba93045d673735361bc98f56125bbc77b
{ "arc:challenge": 61.9, "hellaswag": 84.6, "hendrycksTest": 62.5, "truthfulqa:mc": 52.3 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
65.3
TheBloke/guanaco-13B-HF
main
bd59c700815124df616a17f5b49a0bc51590b231
{ "arc:challenge": 57.8, "hellaswag": 83.8, "hendrycksTest": 48.3, "truthfulqa:mc": 46.7 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
59.1
TheBloke/LongChat-13B-GPTQ
main
8ec25a29033b7be5daeafa26f08e1ea7cf232b98
{ "arc:challenge": 28.3, "hellaswag": 26.1, "hendrycksTest": 25.6, "truthfulqa:mc": 48.3 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
32.1
TheBloke/GPlatty-30B-SuperHOT-8K-fp16
main
e2103a424c1700756df1c0c0b334195f37efe17b
{ "arc:challenge": 28.3, "hellaswag": 33.5, "hendrycksTest": 24.9, "truthfulqa:mc": 46.3 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
33.2
TheBloke/llama-30b-supercot-SuperHOT-8K-fp16
main
7efdff78a90132c1c66e1d27518ad7cbadffa139
{ "arc:challenge": 25.9, "hellaswag": 30.5, "hendrycksTest": 23.5, "truthfulqa:mc": 47 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
31.7
TheBloke/UltraLM-13B-fp16
main
734f5641f6c548474517d1536c46024517f120e0
{ "arc:challenge": 57.6, "hellaswag": 80.2, "hendrycksTest": 51.9, "truthfulqa:mc": 51.6 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
60.3
TheBloke/gpt4-alpaca-lora_mlp-65B-HF
main
664ff8e3e1d446971a16a6c9018ab24de7664684
{ "arc:challenge": 65, "hellaswag": 86.1, "hendrycksTest": 62.7, "truthfulqa:mc": 59.2 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
68.2
TheBloke/Wizard-Vicuna-30B-Uncensored-fp16
main
c7b7cecb5a314fc66deebabcb67c230a3fbe84f7
{ "arc:challenge": 62.1, "hellaswag": 83.4, "hendrycksTest": 58.2, "truthfulqa:mc": 50.8 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
63.6
TheBloke/Llama-2-13B-fp16
main
b2e65e8ad4bb35e5abaee0170ebd5fc2134a50bb
{ "arc:challenge": 59.3, "hellaswag": 82.2, "hendrycksTest": 55.7, "truthfulqa:mc": 37.4 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
58.6
TheBloke/robin-33B-v2-fp16
main
c0ed7d40c3e52379780638dac3bd1f69597b8e18
{ "arc:challenge": 62.4, "hellaswag": 83.6, "hendrycksTest": 54.7, "truthfulqa:mc": 53.9 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
63.6
TheBloke/guanaco-65B-HF
main
7f83ae526f8b83705ca8434535da8fd8c692f9d0
{ "arc:challenge": 65.4, "hellaswag": 86.5, "hendrycksTest": 62.9, "truthfulqa:mc": 52.8 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
66.9
TheBloke/CAMEL-33B-Combined-Data-SuperHOT-8K-fp16
main
14744d11eab7028c5c845f89db2edc9c6fe2becb
{ "arc:challenge": 25.9, "hellaswag": 31.6, "hendrycksTest": 23.7, "truthfulqa:mc": 48.1 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
32.3
TheBloke/airoboros-7b-gpt4-fp16
main
14aa50fba9f6418c0d5e2d24087eb802931040ef
{ "arc:challenge": 53.1, "hellaswag": 78.7, "hendrycksTest": 38.9, "truthfulqa:mc": 40.7 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
52.9
TheBloke/fiction.live-Kimiko-V2-70B-fp16
main
6b0c2cb654133cad2d4920e7da2e3f6cb1c4f7fd
{ "arc:challenge": 67.7, "hellaswag": 87.7, "hendrycksTest": 69.8, "truthfulqa:mc": 49.3 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
68.6
TheBloke/Vicuna-13B-CoT-fp16
main
fe74a0ece9089828b301bd0f067ae5f257516179
{ "arc:challenge": 52.7, "hellaswag": 80.1, "hendrycksTest": 51.9, "truthfulqa:mc": 52.1 }
10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3
"2023-09-20T10:22:33"
59.2

No dataset card yet

New: Create and edit this dataset card directly on the website!

Contribute a Dataset Card
Downloads last month
1
Add dataset card

Spaces using IlyasMoutawwakil/llm-race-dataset 2