model
stringlengths 4
89
| revision
stringclasses 1
value | model_sha
stringlengths 0
40
| results
dict | commit
stringlengths 40
40
| date
timestamp[ns] | score
float64 21.8
83
⌀ |
---|---|---|---|---|---|---|
togethercomputer/RedPajama-INCITE-Instruct-7B-v0.1 | main | 95667a602ff2646bf67fe3a57c4eb9a1edec87fe | {
"arc:challenge": 44.1,
"hellaswag": 72,
"hendrycksTest": 37.6,
"truthfulqa:mc": 34
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 46.9 |
togethercomputer/RedPajama-INCITE-7B-Instruct | main | 95667a602ff2646bf67fe3a57c4eb9a1edec87fe | {
"arc:challenge": 44.1,
"hellaswag": 72,
"hendrycksTest": 37.6,
"truthfulqa:mc": 34
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 46.9 |
togethercomputer/Llama-2-7B-32K-Instruct | main | 35696b9a7ab330dcbe240ff76fb44ab1eccf45bf | {
"arc:challenge": 51.1,
"hellaswag": 78.5,
"hendrycksTest": 46.1,
"truthfulqa:mc": 44.9
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 55.2 |
togethercomputer/LLaMA-2-7B-32K | main | aef6d8946ae1015bdb65c478a2dd73b58daaef47 | {
"arc:challenge": 48,
"hellaswag": 77.5,
"hendrycksTest": 45.4,
"truthfulqa:mc": 38.5
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 52.4 |
togethercomputer/RedPajama-INCITE-Base-3B-v1 | main | 094fbdd0c911feb485ce55de1952ab2e75277e1e | {
"arc:challenge": 40.2,
"hellaswag": 64.8,
"hendrycksTest": 27,
"truthfulqa:mc": 33.2
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 41.3 |
togethercomputer/RedPajama-INCITE-7B-Base | main | 78f7e482443971f4873ba3239f0ac810a367833b | {
"arc:challenge": 46.2,
"hellaswag": 71.6,
"hendrycksTest": 27.7,
"truthfulqa:mc": 33
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 44.6 |
togethercomputer/RedPajama-INCITE-Instruct-3B-v1 | main | 0c66778ee09a036886741707733620b91057909a | {
"arc:challenge": 41.6,
"hellaswag": 65.5,
"hendrycksTest": 25,
"truthfulqa:mc": 36.4
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 42.1 |
togethercomputer/Pythia-Chat-Base-7B | main | 97aa918c383820e1a69f042801091d7deb996c20 | {
"arc:challenge": 40,
"hellaswag": 68.7,
"hendrycksTest": 27.4,
"truthfulqa:mc": 34.6
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 42.7 |
togethercomputer/GPT-NeoXT-Chat-Base-20B | main | d386708e84d862a65f7d2b4989f64750cb657227 | {
"arc:challenge": 45.6,
"hellaswag": 74,
"hendrycksTest": 29.9,
"truthfulqa:mc": 34.5
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 46 |
togethercomputer/GPT-JT-6B-v1 | main | f34aa35f906895602c1f86f5685e598afdea8051 | {
"arc:challenge": 40.9,
"hellaswag": 67.1,
"hendrycksTest": 47.2,
"truthfulqa:mc": 37.1
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 48.1 |
togethercomputer/GPT-JT-Moderation-6B | main | 1297870783f6091294769014afddf94499966a78 | {
"arc:challenge": 40.5,
"hellaswag": 67.7,
"hendrycksTest": 41.6,
"truthfulqa:mc": 37.3
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 46.8 |
togethercomputer/RedPajama-INCITE-Base-7B-v0.1 | main | 78f7e482443971f4873ba3239f0ac810a367833b | {
"arc:challenge": 46.2,
"hellaswag": 71.6,
"hendrycksTest": 27.7,
"truthfulqa:mc": 33
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 44.6 |
togethercomputer/RedPajama-INCITE-Chat-3B-v1 | main | f0e0995eba801096ed04cb87931d96a8316871af | {
"arc:challenge": 42.8,
"hellaswag": 67.6,
"hendrycksTest": 26.2,
"truthfulqa:mc": 34.4
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 42.8 |
Aspik101/trurl-2-13b-pl-instruct_unload | main | 17f57642165e30a4025d6817bd47dcd80d0c5c4d | {
"arc:challenge": 59.9,
"hellaswag": 80,
"hendrycksTest": 78.7,
"truthfulqa:mc": 45.6
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 66.1 |
Aspik101/trurl-2-7b-pl-instruct_unload | main | 768d800e4dbe3fc95334f30ca7cd02113d3e3fd3 | {
"arc:challenge": 53.2,
"hellaswag": 74.6,
"hendrycksTest": 49.9,
"truthfulqa:mc": 45.7
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 55.8 |
Aspik101/Vicuzard-30B-Uncensored-instruct-PL-lora_unload | main | 652f03ac67b4293198d98b618e64285fb32a28e9 | {
"arc:challenge": 62.5,
"hellaswag": 83.7,
"hendrycksTest": 57.8,
"truthfulqa:mc": 50.9
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 63.7 |
Aspik101/Nous-Hermes-13b-pl-lora_unload | main | d0ef3991a11c4dc2ea2f832d4082c89c3c5e810c | {
"arc:challenge": 57.1,
"hellaswag": 81.5,
"hendrycksTest": 49.2,
"truthfulqa:mc": 48.3
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 59 |
Aspik101/vicuna-13b-v1.5-PL-lora_unload | main | 5c8aeb722e11d1c7258abd45f9f2840f57976c28 | {
"arc:challenge": 56.9,
"hellaswag": 81.2,
"hendrycksTest": 56.1,
"truthfulqa:mc": 49.8
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 61 |
Aspik101/Redmond-Puffin-13B-instruct-PL-lora_unload | main | b933009635299bca32c694336aa2007d756a2dda | {
"arc:challenge": 60.9,
"hellaswag": 82.4,
"hendrycksTest": 55.6,
"truthfulqa:mc": 44.3
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 60.8 |
Aspik101/tulu-7b-instruct-pl-lora_unload | main | 962d4e5d8da5a4ec0ec047b6f8f08f1bb9e509fe | {
"arc:challenge": 28.7,
"hellaswag": 26.1,
"hendrycksTest": 23.1,
"truthfulqa:mc": 48.6
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 31.6 |
Aspik101/llama-30b-2048-instruct-PL-lora_unload | main | b15f4310ea37fef99e4f16372a4b1f2342e27613 | {
"arc:challenge": 63.8,
"hellaswag": 84.7,
"hendrycksTest": 61.5,
"truthfulqa:mc": 52.5
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 65.6 |
Aspik101/30B-Lazarus-instruct-PL-lora_unload | main | eeb29b35ceb6dd5c532f1e4e1235f1cdd3f51f23 | {
"arc:challenge": 62.8,
"hellaswag": 84.1,
"hendrycksTest": 56.9,
"truthfulqa:mc": 55.5
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 64.8 |
Aspik101/vicuna-7b-v1.3-instruct-pl-lora_unload | main | e4b19d9d6168b32402da4ab2b5ec7ff27cf40d9b | {
"arc:challenge": 48,
"hellaswag": 76.3,
"hendrycksTest": 47.4,
"truthfulqa:mc": 44.4
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 54 |
Aspik101/StableBeluga-13B-instruct-PL-lora_unload | main | 6e1a6e1f91f6ac97b643be1bd24be6096e2e7dd3 | {
"arc:challenge": 60.9,
"hellaswag": 82.1,
"hendrycksTest": 57,
"truthfulqa:mc": 48.6
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 62.2 |
Aspik101/llama-30b-instruct-2048-PL-lora | main | 1a076bce564f03bd47951eecab628c541fb1a6ad | {
"arc:challenge": 63.3,
"hellaswag": 84.7,
"hendrycksTest": 61.7,
"truthfulqa:mc": 53.3
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 65.8 |
Aspik101/WizardVicuna-Uncensored-3B-instruct-PL-lora_unload | main | e471ec778771f29992293d1660cc108f29c9c69e | {
"arc:challenge": 42,
"hellaswag": 66.8,
"hendrycksTest": 25.7,
"truthfulqa:mc": 39.7
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 43.6 |
Aspik101/Llama-2-7b-hf-instruct-pl-lora_unload | main | 3dfef350be9c8ce92c2d314dbe96a002bd6ca97d | {
"arc:challenge": 53.8,
"hellaswag": 78.3,
"hendrycksTest": 46.8,
"truthfulqa:mc": 42.3
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 55.3 |
NewstaR/Starlight-13B | main | cb9fced568b1abd881133c642c427aaa488f00cc | {
"arc:challenge": 59.3,
"hellaswag": 82.2,
"hendrycksTest": 55.7,
"truthfulqa:mc": 37.4
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 58.6 |
NewstaR/Starlight-7B | main | 1f7436c458ebc3d8d31b91091c1a7a48e942cd3b | {
"arc:challenge": 53.1,
"hellaswag": 78.6,
"hendrycksTest": 46.8,
"truthfulqa:mc": 38.8
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 54.3 |
NewstaR/Morningstar-13b-hf | main | 2605b5b3b0ecba906ac26d39aab40f33c2ec81c9 | {
"arc:challenge": 59,
"hellaswag": 81.9,
"hendrycksTest": 54.6,
"truthfulqa:mc": 44.1
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 59.9 |
AtomEchoAI/AtomGPT_56k | main | f69ecfd630ec89afffa4ca7bd8a5eda0daf57643 | {
"arc:challenge": 53.2,
"hellaswag": 76.7,
"hendrycksTest": 45.3,
"truthfulqa:mc": 40.3
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 53.9 |
MayaPH/GodziLLa-30B-plus | main | a66b1860d11ebf8aed07237cf636fdd2b3a07f06 | {
"arc:challenge": 28.9,
"hellaswag": 26.4,
"hendrycksTest": 24.6,
"truthfulqa:mc": 48.9
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 32.2 |
MayaPH/GodziLLa-30B-instruct | main | 642bf3683801e20e4b7cf28d94374d5e6054c007 | {
"arc:challenge": 29,
"hellaswag": 26.5,
"hendrycksTest": 24.9,
"truthfulqa:mc": 48.8
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 32.3 |
MayaPH/FinOPT-Washington | main | cdd8a6cde7902de39757cf31d73af1f51df0d8e8 | {
"arc:challenge": 25.2,
"hellaswag": 26.2,
"hendrycksTest": 24.8,
"truthfulqa:mc": 45.8
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 30.5 |
MayaPH/GodziLLa2-70B | main | 7b78087db07eec97f7b461d10758ece76d685543 | {
"arc:challenge": 71.4,
"hellaswag": 87.5,
"hendrycksTest": 69.9,
"truthfulqa:mc": 61.5
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 72.6 |
MayaPH/GodziLLa-30B | main | aa9912a2ac60abeac28b4566731cd903dcc582ac | {
"arc:challenge": 61.5,
"hellaswag": 82.1,
"hendrycksTest": 54.2,
"truthfulqa:mc": 55.9
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 63.4 |
MayaPH/FinOPT-Franklin | main | 1b13331834190bfe49a176f1661ba4d8309a5051 | {
"arc:challenge": 27.7,
"hellaswag": 24.9,
"hendrycksTest": 23.1,
"truthfulqa:mc": 52.4
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 32 |
MayaPH/FinOPT-Lincoln | main | 7ddc381fa3968df22f72acb6cf03b75d3ac49661 | {
"arc:challenge": 26.7,
"hellaswag": 25.6,
"hendrycksTest": 23,
"truthfulqa:mc": 50.6
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 31.5 |
MayaPH/opt-flan-iml-6.7b | main | cbe8d60db6f3c52e653ca73e23a1c34c08127d02 | {
"arc:challenge": 30.1,
"hellaswag": 58.8,
"hendrycksTest": 25.1,
"truthfulqa:mc": 36.7
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 37.7 |
Mikivis/gpt2-large-lora-sft1 | main | 8e26a8d2dc1661d87a8652c75f00b805d63e7330 | {
"arc:challenge": 24.7,
"hellaswag": 42.7,
"hendrycksTest": 24.9,
"truthfulqa:mc": 39.4
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 32.9 |
Mikivis/gpt2-large-lora-sft | main | 1c0c5a686f3c83692e033416197155557e4d3a0d | {
"arc:challenge": 26.8,
"hellaswag": 44.2,
"hendrycksTest": 25.8,
"truthfulqa:mc": 39.1
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 34 |
Mikivis/gpt2-large-lora-sft2 | main | 1244efb5d20765beb54f6b4a4e1426cf6d5daf44 | {
"arc:challenge": 26.6,
"hellaswag": 42.7,
"hendrycksTest": 24.7,
"truthfulqa:mc": 40.3
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 33.6 |
Mikivis/gpt2-large-lora-stf4 | main | 82eff3a62116fd589ad7319c9d75ff6b12f42f72 | {
"arc:challenge": 26.9,
"hellaswag": 42.2,
"hendrycksTest": 25.5,
"truthfulqa:mc": 40.8
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 33.8 |
matsuo-lab/weblab-10b | main | d6fc432983b1633a4c1568d121c60de6b8c3e511 | {
"arc:challenge": 39.5,
"hellaswag": 65.8,
"hendrycksTest": 26.3,
"truthfulqa:mc": 36
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 41.9 |
matsuo-lab/weblab-10b-instruction-sft | main | 112a5ad9f556078ab14a5cd93511b9db4a0d4413 | {
"arc:challenge": 40.1,
"hellaswag": 65.3,
"hendrycksTest": 26.7,
"truthfulqa:mc": 36.8
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 42.2 |
Norquinal/llama-2-7b-claude-chat-rp | main | 4309eedebe8ba5709e0cc7cf186cb783f3bc8060 | {
"arc:challenge": 54.9,
"hellaswag": 80.1,
"hendrycksTest": 47,
"truthfulqa:mc": 43.5
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 56.4 |
anton-l/gpt-j-tiny-random | main | feea91564dac0081f73aeb6744979c6cfe553fff | {
"arc:challenge": 26.4,
"hellaswag": 25.8,
"hendrycksTest": 24.5,
"truthfulqa:mc": 47.4
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 31 |
GeorgiaTechResearchInstitute/starcoder-gpteacher-code-instruct | main | d866b68daa719239dc44979dbf39a608ed6f7bce | {
"arc:challenge": 32.7,
"hellaswag": 47.6,
"hendrycksTest": 28.6,
"truthfulqa:mc": 40.4
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 37.3 |
GeorgiaTechResearchInstitute/galactica-6.7b-evol-instruct-70k | main | 14fa470051d0bc38fd871643186a9edfd3a8a9aa | {
"arc:challenge": 42.6,
"hellaswag": 49.3,
"hendrycksTest": 33,
"truthfulqa:mc": 42.1
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 41.8 |
GeorgiaTechResearchInstitute/galpaca-30b | main | a1f0c4bedd65b485a0d4d3a3bd60d7a4599f1eaf | {
"arc:challenge": 49.6,
"hellaswag": 58.2,
"hendrycksTest": 43.8,
"truthfulqa:mc": 41.2
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 48.2 |
xzuyn/MedicWizard-7B | main | 0b3ef975fb5e8ac1eae775160ab54c98221889df | {
"arc:challenge": 53.5,
"hellaswag": 78.4,
"hendrycksTest": 44.6,
"truthfulqa:mc": 41.3
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 54.4 |
xzuyn/Alpacino-SuperCOT-13B | main | 3a82b04684fe99d59556421c3f96a187049a3cec | {
"arc:challenge": 58.4,
"hellaswag": 81.7,
"hendrycksTest": 47.9,
"truthfulqa:mc": 45.4
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 58.4 |
tianyil1/denas-llama2 | main | b8aebc9157c0e427536aeac9132021fd66615702 | {
"arc:challenge": 53.9,
"hellaswag": 77.8,
"hendrycksTest": 45.5,
"truthfulqa:mc": 45.2
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 55.6 |
NekoPunchBBB/Llama-2-13b-hf_Open-Platypus-QLoRA-multigpu | main | f65029ea8f030731ace568e40bab33a7097a13de | {
"arc:challenge": 57.5,
"hellaswag": 82.5,
"hendrycksTest": 54.8,
"truthfulqa:mc": 43.8
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 59.6 |
NekoPunchBBB/Llama-2-13b-hf_Open-Platypus-8bit-att | main | 83a8e51d0a72dcfbe5de13dc7ee10dc20e91602e | {
"arc:challenge": 57.5,
"hellaswag": 82.1,
"hendrycksTest": 54.6,
"truthfulqa:mc": 42.2
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 59.1 |
NekoPunchBBB/Llama-2-13b-hf_Open-Platypus | main | c318a24121bd69509f395e17a9636093213ece21 | {
"arc:challenge": 58.9,
"hellaswag": 82.1,
"hendrycksTest": 55,
"truthfulqa:mc": 42.8
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 59.7 |
FlagAlpha/Llama2-Chinese-13b-Chat | main | cb69cda10a72bc9736b1c10181ac41f28b69ff9b | {
"arc:challenge": 56,
"hellaswag": 82,
"hendrycksTest": 54.7,
"truthfulqa:mc": 48.9
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 60.4 |
jphme/orca_mini_v2_ger_7b | main | 175965f50907c6a8cd40f1a4b10d28342969c066 | {
"arc:challenge": 49.8,
"hellaswag": 75.5,
"hendrycksTest": 39.1,
"truthfulqa:mc": 45.7
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 52.5 |
jaspercatapang/Echidna-30B | main | 20b13b6676d54b555ae2b9b2b4b6fc8a0c7c2e89 | {
"arc:challenge": 28.5,
"hellaswag": 25.5,
"hendrycksTest": 24.9,
"truthfulqa:mc": 48.1
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 31.8 |
timdettmers/guanaco-33b-merged | main | b2e78a916582935b6616d184b22ea5e9e1eb4c34 | {
"arc:challenge": 62.5,
"hellaswag": 84.5,
"hendrycksTest": 53.8,
"truthfulqa:mc": 51.2
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 63 |
timdettmers/guanaco-65b-merged | main | 98c803bb6e70efe9f2aefb12cba36a96f2959d4d | {
"arc:challenge": 27.5,
"hellaswag": 26.6,
"hendrycksTest": 25.2,
"truthfulqa:mc": 48.4
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 31.9 |
xhyi/PT_GPTNEO350_ATG | main | 56ab08aaa6802d0f830d42c352d5d536be72811d | {
"arc:challenge": 25.4,
"hellaswag": 37.6,
"hendrycksTest": 24.8,
"truthfulqa:mc": 43
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 32.7 |
Andron00e/YetAnother_Open-Llama-3B-LoRA-OpenOrca | main | {
"arc:challenge": 24.8,
"hellaswag": 26.3,
"hendrycksTest": 25.2,
"truthfulqa:mc": null
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | null |
|
Andron00e/YetAnother_Open-Llama-3B-LoRA | main | 52c5cb0178831908ed0571f1750fcb0f0fb125f9 | {
"arc:challenge": 25.9,
"hellaswag": 25.8,
"hendrycksTest": 24.7,
"truthfulqa:mc": null
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | null |
WizardLM/WizardLM-30B-V1.0 | main | 815e2dd7daabe446c429f3c9f70ef01582528f81 | {
"arc:challenge": 27.4,
"hellaswag": 25.9,
"hendrycksTest": 23.1,
"truthfulqa:mc": 48.6
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 31.2 |
WizardLM/WizardMath-13B-V1.0 | main | 209316bea6eab73d8b18fca2a730b1dff3dcf999 | {
"arc:challenge": 60.1,
"hellaswag": 82,
"hendrycksTest": 54.8,
"truthfulqa:mc": 42.7
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 59.9 |
WizardLM/WizardMath-70B-V1.0 | main | e85b43e53c5379e35393b970c66d76c2d1060381 | {
"arc:challenge": 68.2,
"hellaswag": 86.5,
"hendrycksTest": 68.9,
"truthfulqa:mc": 52.7
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 69.1 |
WizardLM/WizardLM-70B-V1.0 | main | 6dae38060d70b82dcfe787a612d04aaf0adf0738 | {
"arc:challenge": 65.4,
"hellaswag": 84.4,
"hendrycksTest": 64,
"truthfulqa:mc": 54.8
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 67.2 |
WizardLM/WizardCoder-15B-V1.0 | main | 926ca1b215c4631bc5f8c3e47173381452c23e5c | {
"arc:challenge": 32.3,
"hellaswag": 47.2,
"hendrycksTest": 29.4,
"truthfulqa:mc": 41.6
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 37.6 |
WizardLM/WizardCoder-Python-34B-V1.0 | main | 5cdc34e4a81d202f1d4a3b5d60e028aab895dfeb | {
"arc:challenge": 52.1,
"hellaswag": 74.8,
"hendrycksTest": 49.1,
"truthfulqa:mc": 48.8
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 56.2 |
WizardLM/WizardLM-13B-V1.1 | main | badd80f8a6f46fb15310fedf6d4db54959854897 | {
"arc:challenge": 60.2,
"hellaswag": 81.4,
"hendrycksTest": 50.9,
"truthfulqa:mc": 54.6
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 61.8 |
WizardLM/WizardMath-7B-V1.0 | main | 06dbd3e0da08255c575e585cb82e0554c1d2707a | {
"arc:challenge": 54.1,
"hellaswag": 79.5,
"hendrycksTest": 46,
"truthfulqa:mc": 43.7
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 55.8 |
WizardLM/WizardLM-13B-V1.2 | main | 6760d0c07ffdc2405295ed7a29437cf4dc414bac | {
"arc:challenge": 59,
"hellaswag": 82.2,
"hendrycksTest": 54.6,
"truthfulqa:mc": 47.3
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 60.8 |
sartmis1/starcoder-finetune-openapi | main | fed87393fd749e46c0c82da09d433deb9b7cf9ee | {
"arc:challenge": 30.6,
"hellaswag": 48.1,
"hendrycksTest": 30.4,
"truthfulqa:mc": 41.8
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 37.7 |
sartmis1/starcoder-finetune-selfinstruct | main | b21bd307ea7417185e7dc59557c399a3e4e0092b | {
"arc:challenge": 31.2,
"hellaswag": 47.7,
"hendrycksTest": 29.5,
"truthfulqa:mc": 41.6
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 37.5 |
PY007/TinyLlama-1.1B-step-50K-105b | main | c1f1ef67c12e4bb85fe0bdf1747c645a202cc118 | {
"arc:challenge": 25.9,
"hellaswag": 44.1,
"hendrycksTest": 26.8,
"truthfulqa:mc": 39.5
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 34.1 |
PY007/TinyLlama-1.1B-intermediate-step-240k-503b | main | 213ebf60d7fdd3258fa5574840b06c97a7e8cf5d | {
"arc:challenge": 29.3,
"hellaswag": 49.7,
"hendrycksTest": 26.3,
"truthfulqa:mc": 40.2
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 36.4 |
HanningZhang/Robin-v2 | main | {
"arc:challenge": 48.8,
"hellaswag": 74.5,
"hendrycksTest": 39.3,
"truthfulqa:mc": 42.3
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 51.2 |
|
nicholasKluge/Aira-2-774M | main | f43044cfe7bf0827a176f0d319c63251c2b29373 | {
"arc:challenge": 28.8,
"hellaswag": 40.8,
"hendrycksTest": 25.1,
"truthfulqa:mc": 41.3
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 34 |
nicholasKluge/Aira-2-355M | main | 2479f5b1bb62251ec88e60182ba81390a4c19cf9 | {
"arc:challenge": 27.6,
"hellaswag": 38.9,
"hendrycksTest": 27.3,
"truthfulqa:mc": 38.5
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 33.1 |
nicholasKluge/Aira-124M | main | 0c0d509ec9ce057e7b506e15c868eecf79cc8ae5 | {
"arc:challenge": 24.6,
"hellaswag": 31.3,
"hendrycksTest": 25.3,
"truthfulqa:mc": 41
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 30.6 |
DanielSc4/RedPajama-INCITE-Chat-3B-v1-RL-LoRA-8bit-test1 | main | a2ee88a9fa1c9ad41e0a8c15217a4b1230ec33c8 | {
"arc:challenge": 41.3,
"hellaswag": 66.8,
"hendrycksTest": 26.1,
"truthfulqa:mc": 35
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 42.3 |
DanielSc4/RedPajama-INCITE-Chat-3B-v1-FT-LoRA-8bit-test1 | main | f477d24b00e05fe4c5f8d5f933080994cfd90e4e | {
"arc:challenge": 38.7,
"hellaswag": 63.5,
"hendrycksTest": 25.2,
"truthfulqa:mc": 36.1
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 40.9 |
IGeniusDev/llama13B-quant8-testv1-openorca-customdataset | main | f364d000bedac80e72aa103c08b77aee1b61b7da | {
"arc:challenge": 60.2,
"hellaswag": 83,
"hendrycksTest": 54.3,
"truthfulqa:mc": 37.3
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 58.7 |
NousResearch/Nous-Hermes-Llama2-13b | main | 8f95aa9cd207db7b24179fc779c2b8973e71bee2 | {
"arc:challenge": 61.3,
"hellaswag": 83.3,
"hendrycksTest": 55,
"truthfulqa:mc": 50.4
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 62.5 |
NousResearch/CodeLlama-13b-hf | main | b7cfbbce945b966607d15ae275704922a6d04afc | {
"arc:challenge": 40.9,
"hellaswag": 63.4,
"hendrycksTest": 32.8,
"truthfulqa:mc": 43.8
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 45.2 |
NousResearch/Nous-Puffin-70B | main | 129e0af93d04b1b9cc85ea48bbb300f1ccb44210 | {
"arc:challenge": 67.4,
"hellaswag": 87.4,
"hendrycksTest": 69.8,
"truthfulqa:mc": 46.8
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 67.9 |
NousResearch/Nous-Hermes-13b | main | 24e8c03148ffd1f3e469744dfc24ad2ad82848f8 | {
"arc:challenge": 56.6,
"hellaswag": 82.1,
"hendrycksTest": 50.4,
"truthfulqa:mc": 51.5
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 60.2 |
NousResearch/Nous-Hermes-llama-2-7b | main | 60e58acecdc1552e1b1752a38d1d91d942d1c3f0 | {
"arc:challenge": 55.1,
"hellaswag": 78.9,
"hendrycksTest": 48.3,
"truthfulqa:mc": 49
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 57.8 |
NousResearch/CodeLlama-34b-hf | main | 4e61ec70eb258047f5bc689fa6a66f7753da52b8 | {
"arc:challenge": 37.5,
"hellaswag": 31.8,
"hendrycksTest": 37.2,
"truthfulqa:mc": 38.9
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 36.4 |
NousResearch/Nous-Hermes-Llama2-70b | main | 13a0b4da159ad95c93e72a002d893c48ed0f257a | {
"arc:challenge": 67.6,
"hellaswag": 86.8,
"hendrycksTest": 69.7,
"truthfulqa:mc": 55
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 69.8 |
NousResearch/Redmond-Puffin-13B | main | 12af25fa7ea02c4fc636952ea8b9dc9cf48e35be | {
"arc:challenge": 60.5,
"hellaswag": 83.2,
"hendrycksTest": 55,
"truthfulqa:mc": 42.1
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 60.2 |
ddobokki/Llama-2-70b-orca-200k | main | 1ab69d47a467f15d8168b119ad24c1842d3ff54e | {
"arc:challenge": 64.8,
"hellaswag": 85.3,
"hendrycksTest": 66.9,
"truthfulqa:mc": 56.2
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 68.3 |
eachadea/vicuna-13b-1.1 | main | bfcc6ca66694310be6c85ba0638597f4256c4143 | {
"arc:challenge": 52.7,
"hellaswag": 80.1,
"hendrycksTest": 51.9,
"truthfulqa:mc": 52.1
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 59.2 |
eachadea/vicuna-13b | main | ac4218770a58baaaaf25201076fe082abb6ffd13 | {
"arc:challenge": 51.7,
"hellaswag": 79.9,
"hendrycksTest": 50.8,
"truthfulqa:mc": 52.7
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 58.8 |
eachadea/vicuna-7b-1.1 | main | 9d8eea215e00b388a22e8f050768ea8911d41f1d | {
"arc:challenge": 53.7,
"hellaswag": 77.5,
"hendrycksTest": 45.6,
"truthfulqa:mc": 48.9
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 56.4 |
lvkaokao/llama2-7b-hf-chat-lora-v2 | main | 0b8e61d3325cddbad207cbf885c2b5db6a83a059 | {
"arc:challenge": 55,
"hellaswag": 78.8,
"hendrycksTest": 51.3,
"truthfulqa:mc": 44
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 57.3 |
lvkaokao/llama2-7b-hf-instruction-lora | main | f660a40323b29040e78097acca320517ed242512 | {
"arc:challenge": 55.4,
"hellaswag": 78.6,
"hendrycksTest": 49.4,
"truthfulqa:mc": 41.8
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 56.3 |
lvkaokao/llama2-7b-hf-chat-lora-v3 | main | 79047f667253c878ad3143b016e3dcb3df707572 | {
"arc:challenge": 57.3,
"hellaswag": 78.6,
"hendrycksTest": 50.6,
"truthfulqa:mc": 50.6
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 59.3 |
OpenBuddy/openbuddy-openllama-13b-v7-fp16 | main | 8690c065bccd3e897ccbf3d8aa24b0216a6f5dba | {
"arc:challenge": 47.6,
"hellaswag": 72.2,
"hendrycksTest": 47.7,
"truthfulqa:mc": 48.7
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 54 |