model
stringlengths 4
89
| revision
stringclasses 1
value | model_sha
stringlengths 0
40
| results
dict | commit
stringlengths 40
40
| date
timestamp[ns] | score
float64 21.8
83
⌀ |
---|---|---|---|---|---|---|
Weyaxi/TekniumAiroboros-Nebula-7B
|
main
|
ef964d514cc25a600b0de78fc469d1acbec34591
|
{
"arc:challenge": 57.2,
"hellaswag": 81.7,
"hendrycksTest": 55.3,
"truthfulqa:mc": 51.6
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 61.4 |
Weyaxi/SlimOpenOrca-Mistral-7B
|
main
|
b0134a7512444dfbb60a2e2d81469a5bbbb18026
|
{
"arc:challenge": 63,
"hellaswag": 83.5,
"hendrycksTest": 62.3,
"truthfulqa:mc": 57.4
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 66.6 |
Weyaxi/SynthIA-v1.3-Nebula-v2-7B
|
main
|
c6030620e9d4390d54ec221a18ff3e530f4dcd84
|
{
"arc:challenge": 59.4,
"hellaswag": 82.8,
"hendrycksTest": 57.6,
"truthfulqa:mc": 50.6
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 62.6 |
Weyaxi/OpenOrca-Zephyr-7B
|
main
|
2a2c7d287a46243cccf3ff6628375d0d190394ac
|
{
"arc:challenge": 64.1,
"hellaswag": 83.8,
"hendrycksTest": 62.5,
"truthfulqa:mc": 54.3
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 66.2 |
Weyaxi/Samantha-Nebula-7B
|
main
|
a7d4b8a1683e33dd3c60064d7dd9d5c35691323f
|
{
"arc:challenge": 57,
"hellaswag": 82.3,
"hendrycksTest": 54.2,
"truthfulqa:mc": 49.6
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 60.8 |
Weyaxi/Dolphin2.1-OpenOrca-7B
|
main
|
076c0f7de93307e8fb3ad3bd820fb5f73325ca70
|
{
"arc:challenge": 63.9,
"hellaswag": 84.3,
"hendrycksTest": 62.7,
"truthfulqa:mc": 53.8
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 66.2 |
Weyaxi/zephyr-beta-Nebula-v2-7B
|
main
|
226caedb50a12730232c1f8fe9c96b6dcf818ba7
|
{
"arc:challenge": 56.6,
"hellaswag": 82.5,
"hendrycksTest": 56.4,
"truthfulqa:mc": 58.7
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 63.6 |
Weyaxi/Dolphin-Nebula-7B
|
main
|
c14b3545066e5ee5562c1724a037b41db95f1f0d
|
{
"arc:challenge": 55.2,
"hellaswag": 78.6,
"hendrycksTest": 53.4,
"truthfulqa:mc": 58
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 61.3 |
Weyaxi/test-help-steer-filtered-orig
|
main
|
bda6d45ddb3ef73df4d198d95416c66872429927
|
{
"arc:challenge": 57.6,
"hellaswag": 80.4,
"hendrycksTest": 57.2,
"truthfulqa:mc": 41.1
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 59.1 |
Weyaxi/Luban-Marcoroni-13B
|
main
|
bf152c36935acd67a9029c017f0c1ff2d7a92314
|
{
"arc:challenge": 63.7,
"hellaswag": 82.9,
"hendrycksTest": 58.7,
"truthfulqa:mc": 55.6
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 65.2 |
jerryjalapeno/nart-100k-7b
|
main
|
50e61b8e6cc17cb3fbcb490fe3dc7e2c8b248378
|
{
"arc:challenge": 54.1,
"hellaswag": 78.5,
"hendrycksTest": 35,
"truthfulqa:mc": 36.7
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 51.1 |
SLAM-group/NewHope
|
main
|
560ca6df8335d6d2998ac8f079218816a5742b02
|
{
"arc:challenge": 60.9,
"hellaswag": 84,
"hendrycksTest": 55.7,
"truthfulqa:mc": 44.9
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 61.4 |
psmathur/model_007
|
main
|
0f5d81b13718a866cb078bd8762ab80a41972663
|
{
"arc:challenge": 71.1,
"hellaswag": 87.7,
"hendrycksTest": 69,
"truthfulqa:mc": 63.1
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 72.7 |
psmathur/model_101
|
main
|
884c53a64a3c5faf7b0706d36a587ca1532ed8f5
|
{
"arc:challenge": 68.7,
"hellaswag": 86.4,
"hendrycksTest": 69.9,
"truthfulqa:mc": 58.9
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 71 |
psmathur/orca_mini_v2_13b
|
main
|
1058709314f7ca090937d0a2b7b37b0b3a8f12a3
|
{
"arc:challenge": 55.1,
"hellaswag": 79.7,
"hendrycksTest": 50.1,
"truthfulqa:mc": 52.6
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 59.4 |
psmathur/orca_mini_13b
|
main
|
ca900c8f3145de40cd188c559b2901a2e4711546
|
{
"arc:challenge": 42.1,
"hellaswag": 63.4,
"hendrycksTest": 35.4,
"truthfulqa:mc": 43.1
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 46 |
psmathur/orca_mini_v3_70b
|
main
|
c1d4f997f8ed685a6efc72229523b2e56fd0774b
|
{
"arc:challenge": 71.2,
"hellaswag": 87.9,
"hendrycksTest": 70.2,
"truthfulqa:mc": 61.3
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 72.6 |
psmathur/orca_mini_v3_7b
|
main
|
a1583d2f02041fb37df28eeae4da644d8dff33eb
|
{
"arc:challenge": 56.9,
"hellaswag": 79.6,
"hendrycksTest": 52.4,
"truthfulqa:mc": 50.5
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 59.8 |
psmathur/orca_mini_7b
|
main
|
6ed0dca683685cb5b9e7df599f87d311f00ba6db
|
{
"arc:challenge": 43.9,
"hellaswag": 65.2,
"hendrycksTest": 30,
"truthfulqa:mc": 42
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 45.3 |
psmathur/model_007_v2
|
main
|
3d95e0f3598f7a76ab97cb2cc0e4aae957d77479
|
{
"arc:challenge": 71.4,
"hellaswag": 87.3,
"hendrycksTest": 68.6,
"truthfulqa:mc": 62.7
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 72.5 |
psmathur/orca_mini_3b
|
main
|
fd2754e80ce80757a3a68a840d7d287dd7def676
|
{
"arc:challenge": 41.6,
"hellaswag": 61.5,
"hendrycksTest": 26.8,
"truthfulqa:mc": 42.4
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 43.1 |
psmathur/model_007_13b
|
main
|
0436ba68d245c8a2c04a2cc9637630d2e163cbbe
|
{
"arc:challenge": 22.7,
"hellaswag": 25,
"hendrycksTest": 23.1,
"truthfulqa:mc": null
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | null |
psmathur/model_42_70b
|
main
|
ca3789cd6b683e97dcd6a5f0367f90a63d7a4e7b
|
{
"arc:challenge": 68.3,
"hellaswag": 87.7,
"hendrycksTest": 70,
"truthfulqa:mc": 48.8
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 68.7 |
psmathur/test_42_70b
|
main
|
ca3789cd6b683e97dcd6a5f0367f90a63d7a4e7b
|
{
"arc:challenge": 68.3,
"hellaswag": 87.7,
"hendrycksTest": 70,
"truthfulqa:mc": 48.8
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 68.7 |
rinna/bilingual-gpt-neox-4b
|
main
|
f02f6f3c8da0093f3c1ce59220409bc2fa9fbb17
|
{
"arc:challenge": 29.2,
"hellaswag": 43.7,
"hendrycksTest": 23.1,
"truthfulqa:mc": 45
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 35.2 |
rinna/bilingual-gpt-neox-4b-instruction-sft
|
main
|
c20e42bd49a3b1b0d0a07151899a322c4760e871
|
{
"arc:challenge": 28.1,
"hellaswag": 47.5,
"hendrycksTest": 23.1,
"truthfulqa:mc": 43.8
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 35.6 |
rinna/bilingual-gpt-neox-4b-8k
|
main
|
ad56d7fc86db4ad5a7036bc9f80e11cd6f435a60
|
{
"arc:challenge": 28.6,
"hellaswag": 43.9,
"hendrycksTest": 25.4,
"truthfulqa:mc": 47.5
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 36.4 |
Kiddyz/testlm
|
main
|
e00d8c50a007eb1da3fbfb4d5f5a73c1af3aa104
|
{
"arc:challenge": 53.5,
"hellaswag": 75.8,
"hendrycksTest": 51.2,
"truthfulqa:mc": 48.4
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 57.2 |
Kiddyz/testlm2
|
main
|
9bffd9acfb12b5da1a1dd09825a633f804126dfa
|
{
"arc:challenge": 53,
"hellaswag": 75.6,
"hendrycksTest": 51.5,
"truthfulqa:mc": 48.7
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 57.2 |
Kiddyz/testlm-1
|
main
|
e00d8c50a007eb1da3fbfb4d5f5a73c1af3aa104
|
{
"arc:challenge": 53.5,
"hellaswag": 75.8,
"hendrycksTest": 51.2,
"truthfulqa:mc": 48.4
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 57.2 |
Kiddyz/testlm-1-1
|
main
|
e00d8c50a007eb1da3fbfb4d5f5a73c1af3aa104
|
{
"arc:challenge": 53.5,
"hellaswag": 75.8,
"hendrycksTest": 51.2,
"truthfulqa:mc": 48.4
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 57.2 |
Kiddyz/testllm-c2
|
main
|
b87c798bc27522824451dfccf5eae50edbd4263b
|
{
"arc:challenge": 60.6,
"hellaswag": 81.9,
"hendrycksTest": 61.2,
"truthfulqa:mc": 49.9
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 63.4 |
Kiddyz/testlm-3
|
main
|
6ba288ac39fc4145144e360a8f2641d6f5a6a33a
|
{
"arc:challenge": 53.6,
"hellaswag": 78.5,
"hendrycksTest": 51.8,
"truthfulqa:mc": 46.4
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 57.6 |
AIDC-ai-business/Marcoroni-70B-v1
|
main
|
55a30d29db194832c0b5de1392a6598a63582144
|
{
"arc:challenge": 73.5,
"hellaswag": 87.6,
"hendrycksTest": 70.7,
"truthfulqa:mc": 64.4
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 74.1 |
cerebras/Cerebras-GPT-1.3B
|
main
|
5b95400ee8d1e3cc9f79f0dec7182ed9c1009c34
|
{
"arc:challenge": 26.3,
"hellaswag": 38.5,
"hendrycksTest": 26.6,
"truthfulqa:mc": 42.7
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 33.5 |
cerebras/Cerebras-GPT-256M
|
main
|
d77812ac95aece1f1edef6745ae2a1b325ad01a4
|
{
"arc:challenge": 22,
"hellaswag": 29,
"hendrycksTest": 26.8,
"truthfulqa:mc": 46
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 31 |
cerebras/Cerebras-GPT-590M
|
main
|
67a653304fd782a34906d59f3795a37f9e053397
|
{
"arc:challenge": 23.7,
"hellaswag": 32.4,
"hendrycksTest": 26,
"truthfulqa:mc": 44.1
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 31.6 |
cerebras/Cerebras-GPT-2.7B
|
main
|
4383dfd80aafdbcfd0876419d246de51e6cbf7c1
|
{
"arc:challenge": 29.1,
"hellaswag": 49.3,
"hendrycksTest": 25.2,
"truthfulqa:mc": 41.4
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 36.2 |
jordiclive/Llama-2-70b-oasst-1-200
|
main
|
153b209007e688d713cd670c9972f2827c597b45
|
{
"arc:challenge": 67.7,
"hellaswag": 87.2,
"hendrycksTest": 70,
"truthfulqa:mc": 51.3
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 69 |
ziqingyang/chinese-alpaca-2-7b
|
main
|
ab2476bffedeed752daedd77e71900578e136e7c
|
{
"arc:challenge": 49.6,
"hellaswag": 72.6,
"hendrycksTest": 46.5,
"truthfulqa:mc": 48.6
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 54.3 |
ziqingyang/chinese-alpaca-2-13b
|
main
|
576094cbf4988baf88b3bb66678be1db70bd720a
|
{
"arc:challenge": 58.7,
"hellaswag": 79.7,
"hendrycksTest": 55.1,
"truthfulqa:mc": 50.2
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 60.9 |
Severian/ANIMA-Phi-Neptune-Mistral-7B-v1
|
main
|
35dd5fee8563b61c41743e88be6c557f409c1c10
|
{
"arc:challenge": 52.9,
"hellaswag": 74.7,
"hendrycksTest": 52.2,
"truthfulqa:mc": 59.4
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 59.8 |
Severian/ANIMA-Phi-Neptune-Mistral-7B-v3
|
main
|
fd6fda131561917202905be1f4f3b0adc13efdb5
|
{
"arc:challenge": 56.8,
"hellaswag": 78.8,
"hendrycksTest": 53.8,
"truthfulqa:mc": 59.4
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 62.2 |
Severian/ANIMA-Phi-Neptune-Mistral-7B-v4
|
main
|
a8e18f970f7ca994740177d6c228adee9e17aba9
|
{
"arc:challenge": 55.5,
"hellaswag": 77.6,
"hendrycksTest": 53.1,
"truthfulqa:mc": 59
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 61.3 |
Severian/ANIMA-Phi-Neptune-Mistral-LoRa
|
main
|
feef1ab8eeb7ba21685b93e074141136d95174bf
|
{
"arc:challenge": 53.1,
"hellaswag": 74.7,
"hendrycksTest": 52.1,
"truthfulqa:mc": 59.4
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 59.8 |
Yukang/Llama-2-13b-chat-longlora-32k-sft
|
main
|
6f2924e354c3ab035aa2ff7c7e28d0e5327e2667
|
{
"arc:challenge": 26.1,
"hellaswag": 26.2,
"hendrycksTest": 23.1,
"truthfulqa:mc": 49.1
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 31.1 |
Yukang/Llama-2-7b-longlora-32k-ft
|
main
|
ab48674ffc55568ffe2a1207ef0e711c2febbaaf
|
{
"arc:challenge": 27.9,
"hellaswag": 25.6,
"hendrycksTest": 23.1,
"truthfulqa:mc": 49.6
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 31.6 |
Yukang/Llama-2-7b-longlora-16k-ft
|
main
|
c86de31b80866d047e680e08dbd3572e2965d4c5
|
{
"arc:challenge": 26.4,
"hellaswag": 26.4,
"hendrycksTest": 23.7,
"truthfulqa:mc": 47.8
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 31.1 |
Yukang/Llama-2-7b-longlora-100k-ft
|
main
|
242c6469cab41b41d30826e850afa4687e422f24
|
{
"arc:challenge": 28.2,
"hellaswag": 25.4,
"hendrycksTest": 23.5,
"truthfulqa:mc": 49.1
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 31.6 |
Yukang/Llama-2-13b-longlora-16k-ft
|
main
|
5f0cfdef590fc9bd7642042fb5f1ed9679260b93
|
{
"arc:challenge": 25.9,
"hellaswag": 27.6,
"hendrycksTest": 23.1,
"truthfulqa:mc": 48.9
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 31.4 |
Yukang/Llama-2-13b-longlora-32k-ft
|
main
|
6d17c854025b0bd54ce572ac803f1bb052875dbf
|
{
"arc:challenge": 59.5,
"hellaswag": 82.6,
"hendrycksTest": 52.1,
"truthfulqa:mc": 37.4
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 57.9 |
dfurman/llama-2-7b-instruct-peft
|
main
|
0fc43413117187e0723cdac133068ab527c80fe2
|
{
"arc:challenge": 51.2,
"hellaswag": 78.9,
"hendrycksTest": 46.6,
"truthfulqa:mc": 48.5
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 56.3 |
dfurman/Llama-2-13B-Instruct-v0.2
|
main
|
ac4b0962df8430f0b31c76a3d97a61134114c87e
|
{
"arc:challenge": 60.6,
"hellaswag": 82,
"hendrycksTest": 55.5,
"truthfulqa:mc": 45.7
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 61 |
dfurman/llama-2-13b-guanaco-peft
|
main
|
10b58a7c31d5513fa56a9b8b38739253bf5cc0b4
|
{
"arc:challenge": 60,
"hellaswag": 82.4,
"hendrycksTest": 55.8,
"truthfulqa:mc": 42.6
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 60.2 |
dfurman/llama-2-70b-dolphin-peft
|
main
|
a1190dee60b5854e80d340958dc3cc956bc56f68
|
{
"arc:challenge": 69.6,
"hellaswag": 86.8,
"hendrycksTest": 69.2,
"truthfulqa:mc": 57.4
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 70.7 |
PY007/TinyLlama-1.1B-intermediate-step-240k-503b
|
main
|
213ebf60d7fdd3258fa5574840b06c97a7e8cf5d
|
{
"arc:challenge": 29.3,
"hellaswag": 49.7,
"hendrycksTest": 26.3,
"truthfulqa:mc": 40.2
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 36.4 |
PY007/TinyLlama-1.1B-step-50K-105b
|
main
|
c1f1ef67c12e4bb85fe0bdf1747c645a202cc118
|
{
"arc:challenge": 25.9,
"hellaswag": 44.1,
"hendrycksTest": 26.8,
"truthfulqa:mc": 39.5
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 34.1 |
FelixChao/llama2-13b-math1.1
|
main
|
3c4d83d3525e54a493ff510443fdcca44bf63b59
|
{
"arc:challenge": 57.3,
"hellaswag": 80.7,
"hendrycksTest": 53.6,
"truthfulqa:mc": 48.4
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 60 |
sauce1337/AppleSauce-L2-13b
|
main
|
ba253c52eb85e24987c81e5d36b5a9a00e276ce7
|
{
"arc:challenge": 61,
"hellaswag": 83.6,
"hendrycksTest": 57.1,
"truthfulqa:mc": 47.8
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 62.4 |
glaiveai/glaive-coder-7b
|
main
|
72a255a58480ef0713eed988312fe82f77f94f37
|
{
"arc:challenge": 42.7,
"hellaswag": 64.7,
"hendrycksTest": 37.2,
"truthfulqa:mc": 39.9
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 46.1 |
Voicelab/trurl-2-7b
|
main
|
e26ca5f157c60fc527170cc04db7fc0ea04ad26f
|
{
"arc:challenge": 53.4,
"hellaswag": 75.3,
"hendrycksTest": 50,
"truthfulqa:mc": 45.4
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 56 |
lloorree/kssht-castor-70b
|
main
|
e49a6bdc5e6024fb0e60dbba4601b346b4369377
|
{
"arc:challenge": 69.5,
"hellaswag": 87.5,
"hendrycksTest": 70.4,
"truthfulqa:mc": 56.3
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 70.9 |
lloorree/jfdslijsijdgis
|
main
|
1e67eaa4ef618a5a0d8c52e5e107635c706b34c5
|
{
"arc:challenge": 69.6,
"hellaswag": 87.3,
"hendrycksTest": 70,
"truthfulqa:mc": 59.2
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 71.5 |
lloorree/kssht-dahj-70b
|
main
|
85901aab2c3faf09de5ba8e9d65ec03aee4b20e4
|
{
"arc:challenge": 70.8,
"hellaswag": 87.3,
"hendrycksTest": 70.4,
"truthfulqa:mc": 58.9
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 71.8 |
lloorree/kssht-euripedes-70b
|
main
|
04ae5f2187697a7e9a2d57f327a7131f23d3e927
|
{
"arc:challenge": 69.8,
"hellaswag": 87.6,
"hendrycksTest": 70.4,
"truthfulqa:mc": 55.5
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 70.8 |
chavinlo/gpt4-x-alpaca
|
main
|
6a571f458cab9a23d14324ec63e0abd1744c8353
|
{
"arc:challenge": 52.8,
"hellaswag": 79.6,
"hendrycksTest": 48.2,
"truthfulqa:mc": 48.9
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 57.4 |
SkunkworksAI/Mistralic-7B-1
|
main
|
ebf138de4fb7a57f0d187ad0ab43abd6b35bfb62
|
{
"arc:challenge": 60.8,
"hellaswag": 82.3,
"hendrycksTest": 60.8,
"truthfulqa:mc": 52.4
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 64.1 |
yulan-team/YuLan-Chat-2-13b-fp16
|
main
|
2d439187efd6edd91a0c0146f08dff52d92aa7bc
|
{
"arc:challenge": 59,
"hellaswag": 80.7,
"hendrycksTest": 56.7,
"truthfulqa:mc": 52.2
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 62.1 |
Undi95/MXLewd-L2-20B
|
main
|
ac279478abd9ddb8d1f5adcc548be0287b963adf
|
{
"arc:challenge": 63.2,
"hellaswag": 85.3,
"hendrycksTest": 57.4,
"truthfulqa:mc": 51.6
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 64.4 |
Undi95/CreativityEngine
|
main
|
7870cc50b82b5cbebfa9935b6d73a9d20170299a
|
{
"arc:challenge": 59.3,
"hellaswag": 82.4,
"hendrycksTest": 53.6,
"truthfulqa:mc": 52.5
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 62 |
Undi95/OpenRP-13B
|
main
|
d11815287c51ef51485fb003f8f72773cf6f19a4
|
{
"arc:challenge": 62.1,
"hellaswag": 82.6,
"hendrycksTest": 57.5,
"truthfulqa:mc": 48.3
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 62.6 |
Undi95/U-Amethyst-20B
|
main
|
c0cbe0b3c88041bb6beef27dbe85146af8dddec9
|
{
"arc:challenge": 62.2,
"hellaswag": 83.1,
"hendrycksTest": 55.9,
"truthfulqa:mc": 53.2
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 63.6 |
Undi95/Amethyst-13B
|
main
|
d4a85b1006f0b9439e64f0e7400533a7b867c24d
|
{
"arc:challenge": 62.6,
"hellaswag": 83.2,
"hendrycksTest": 55.9,
"truthfulqa:mc": 52.4
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 63.5 |
Undi95/Nous-Hermes-13B-Code
|
main
|
5a45cb2a6442581ce32cc19c561c49cec1db4ebb
|
{
"arc:challenge": 61.2,
"hellaswag": 83.2,
"hendrycksTest": 55.1,
"truthfulqa:mc": 50.6
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 62.5 |
Undi95/ReMM-v2-L2-13B
|
main
|
bc42c77f88482c37c72c85c66135e99972bbca1b
|
{
"arc:challenge": 61.9,
"hellaswag": 84,
"hendrycksTest": 56.1,
"truthfulqa:mc": 50.8
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 63.2 |
Undi95/Llama2-13B-no_robots-alpaca-lora
|
main
|
581aba329e607533c299746bb9eb4154a7aab139
|
{
"arc:challenge": 58.9,
"hellaswag": 82.4,
"hendrycksTest": 53.1,
"truthfulqa:mc": 40.5
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 58.7 |
Undi95/MLewd-v2.4-13B
|
main
|
6f6ec6024ee054020e49fd96f149919692848f0b
|
{
"arc:challenge": 61.7,
"hellaswag": 83.8,
"hendrycksTest": 55.1,
"truthfulqa:mc": 53.3
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 63.5 |
Undi95/Mistral-11B-TestBench3
|
main
|
7eb397ad2ec67400e31dc010f9b364a72d64d965
|
{
"arc:challenge": 62,
"hellaswag": 83.9,
"hendrycksTest": 63.1,
"truthfulqa:mc": 53.7
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 65.7 |
Undi95/ReMM-L2-13B
|
main
|
c4710577003a23ca8e9040d16dfb8f3e9bc5d636
|
{
"arc:challenge": 59.7,
"hellaswag": 83.1,
"hendrycksTest": 54.1,
"truthfulqa:mc": 49.9
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 61.7 |
Undi95/ReMM-v2.1-L2-13B
|
main
|
e6b5ac97f74355cb281a621261debe5720fb4da2
|
{
"arc:challenge": 61.4,
"hellaswag": 83.9,
"hendrycksTest": 56,
"truthfulqa:mc": 50.3
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 62.9 |
Undi95/CodeEngine
|
main
|
f57879831c39f2dcb656cb2c9e9ce5878e92bb44
|
{
"arc:challenge": 58.4,
"hellaswag": 82.3,
"hendrycksTest": 54.2,
"truthfulqa:mc": 45.2
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 60 |
Undi95/MM-ReMM-L2-20B
|
main
|
37869800c15fb37d017ea83bb50fec6d6141f6ba
|
{
"arc:challenge": 60.8,
"hellaswag": 85.2,
"hendrycksTest": 56.5,
"truthfulqa:mc": 53.3
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 64 |
Undi95/ReMM-L2-13B-PIPPA
|
main
|
79e711178c6881496ae1f5635b08bc193f370709
|
{
"arc:challenge": 59.7,
"hellaswag": 83.1,
"hendrycksTest": 54.1,
"truthfulqa:mc": 49.9
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 61.7 |
Undi95/Mistral-11B-TestBench11
|
main
|
9aae2b156b24557bb98e515f3a90c7865529d2e9
|
{
"arc:challenge": 64.4,
"hellaswag": 83.9,
"hendrycksTest": 63.8,
"truthfulqa:mc": 56.7
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 67.2 |
Undi95/ReMM-v2.2-L2-13B
|
main
|
d55031fbcd41d749bc0c0ffbcd85636718d373b6
|
{
"arc:challenge": 61.3,
"hellaswag": 84.2,
"hendrycksTest": 56.2,
"truthfulqa:mc": 51.4
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 63.3 |
Undi95/Mistral-11B-TestBench7
|
main
|
3d4d99f90ec582e0d532e11f6da419d6b962c536
|
{
"arc:challenge": 63.3,
"hellaswag": 82.9,
"hendrycksTest": 64.1,
"truthfulqa:mc": 46.9
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 64.3 |
Undi95/UndiMix-v4-13B
|
main
|
6dd97c74cfe1d22432d5c993814e230f333ba401
|
{
"arc:challenge": 61.9,
"hellaswag": 83.9,
"hendrycksTest": 56.9,
"truthfulqa:mc": 49
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 62.9 |
Undi95/Mistral-11B-TestBench10
|
main
|
730429d6132c7702885840098885081c2df878df
|
{
"arc:challenge": 64.2,
"hellaswag": 84.2,
"hendrycksTest": 63.9,
"truthfulqa:mc": 55.6
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 67 |
Undi95/Emerald-13B
|
main
|
f7696299463d8ec402a4e1eb001f3a447f1c5552
|
{
"arc:challenge": 62.3,
"hellaswag": 83.7,
"hendrycksTest": 55.7,
"truthfulqa:mc": 50.9
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 63.2 |
Undi95/ReMM-Mistral-13B
|
main
|
a5ef9385d9430a81778183d71b58eb2b869d6a7e
|
{
"arc:challenge": 62.2,
"hellaswag": 83.8,
"hendrycksTest": 55.4,
"truthfulqa:mc": 53.3
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 63.7 |
IGeniusDev/llama13B-quant8-testv1-openorca-customdataset
|
main
|
f364d000bedac80e72aa103c08b77aee1b61b7da
|
{
"arc:challenge": 60.2,
"hellaswag": 83,
"hendrycksTest": 54.3,
"truthfulqa:mc": 37.3
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 58.7 |
harborwater/open-llama-3b-claude-30k
|
main
|
049db7fda44e5ce1e8febf5c3f45e3a93aaaa859
|
{
"arc:challenge": 41.7,
"hellaswag": 72.6,
"hendrycksTest": 24,
"truthfulqa:mc": 38.5
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 44.2 |
harborwater/open-llama-3b-everything-v2
|
main
|
31ce2c1611d9f7d56184ceb5bff6a7e95a180c03
|
{
"arc:challenge": 42.8,
"hellaswag": 73.3,
"hendrycksTest": 26.9,
"truthfulqa:mc": 37.3
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 45.1 |
AA051610/VA
|
main
|
3c60daea2db0840475b3f67149122d9e033eab5b
|
{
"arc:challenge": 41.4,
"hellaswag": 62.5,
"hendrycksTest": 50,
"truthfulqa:mc": 44.9
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 49.7 |
AA051610/T1B
|
main
|
6f3664328e9f07a6578ccb0c5713b747cc0549d5
|
{
"arc:challenge": 56.1,
"hellaswag": 79.8,
"hendrycksTest": 60,
"truthfulqa:mc": 47
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 60.7 |
AA051610/T2A
|
main
|
c30e3b053299c7ecf250af143a816ef8a9a45c22
|
{
"arc:challenge": 51.5,
"hellaswag": 74,
"hendrycksTest": 62.1,
"truthfulqa:mc": 47
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 58.6 |
AA051610/T1C
|
main
|
1b1185ddc427341df12dd1aa8c68090fde16b5f3
|
{
"arc:challenge": 50.2,
"hellaswag": 72.2,
"hendrycksTest": 56.3,
"truthfulqa:mc": 42.5
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 55.3 |
huashiyiqike/testmodel
|
main
|
1ac5d244402e2433b6abfcff1fe65e84af15766b
|
{
"arc:challenge": 19.7,
"hellaswag": 26.7,
"hendrycksTest": 25.3,
"truthfulqa:mc": 43.7
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 28.8 |
maywell/Synatra-7B-v0.3-RP
|
main
|
372f6e0ab2c20b93e0c42218f76a71a4f9bb282e
|
{
"arc:challenge": 62.2,
"hellaswag": 82.3,
"hendrycksTest": 60.8,
"truthfulqa:mc": 52.6
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 64.5 |
maywell/Synatra-7B-v0.3-dpo
|
main
|
405a4f1e6513cd1b8de5eb4e003bb49cc86d1f8a
|
{
"arc:challenge": 62.8,
"hellaswag": 82.6,
"hendrycksTest": 61.5,
"truthfulqa:mc": 56.5
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 65.8 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.