model
stringlengths 4
89
| revision
stringclasses 1
value | model_sha
stringlengths 0
40
| results
dict | commit
stringlengths 40
40
| date
timestamp[ns] | score
float64 21.8
83
⌀ |
---|---|---|---|---|---|---|
RWKV/rwkv-4-169m-pile
|
main
|
46bdc280eb97b6141d5d51a935e0c4870ecaefcc
|
{
"arc:challenge": 23.6,
"hellaswag": 31.7,
"hendrycksTest": 23.2,
"truthfulqa:mc": 41.9
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 30.1 |
RWKV/rwkv-4-430m-pile
|
main
|
a4f6ec80438d4262d1bbc8f385feb2ef1a4a9d6b
|
{
"arc:challenge": 26.7,
"hellaswag": 40,
"hendrycksTest": 24.9,
"truthfulqa:mc": 39.6
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 32.8 |
RWKV/rwkv-4-7b-pile
|
main
|
922e22a761427e50d7be457b31a76b1126021b8b
|
{
"arc:challenge": 39.7,
"hellaswag": 66.3,
"hendrycksTest": 25,
"truthfulqa:mc": 33.7
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 41.2 |
RWKV/rwkv-raven-14b
|
main
|
359c0649b4f1d10a26ebea32908035bc00d152ee
|
{
"arc:challenge": 44.6,
"hellaswag": 71.3,
"hendrycksTest": 25.9,
"truthfulqa:mc": 41.9
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 45.9 |
RWKV/rwkv-raven-1b5
|
main
|
571a3bd891ce33f2ee3fc6de09218178edb0dae2
|
{
"arc:challenge": 31.8,
"hellaswag": 52.6,
"hendrycksTest": 26,
"truthfulqa:mc": 37.1
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 36.9 |
jaspercatapang/Echidna-30B
|
main
|
20b13b6676d54b555ae2b9b2b4b6fc8a0c7c2e89
|
{
"arc:challenge": 28.5,
"hellaswag": 25.5,
"hendrycksTest": 24.9,
"truthfulqa:mc": 48.1
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 31.8 |
CoolWP/llama-2-13b-guanaco-fp16
|
main
|
a60e8e39e4fbe271655e1c78eb1ceb2081518231
|
{
"arc:challenge": 59.6,
"hellaswag": 82.4,
"hendrycksTest": 55.5,
"truthfulqa:mc": 43.4
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 60.2 |
TigerResearch/tigerbot-7b-sft
|
main
|
98b847905d63f74624e834db1ff95ee2814cbbd3
|
{
"arc:challenge": 41.6,
"hellaswag": 60.6,
"hendrycksTest": 29.9,
"truthfulqa:mc": 58.2
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 47.6 |
TigerResearch/tigerbot-7b-base
|
main
|
300831494aa1eb16e59799310a09531f60dcc904
|
{
"arc:challenge": 47.7,
"hellaswag": 72.1,
"hendrycksTest": 45.1,
"truthfulqa:mc": 42.3
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 51.8 |
MayaPH/FinOPT-Washington
|
main
|
cdd8a6cde7902de39757cf31d73af1f51df0d8e8
|
{
"arc:challenge": 25.2,
"hellaswag": 26.2,
"hendrycksTest": 24.8,
"truthfulqa:mc": 45.8
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 30.5 |
MayaPH/GodziLLa-30B
|
main
|
aa9912a2ac60abeac28b4566731cd903dcc582ac
|
{
"arc:challenge": 61.5,
"hellaswag": 82.1,
"hendrycksTest": 54.2,
"truthfulqa:mc": 55.9
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 63.4 |
MayaPH/GodziLLa-30B-instruct
|
main
|
642bf3683801e20e4b7cf28d94374d5e6054c007
|
{
"arc:challenge": 29,
"hellaswag": 26.5,
"hendrycksTest": 24.9,
"truthfulqa:mc": 48.8
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 32.3 |
MayaPH/GodziLLa-30B-plus
|
main
|
a66b1860d11ebf8aed07237cf636fdd2b3a07f06
|
{
"arc:challenge": 28.9,
"hellaswag": 26.4,
"hendrycksTest": 24.6,
"truthfulqa:mc": 48.9
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 32.2 |
MayaPH/opt-flan-iml-6.7b
|
main
|
cbe8d60db6f3c52e653ca73e23a1c34c08127d02
|
{
"arc:challenge": 30.1,
"hellaswag": 58.8,
"hendrycksTest": 25.1,
"truthfulqa:mc": 36.7
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 37.7 |
MayaPH/GodziLLa2-70B
|
main
|
7b78087db07eec97f7b461d10758ece76d685543
|
{
"arc:challenge": 71.4,
"hellaswag": 87.5,
"hendrycksTest": 69.9,
"truthfulqa:mc": 61.5
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 72.6 |
abhishek/llama2guanacotest
|
main
|
679d17809939a0bf9b79bbb027898cbea64045b2
|
{
"arc:challenge": 51.6,
"hellaswag": 77.6,
"hendrycksTest": 48.5,
"truthfulqa:mc": 43.9
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 55.4 |
abhishek/autotrain-llama-alpaca-peft-52508123785
|
main
|
{
"arc:challenge": 52.2,
"hellaswag": 76.9,
"hendrycksTest": 37.6,
"truthfulqa:mc": 32.9
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 49.9 |
|
guardrail/llama-2-7b-guanaco-instruct-sharded
|
main
|
fc7a3abbc3b9a9b3e163ef3c4844307ac270fca7
|
{
"arc:challenge": 53.8,
"hellaswag": 78.7,
"hendrycksTest": 46.7,
"truthfulqa:mc": 43.9
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 55.8 |
CobraMamba/mamba-gpt-3b-v2
|
main
|
935f4d90bd0fc7117113d3c7b6b6af9dba93183d
|
{
"arc:challenge": 42.2,
"hellaswag": 71.5,
"hendrycksTest": 27.1,
"truthfulqa:mc": 36.7
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 44.4 |
CobraMamba/mamba-gpt-7b
|
main
|
cb0b04b1bff7921614efbd87d5b87bac04c58d13
|
{
"arc:challenge": 51.2,
"hellaswag": 75.4,
"hendrycksTest": 47.5,
"truthfulqa:mc": 42.1
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 54 |
CobraMamba/mamba-gpt-7b-v1
|
main
|
e64d658b397748e409d9633fd24fc5a6df429600
|
{
"arc:challenge": 61.3,
"hellaswag": 84.1,
"hendrycksTest": 63.5,
"truthfulqa:mc": 46.3
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 63.8 |
CobraMamba/mamba-gpt-3b-v3
|
main
|
d860a90ef6b30c695b985dd2ff382d4bbb80e857
|
{
"arc:challenge": 41.7,
"hellaswag": 71.1,
"hendrycksTest": 27.3,
"truthfulqa:mc": 37.9
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 44.5 |
CobraMamba/mamba-gpt-7b-v2
|
main
|
6439444e2c0b61253d3e61ae04fe0436717acc2f
|
{
"arc:challenge": 61.9,
"hellaswag": 83.8,
"hendrycksTest": 61.7,
"truthfulqa:mc": 46.6
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 63.5 |
blueapple8259/TinyStories-Alpaca
|
main
|
18e0bde7e72e477757832f0624a0410efc066216
|
{
"arc:challenge": 24,
"hellaswag": 24.9,
"hendrycksTest": 23.4,
"truthfulqa:mc": 46.7
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 29.8 |
DanielSc4/RedPajama-INCITE-Chat-3B-v1-FT-LoRA-8bit-test1
|
main
|
f477d24b00e05fe4c5f8d5f933080994cfd90e4e
|
{
"arc:challenge": 38.7,
"hellaswag": 63.5,
"hendrycksTest": 25.2,
"truthfulqa:mc": 36.1
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 40.9 |
DanielSc4/RedPajama-INCITE-Chat-3B-v1-RL-LoRA-8bit-test1
|
main
|
a2ee88a9fa1c9ad41e0a8c15217a4b1230ec33c8
|
{
"arc:challenge": 41.3,
"hellaswag": 66.8,
"hendrycksTest": 26.1,
"truthfulqa:mc": 35
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 42.3 |
THUDM/chatglm2-6b
|
main
|
162b620e3078b03eefff94eb5f762d4093425fb5
|
{
"arc:challenge": 38.8,
"hellaswag": 59,
"hendrycksTest": 46.7,
"truthfulqa:mc": 48.1
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 48.2 |
Salesforce/codegen-6B-multi
|
main
|
2d58b1e73791e8f0be7ea59c2720dccb6f4d0f06
|
{
"arc:challenge": 27.2,
"hellaswag": 41.1,
"hendrycksTest": 25.7,
"truthfulqa:mc": 45.7
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 34.9 |
mrm8488/llama-2-coder-7b
|
main
|
f21c0d5e3f9f8c5addf093358e6885afa9602296
|
{
"arc:challenge": 54,
"hellaswag": 78.4,
"hendrycksTest": 46.3,
"truthfulqa:mc": 38.5
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 54.3 |
mrm8488/mistral-7b-ft-h4-no_robots_instructions
|
main
|
785446da9a53ceae48795069bf7ccaf46a91a5ba
|
{
"arc:challenge": 60.9,
"hellaswag": 83.2,
"hendrycksTest": 63.7,
"truthfulqa:mc": 43.6
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 62.8 |
Riiid/sheep-duck-llama-2
|
main
|
e196dd0fe1d604c4975d972b177b09e4f1572cd5
|
{
"arc:challenge": 72.4,
"hellaswag": 87.8,
"hendrycksTest": 70.8,
"truthfulqa:mc": 63.8
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 73.7 |
Riiid/sheep-duck-llama-2-70b-v1.1
|
main
|
a8fc5c02c995733af6339ec882bef4ed93db1e8f
|
{
"arc:challenge": 73,
"hellaswag": 87.8,
"hendrycksTest": 70.8,
"truthfulqa:mc": 64.6
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 74.1 |
PSanni/Deer-3b
|
main
|
53ea8f8862fc1820f0cd31f62953b7290fd79867
|
{
"arc:challenge": 38.5,
"hellaswag": 57.4,
"hendrycksTest": 25.6,
"truthfulqa:mc": 40
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 40.4 |
w601sxs/b1ade-1b
|
main
|
b4b0fd71589e6590089e1ec14a840ecab10894ae
|
{
"arc:challenge": 28.6,
"hellaswag": 46.1,
"hendrycksTest": 25.1,
"truthfulqa:mc": 41.3
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 35.3 |
bofenghuang/vigogne-13b-chat
|
main
|
27002e974774c3599e6a4d731dd44e68b9e41f92
|
{
"arc:challenge": 58.6,
"hellaswag": 80.9,
"hendrycksTest": 47.8,
"truthfulqa:mc": 48.7
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 59 |
bofenghuang/vigogne-33b-instruct
|
main
|
9c2b558b888e0ef8b4a72e0771db72a06a5c8474
|
{
"arc:challenge": 63.1,
"hellaswag": 85,
"hendrycksTest": 58.3,
"truthfulqa:mc": 52.1
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 64.6 |
bofenghuang/vigostral-7b-chat
|
main
|
969fbfc7a91f53c8562a2c48a3c24dd3745d5a97
|
{
"arc:challenge": 62.6,
"hellaswag": 84.3,
"hendrycksTest": 63.5,
"truthfulqa:mc": 49.2
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 64.9 |
u-chom/preded-title-amazongoogle-abtbuy
|
main
|
ab36321d76775d6e276d157e27de23854d21be3a
|
{
"arc:challenge": 50.9,
"hellaswag": 78.1,
"hendrycksTest": 38,
"truthfulqa:mc": 41.6
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 52.2 |
yeen214/test_llama2_7b
|
main
|
69a4886f51ed752216cdd7f41a584d14240126f9
|
{
"arc:challenge": 53.1,
"hellaswag": 78.6,
"hendrycksTest": 46.9,
"truthfulqa:mc": 38.8
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 54.4 |
yeen214/llama2_7b_merge_orcafamily
|
main
|
fb65f697de632f2f3fef57fc3cd12fb5e4913a89
|
{
"arc:challenge": 56.9,
"hellaswag": 81.2,
"hendrycksTest": 51.5,
"truthfulqa:mc": 49.7
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 59.8 |
DevaMalla/llama_7b_qlora_cds
|
main
|
b6b5c65c5c1cce34d24c8f790bb0cc011e0f0808
|
{
"arc:challenge": 52.5,
"hellaswag": 77.8,
"hendrycksTest": 32.4,
"truthfulqa:mc": 46.1
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 52.2 |
DevaMalla/llama_7b_qlora_pds-eval
|
main
|
d20419e1d9e9a6a59ced3edf5169e8e7b3e8394c
|
{
"arc:challenge": 53.9,
"hellaswag": 78.1,
"hendrycksTest": 33,
"truthfulqa:mc": 45.6
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 52.6 |
DevaMalla/llama-base-7b
|
main
|
e01d89d8e444f7d751ea58feaf22ff8c9af69d2a
|
{
"arc:challenge": 50.9,
"hellaswag": 77.8,
"hendrycksTest": 35.7,
"truthfulqa:mc": 34.3
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 49.7 |
NekoPunchBBB/Llama-2-13b-hf_Open-Platypus-QLoRA-multigpu
|
main
|
f65029ea8f030731ace568e40bab33a7097a13de
|
{
"arc:challenge": 57.5,
"hellaswag": 82.5,
"hendrycksTest": 54.8,
"truthfulqa:mc": 43.8
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 59.6 |
NekoPunchBBB/Llama-2-13b-hf_Open-Platypus
|
main
|
c318a24121bd69509f395e17a9636093213ece21
|
{
"arc:challenge": 58.9,
"hellaswag": 82.1,
"hendrycksTest": 55,
"truthfulqa:mc": 42.8
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 59.7 |
NekoPunchBBB/Llama-2-13b-hf_Open-Platypus-8bit-att
|
main
|
83a8e51d0a72dcfbe5de13dc7ee10dc20e91602e
|
{
"arc:challenge": 57.5,
"hellaswag": 82.1,
"hendrycksTest": 54.6,
"truthfulqa:mc": 42.2
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 59.1 |
NewstaR/Koss-7B-chat
|
main
|
b1ab836d9ebf7029fafa07949b51d3838501d537
|
{
"arc:challenge": 53.7,
"hellaswag": 78.8,
"hendrycksTest": 46.7,
"truthfulqa:mc": 44
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 55.8 |
NewstaR/Starlight-13B
|
main
|
cb9fced568b1abd881133c642c427aaa488f00cc
|
{
"arc:challenge": 59.3,
"hellaswag": 82.2,
"hendrycksTest": 55.7,
"truthfulqa:mc": 37.4
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 58.6 |
NewstaR/Morningstar-13b-hf
|
main
|
2605b5b3b0ecba906ac26d39aab40f33c2ec81c9
|
{
"arc:challenge": 59,
"hellaswag": 81.9,
"hendrycksTest": 54.6,
"truthfulqa:mc": 44.1
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 59.9 |
NewstaR/Starlight-7B
|
main
|
1f7436c458ebc3d8d31b91091c1a7a48e942cd3b
|
{
"arc:challenge": 53.1,
"hellaswag": 78.6,
"hendrycksTest": 46.8,
"truthfulqa:mc": 38.8
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 54.3 |
Sao10K/Stheno-1.8-L2-13B
|
main
|
fe054ab749a69375285df40913a88bd40f1e2bf6
|
{
"arc:challenge": 63.5,
"hellaswag": 84.1,
"hendrycksTest": 58.6,
"truthfulqa:mc": 52.9
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 64.8 |
Sao10K/Mythical-Destroyer-V2-L2-13B
|
main
|
cbc8b2e4a3beafc311b9e61f8fa9f7526a77c360
|
{
"arc:challenge": 59.3,
"hellaswag": 82.7,
"hendrycksTest": 57.4,
"truthfulqa:mc": 57.1
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 64.1 |
Sao10K/Stheno-Inverted-1.2-L2-13B
|
main
|
8d2e9087093eef1c9173e167beb40b9d034a4655
|
{
"arc:challenge": 59.4,
"hellaswag": 83,
"hendrycksTest": 55.8,
"truthfulqa:mc": 51.2
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 62.4 |
Sao10K/Zephyrus-L1-33B
|
main
|
679aae34440d576456b283070371b2a15dbb948b
|
{
"arc:challenge": 64.5,
"hellaswag": 84.1,
"hendrycksTest": 57.4,
"truthfulqa:mc": 53.9
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 65 |
Sao10K/SthenoWriter-L2-13B
|
main
|
a6d9e26ab765eb170cc0aa428ee5e25b08524657
|
{
"arc:challenge": 62.3,
"hellaswag": 83.3,
"hendrycksTest": 56.1,
"truthfulqa:mc": 44.7
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 61.6 |
Sao10K/Chat-Stheno-L2-13B
|
main
|
20419fdd5b4bdcbbf075223c33b396958c48a6cf
|
{
"arc:challenge": 58.4,
"hellaswag": 81,
"hendrycksTest": 54.8,
"truthfulqa:mc": 43.3
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 59.4 |
Sao10K/Euryale-1.3-L2-70B
|
main
|
6e3ce78eb5346bf3a5ee88cd60c25dc0d73de639
|
{
"arc:challenge": 70.8,
"hellaswag": 87.9,
"hendrycksTest": 70.4,
"truthfulqa:mc": 59.9
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 72.2 |
Sao10K/Mythical-Destroyer-L2-13B
|
main
|
7c87376b201b1c30c4e12c0b7bc2f28f017ce7bc
|
{
"arc:challenge": 58.7,
"hellaswag": 82,
"hendrycksTest": 57.7,
"truthfulqa:mc": 56.3
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 63.7 |
Sao10K/BrainDerp2
|
main
|
948ee7af94a8b092807df4becfc0a8c1cd042878
|
{
"arc:challenge": 60.9,
"hellaswag": 81.9,
"hendrycksTest": 58.9,
"truthfulqa:mc": 57.2
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 64.7 |
Sao10K/Euryale-L2-70B
|
main
|
6589310a57ce5d9d6877f353f3d00cda8fa9101c
|
{
"arc:challenge": 68.9,
"hellaswag": 87.1,
"hendrycksTest": 68.8,
"truthfulqa:mc": 54.5
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 69.8 |
Sao10K/Stheno-Mix-L2-20B
|
main
|
6f9dcdaae6ef9071effe63d2107abe8b9712345b
|
{
"arc:challenge": 57.8,
"hellaswag": 79.6,
"hendrycksTest": 52.5,
"truthfulqa:mc": 51.8
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 60.4 |
nomic-ai/gpt4all-j
|
main
|
c7244e40ac6f3a52caecc96cd45481caae57ae8c
|
{
"arc:challenge": 42,
"hellaswag": 64.1,
"hendrycksTest": 28.2,
"truthfulqa:mc": 42.8
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 44.3 |
tianyil1/denas-llama2
|
main
|
b8aebc9157c0e427536aeac9132021fd66615702
|
{
"arc:challenge": 53.9,
"hellaswag": 77.8,
"hendrycksTest": 45.5,
"truthfulqa:mc": 45.2
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 55.6 |
llm-agents/tora-70b-v1.0
|
main
|
e95fd7daf017e7c414ec07ebef4ddf013c16f9a4
|
{
"arc:challenge": 67.7,
"hellaswag": 85.8,
"hendrycksTest": 69.2,
"truthfulqa:mc": 51.8
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 68.6 |
rishiraj/bloom-560m-guanaco
|
main
|
17b886fe53bdb4cea75a7f40da1e8e987124edef
|
{
"arc:challenge": 27.9,
"hellaswag": 26.1,
"hendrycksTest": 24.5,
"truthfulqa:mc": 49.4
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 32 |
ByteWave/Yi-8B-Llama
|
main
|
4f3f4d73ff3962487d1c51702b02d795bf1f33a4
|
{
"arc:challenge": 25.7,
"hellaswag": 26.8,
"hendrycksTest": 24.1,
"truthfulqa:mc": 47.8
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 31.1 |
codellama/CodeLlama-34b-hf
|
main
|
c778b02fdecd4663d2b0a42bfb340fd29969533b
|
{
"arc:challenge": 37.5,
"hellaswag": 31.8,
"hendrycksTest": 37.2,
"truthfulqa:mc": 38.9
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 36.4 |
codellama/CodeLlama-7b-Instruct-hf
|
main
|
7affc442e639b8aa1c4b3e98a10a2f45a21b8b4f
|
{
"arc:challenge": 36.5,
"hellaswag": 55.4,
"hendrycksTest": 34.5,
"truthfulqa:mc": 41.2
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 41.9 |
habanoz/TinyLlama-1.1B-intermediate-step-715k-1.5T-lr-5-2.2epochs-oasst1-top1-instruct-V1
|
main
|
74cd9eba94e77832b3081689fc5c99c37c063790
|
{
"arc:challenge": 31.5,
"hellaswag": 54.4,
"hendrycksTest": 25.5,
"truthfulqa:mc": 42.3
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 38.4 |
habanoz/TinyLlama-1.1B-intermediate-step-715k-1.5T-lr-5-3epochs-oasst1-top1-instruct-V1
|
main
|
b1ec2a1e08eb790b9a32a43053316650921af943
|
{
"arc:challenge": 31.4,
"hellaswag": 54.2,
"hendrycksTest": 25.4,
"truthfulqa:mc": 42.5
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 38.4 |
habanoz/TinyLlama-1.1B-intermediate-step-715k-1.5T-lr-5-4epochs-oasst1-top1-instruct-V1
|
main
|
7cd6d5ad10180127771e4326772eae3d40fa8445
|
{
"arc:challenge": 31.1,
"hellaswag": 54.3,
"hendrycksTest": 25.4,
"truthfulqa:mc": 41.7
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 38.1 |
habanoz/tinyllama-oasst1-top1-instruct-full-lr1-5-v0.1
|
main
|
e55b262cbd0ee52f7a4cbda136dbf1a027987c47
|
{
"arc:challenge": 32.8,
"hellaswag": 58.2,
"hendrycksTest": 26,
"truthfulqa:mc": 38.3
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 38.8 |
Deci/DeciCoder-1b
|
main
|
af2ef45ef8cbe82eb7eb4074f260412bc14c7b11
|
{
"arc:challenge": 21.2,
"hellaswag": 31.1,
"hendrycksTest": 24.3,
"truthfulqa:mc": 47.1
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 30.9 |
Andron00e/YetAnother_Open-Llama-3B-LoRA
|
main
|
52c5cb0178831908ed0571f1750fcb0f0fb125f9
|
{
"arc:challenge": 25.9,
"hellaswag": 25.8,
"hendrycksTest": 24.7,
"truthfulqa:mc": null
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | null |
Vmware/open-llama-7b-v2-open-instruct
|
main
|
b8fbe09571a71603ab517fe897a1281005060b62
|
{
"arc:challenge": 39.8,
"hellaswag": 70.3,
"hendrycksTest": 35.2,
"truthfulqa:mc": 39.5
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 46.2 |
gpt2-medium
|
main
|
f65d4965d1221eff2bcf34f53a2ba12120e18f24
|
{
"arc:challenge": 27,
"hellaswag": 40.2,
"hendrycksTest": 26.6,
"truthfulqa:mc": 40.8
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 33.7 |
TinyLlama/TinyLlama-1.1B-intermediate-step-955k-token-2T
|
main
|
f62ecb34ea0d4acea9d896040a4616a9538e2f36
|
{
"arc:challenge": 30.3,
"hellaswag": 54.8,
"hendrycksTest": 26.5,
"truthfulqa:mc": 36.1
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 36.9 |
TinyLlama/TinyLlama-1.1B-Chat-v0.6
|
main
|
bf9ae1c8bf026667e6f810768de259bb4a7f4777
|
{
"arc:challenge": 31.7,
"hellaswag": 55.8,
"hendrycksTest": 26,
"truthfulqa:mc": 34.7
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 37 |
ai-forever/rugpt3large_based_on_gpt2
|
main
|
8201db0de8deb68f25e7309db04d163b71970494
|
{
"arc:challenge": 22.6,
"hellaswag": 32.8,
"hendrycksTest": 24.9,
"truthfulqa:mc": 43.4
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 30.9 |
beomi/KoAlpaca-KoRWKV-6B
|
main
|
427ee72c4350f26de1b287a0c07b842e7d168dbc
|
{
"arc:challenge": 23.5,
"hellaswag": 31.6,
"hendrycksTest": 24.9,
"truthfulqa:mc": 39.8
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 30 |
beomi/KoRWKV-6B
|
main
|
541600070459baf0f1be9560181d5ceb77794085
|
{
"arc:challenge": 22.1,
"hellaswag": 32.2,
"hendrycksTest": 24.7,
"truthfulqa:mc": 39
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 29.5 |
gpt2
|
main
|
11c5a3d5811f50298f278a704980280950aedb10
|
{
"arc:challenge": 22.1,
"hellaswag": null,
"hendrycksTest": null,
"truthfulqa:mc": null
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 22.1 |
pszemraj/pythia-31m-simplewiki-2048
|
main
|
95d47818055661250b55144c7d9beaf05dc126d8
|
{
"arc:challenge": 22.2,
"hellaswag": 25.6,
"hendrycksTest": 23.1,
"truthfulqa:mc": 49.4
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 30.1 |
pszemraj/pythia-31m-goodwiki-deduped-2048-scratch
|
main
|
01a3cd918dd7c233bc0c3c0c948a9a462a5359d1
|
{
"arc:challenge": 23.1,
"hellaswag": 25.7,
"hendrycksTest": 23.1,
"truthfulqa:mc": 51.3
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 30.8 |
bavest/fin-llama-33b-merged
|
main
|
17114520801da7b9599fe7a9fdf238915713a59b
|
{
"arc:challenge": 65,
"hellaswag": 86.2,
"hendrycksTest": 58.7,
"truthfulqa:mc": 49.7
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 64.9 |
totally-not-an-llm/EverythingLM-13b-V2-16k
|
main
|
943f932ae1ae462389e6d2db5273158530749fff
|
{
"arc:challenge": 58.7,
"hellaswag": 80.9,
"hendrycksTest": 49.7,
"truthfulqa:mc": 47.4
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 59.2 |
totally-not-an-llm/EverythingLM-13b-V3-peft
|
main
|
7a2eed5038addcf4fa3b8dd358b45eb96134e749
|
{
"arc:challenge": 58.4,
"hellaswag": 81,
"hendrycksTest": 54.7,
"truthfulqa:mc": 53
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 61.8 |
totally-not-an-llm/EverythingLM-13b-16k
|
main
|
8456a856a8b115b05e76a7d0d945853b10ac71e2
|
{
"arc:challenge": 56.6,
"hellaswag": 80.6,
"hendrycksTest": 50.2,
"truthfulqa:mc": 47.5
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 58.7 |
totally-not-an-llm/PuddleJumper-13b
|
main
|
f3a8a475ff0c6ae37ac8ae0690980be11cac731a
|
{
"arc:challenge": 58.7,
"hellaswag": 81.2,
"hendrycksTest": 58.3,
"truthfulqa:mc": 56.4
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 63.6 |
totally-not-an-llm/EverythingLM-13b-V3-16k
|
main
|
1de9244bfadb947f80872727f76790cbc76e7142
|
{
"arc:challenge": 58.2,
"hellaswag": 80.1,
"hendrycksTest": 50.5,
"truthfulqa:mc": 45.2
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 58.5 |
dvruette/gpt-neox-20b-full-precision
|
main
|
20b347273d90da7c2c9eb4c32d4173dba862a0d2
|
{
"arc:challenge": 48.8,
"hellaswag": 74.4,
"hendrycksTest": 26.2,
"truthfulqa:mc": 36.9
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 46.6 |
dvruette/oasst-pythia-12b-6000-steps
|
main
|
e2ccc0ef8d1cc5ffc8b0e2e885f03ef50597ea8a
|
{
"arc:challenge": 45.4,
"hellaswag": 69.7,
"hendrycksTest": 26,
"truthfulqa:mc": 39.8
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 45.2 |
dvruette/oasst-pythia-6.9b-4000-steps
|
main
|
0e201b6f344ac6382dda40d389e1c9144a87d027
|
{
"arc:challenge": 41.6,
"hellaswag": 64.2,
"hendrycksTest": 26.3,
"truthfulqa:mc": 40.4
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 43.1 |
dvruette/oasst-llama-13b-1000-steps
|
main
|
d2cd599cc40db3370009f45d6caa7e486cb6d31f
|
{
"arc:challenge": 58.1,
"hellaswag": 81.5,
"hendrycksTest": 48.6,
"truthfulqa:mc": 36
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 56 |
dvruette/oasst-llama-13b-2-epochs
|
main
|
0e3796192f7edf43968541b9454ea35da4a2b1c5
|
{
"arc:challenge": 57.9,
"hellaswag": 82.4,
"hendrycksTest": 48.6,
"truthfulqa:mc": 47.3
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 59 |
dvruette/oasst-gpt-neox-20b-3000-steps
|
main
|
f0462a8b7908f61202d86e6a9a2996d8339363b5
|
{
"arc:challenge": 46.4,
"hellaswag": 72.1,
"hendrycksTest": 26.2,
"truthfulqa:mc": 35.5
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 45 |
dvruette/llama-13b-pretrained
|
main
|
c28cc0cf5a1a1bf4de96b23d06b02129dca85eb9
|
{
"arc:challenge": 56.3,
"hellaswag": 79.3,
"hendrycksTest": 47,
"truthfulqa:mc": 48.4
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 57.8 |
dvruette/oasst-pythia-12b-reference
|
main
|
c5a9b7fad884e6c45ce5d2ca551aa1c03db6865f
|
{
"arc:challenge": 43,
"hellaswag": 67.9,
"hendrycksTest": 28.3,
"truthfulqa:mc": 36.6
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 44 |
h2oai/h2ogpt-gm-oasst1-en-2048-open-llama-7b-preview-300bt-v2
|
main
|
fdc6ff469295d0aaabec8948525b70d6688728ac
|
{
"arc:challenge": 36.4,
"hellaswag": 61.4,
"hendrycksTest": 25,
"truthfulqa:mc": 37.6
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 40.1 |
h2oai/h2ogpt-gm-oasst1-multilang-1024-20b
|
main
|
b3a6bf4250a037c09e451344e2a4e987011b79de
|
{
"arc:challenge": 47.4,
"hellaswag": 72.6,
"hendrycksTest": 26.4,
"truthfulqa:mc": 34.4
}
|
9ba100d35ce48d3d4c132947464c93c861932caa
| 2023-11-23T17:28:23 | 45.2 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.