Mobile-MMLU-Challenge / leaderboard_pro.csv
SondosMB's picture
Update leaderboard_pro.csv
6954e0f verified
raw
history blame contribute delete
834 Bytes
Model Name,Overall Accuracy,Correct Predictions,Total Questions,Timestamp,Team Name
Nemotron-Mini-4B-Instruct,30.8,,9497,,
Llama-3.2-1B-Instruct,31.1,,9497,,
Gemma-2-2b-it,31.2,,9497,,
Falcon3-3B-Instruct,37.2,,9497,,
Granite-3.1-3b-a800m-instruct,39.4,,9497,,
Llama-3.2-3B-Instruct,42,,9497,,
Granite-3.1-2b-instruct,42.2,,9497,,
Qwen2.5-1.5B-Instruct,43,,9497,,
Exaone-3.5-2.4B-Instruct,47.7,,9497,,
Phi-3.5-mini-instruct,54.8,,9497,,
Qwen2.5-3B-Instruct,60.6,,9497,,
Olmo-2-1124-7B-Instruct,42.9,,9497,,
Falcon3-7B-Instruct,46.8,,9497,,
Falcon3-10B-Instruct,49.1,,9497,,
Yi-1.5-6B-Chat,54.7,,9497,,
Llama-3.1-8B-Instruct,57.1,,9497,,
Granite-3.1-8b-instruct,57.7,,9497,,
Internlm2.5-7b-chat,58.6,,9497,,
Ministral-8B-Instruct-2410,63.6,,9497,,
Yi-1.5-9B-Chat,67.7,,9497,,
Qwen2.5-7B-Instruct,68.4,,9497,,
Gemma-2-9b-it,69.1,,9497,,