Update status of Weni/WeniGPT-Mistral-7B-instructBase_eval_request_False_float16_Original to FINISHED
Browse files
Weni/WeniGPT-Mistral-7B-instructBase_eval_request_False_float16_Original.json
CHANGED
@@ -8,10 +8,24 @@
|
|
8 |
"architectures": "MistralForCausalLM",
|
9 |
"weight_type": "Original",
|
10 |
"main_language": "Other",
|
11 |
-
"status": "
|
12 |
"submitted_time": "2024-04-09T02:59:01Z",
|
13 |
"model_type": "💬 : chat models (RLHF, DPO, IFT, ...)",
|
14 |
"source": "leaderboard",
|
15 |
"job_id": 433,
|
16 |
-
"job_start_time": "2024-04-13T05-37-06.823274"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
17 |
}
|
|
|
8 |
"architectures": "MistralForCausalLM",
|
9 |
"weight_type": "Original",
|
10 |
"main_language": "Other",
|
11 |
+
"status": "FINISHED",
|
12 |
"submitted_time": "2024-04-09T02:59:01Z",
|
13 |
"model_type": "💬 : chat models (RLHF, DPO, IFT, ...)",
|
14 |
"source": "leaderboard",
|
15 |
"job_id": 433,
|
16 |
+
"job_start_time": "2024-04-13T05-37-06.823274",
|
17 |
+
"eval_version": "1.1.0",
|
18 |
+
"result_metrics": {
|
19 |
+
"enem_challenge": 0.3261021693491952,
|
20 |
+
"bluex": 0.23504867872044508,
|
21 |
+
"oab_exams": 0.2883826879271071,
|
22 |
+
"assin2_rte": 0.6142066914966826,
|
23 |
+
"assin2_sts": 0.40022445504959214,
|
24 |
+
"faquad_nli": 0.344957407970794,
|
25 |
+
"hatebr_offensive": 0.44162029377851314,
|
26 |
+
"portuguese_hate_speech": 0.39714398155956604,
|
27 |
+
"tweetsentbr": 0.5118077502764405
|
28 |
+
},
|
29 |
+
"result_metrics_average": 0.39549934623648175,
|
30 |
+
"result_metrics_npm": 0.07177075983532584
|
31 |
}
|