Update status of Danielbrdz/Barcenas-14b-Phi-3-medium-ORPO_eval_request_False_float16_Original to FINISHED
Browse files
Danielbrdz/Barcenas-14b-Phi-3-medium-ORPO_eval_request_False_float16_Original.json
CHANGED
|
@@ -8,10 +8,24 @@
|
|
| 8 |
"architectures": "MistralForCausalLM",
|
| 9 |
"weight_type": "Original",
|
| 10 |
"main_language": "English",
|
| 11 |
-
"status": "
|
| 12 |
"submitted_time": "2024-07-31T19:07:39Z",
|
| 13 |
"model_type": "💬 : chat (RLHF, DPO, IFT, ...)",
|
| 14 |
"source": "leaderboard",
|
| 15 |
"job_id": 977,
|
| 16 |
-
"job_start_time": "2024-08-08T02-43-35.640819"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 17 |
}
|
|
|
|
| 8 |
"architectures": "MistralForCausalLM",
|
| 9 |
"weight_type": "Original",
|
| 10 |
"main_language": "English",
|
| 11 |
+
"status": "FINISHED",
|
| 12 |
"submitted_time": "2024-07-31T19:07:39Z",
|
| 13 |
"model_type": "💬 : chat (RLHF, DPO, IFT, ...)",
|
| 14 |
"source": "leaderboard",
|
| 15 |
"job_id": 977,
|
| 16 |
+
"job_start_time": "2024-08-08T02-43-35.640819",
|
| 17 |
+
"eval_version": "1.1.0",
|
| 18 |
+
"result_metrics": {
|
| 19 |
+
"enem_challenge": 0.7319804058782365,
|
| 20 |
+
"bluex": 0.6578581363004172,
|
| 21 |
+
"oab_exams": 0.510250569476082,
|
| 22 |
+
"assin2_rte": 0.9259642329554806,
|
| 23 |
+
"assin2_sts": 0.714480317302389,
|
| 24 |
+
"faquad_nli": 0.6906170752324599,
|
| 25 |
+
"hatebr_offensive": 0.8460180802244769,
|
| 26 |
+
"portuguese_hate_speech": 0.7355214633181597,
|
| 27 |
+
"tweetsentbr": 0.5025508930131336
|
| 28 |
+
},
|
| 29 |
+
"result_metrics_average": 0.7016934637445373,
|
| 30 |
+
"result_metrics_npm": 0.5569268035545414
|
| 31 |
}
|