llm_pt_leaderboard_requests
/
JJhooww
/Mistral_Relora_Step2k_eval_request_False_float16_Original.json
![eduagarcia's picture](https://cdn-avatars.huggingface.co/v1/production/uploads/620bf258ce371f5bad535d3a/p7H2GP_QTZqRX1Lcu3DtR.jpeg)
Update status of JJhooww/Mistral_Relora_Step2k_eval_request_False_float16_Original to FINISHED
adbb125
verified
{ | |
"model": "JJhooww/Mistral_Relora_Step2k", | |
"base_model": "", | |
"revision": "main", | |
"private": false, | |
"precision": "float16", | |
"params": 7.242, | |
"architectures": "MistralForCausalLM", | |
"weight_type": "Original", | |
"status": "FINISHED", | |
"submitted_time": "2024-03-08T02:22:23Z", | |
"model_type": "๐ : language adapted models (FP, FT, ...)", | |
"source": "leaderboard", | |
"job_id": 320, | |
"job_start_time": "2024-03-09T08-42-21.029909", | |
"eval_version": "1.1.0", | |
"result_metrics": { | |
"enem_challenge": 0.615815255423373, | |
"bluex": 0.5257301808066759, | |
"oab_exams": 0.3981776765375854, | |
"assin2_rte": 0.9113496854193482, | |
"assin2_sts": 0.7074610038971542, | |
"faquad_nli": 0.6526577185427341, | |
"hatebr_offensive": 0.8133973664850924, | |
"portuguese_hate_speech": 0.6536416538696902, | |
"tweetsentbr": 0.5193585604823832 | |
}, | |
"result_metrics_average": 0.6441765668293375, | |
"result_metrics_npm": 0.4715334429973541 | |
} |