{ "model": "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO", "base_model": "", "revision": "main", "private": false, "precision": "bfloat16", "params": 46.703, "architectures": "MixtralForCausalLM", "weight_type": "Original", "status": "FINISHED", "submitted_time": "2024-02-27T00:38:29Z", "model_type": "💬 : chat models (RLHF, DPO, IFT, ...)", "source": "leaderboard", "job_id": 296, "job_start_time": "2024-03-07T12-23-27.230508", "eval_version": "1.1.0", "result_metrics": { "enem_challenge": 0.6641007697690693, "bluex": 0.5535465924895688, "oab_exams": 0.47289293849658315, "assin2_rte": 0.9023645725471283, "assin2_sts": 0.734857329244095, "faquad_nli": 0.7498307874026198, "hatebr_offensive": 0.7666031472700376, "portuguese_hate_speech": 0.5860877435617644, "tweetsentbr": 0.6190805774400715 }, "result_metrics_average": 0.6721516064689931, "result_metrics_npm": 0.5058743157671289, "main_language": "English" }