File size: 391 Bytes
4feccb1
1
{"model": "NousResearch/Nous-Hermes-2-Mixtral-8x7B-SFT", "base_model": null, "revision": "4c06af2684730f75a6874b95e8bf6058105d9612", "precision": "bfloat16", "params": 46.703, "architectures": "MixtralForCausalLM", "weight_type": "Original", "status": "PENDING", "submitted_time": "2024-06-12T12:08:40Z", "model_type": "chat", "job_id": -1, "job_start_time": null, "use_chat_template": true}