{ "model": "J-AI/Phi_3-CREWAI-PTBR", "base_model": "", "revision": "main", "private": false, "precision": "float16", "params": 0, "architectures": "MistralForCausalLM", "weight_type": "Original", "main_language": "Portuguese", "status": "FAILED", "submitted_time": "2024-05-15T20:00:58Z", "model_type": "🔶 : fine-tuned/fp on domain-specific datasets", "source": "leaderboard", "job_id": -1, "job_start_time": null, "error_msg": "Failed to download and/or use the AutoModel class, trust_remote_code=True - Original Exception: Can't load the configuration of 'J-AI/Phi_3-CREWAI-PTBR'. If you were trying to load it from 'https://huggingface.co/models', make sure you don't have a local directory with the same name. Otherwise, make sure 'J-AI/Phi_3-CREWAI-PTBR' is the correct path to a directory containing a config.json file", "traceback": "Traceback (most recent call last):\n File \"/workspace/repos/llm_leaderboard/llm_leaderboard_eval_bot/evaluate_llms.py\", line 204, in wait_download_and_run_request\n raise Exception(f\"Failed to download and/or use the AutoModel class, trust_remote_code={TRUST_REMOTE_CODE} - Original Exception: {exception_msg}\")\nException: Failed to download and/or use the AutoModel class, trust_remote_code=True - Original Exception: Can't load the configuration of 'J-AI/Phi_3-CREWAI-PTBR'. If you were trying to load it from 'https://huggingface.co/models', make sure you don't have a local directory with the same name. Otherwise, make sure 'J-AI/Phi_3-CREWAI-PTBR' is the correct path to a directory containing a config.json file\n" }