{ "model": "fernandosola/bluearara-7B", "base_model": "", "revision": "main", "private": false, "precision": "bfloat16", "params": 7.0, "architectures": "MistralForCausalLM", "weight_type": "Original", "status": "FAILED", "submitted_time": "2024-02-17T01:32:05Z", "model_type": "🆎 : language adapted models (FP, FT, ...)", "source": "leaderboard", "job_id": -1, "job_start_time": null, "error_msg": "Failed to download and/or use the AutoModel class, trust_remote_code=True - Original Exception: Can't load the configuration of 'fernandosola/bluearara-7B'. If you were trying to load it from 'https://huggingface.co/models', make sure you don't have a local directory with the same name. Otherwise, make sure 'fernandosola/bluearara-7B' is the correct path to a directory containing a config.json file", "traceback": "Traceback (most recent call last):\n File \"/workspace/repos/llm_leaderboard/llm_leaderboard_eval_bot/evaluate_llms.py\", line 215, in wait_download_and_run_request\n raise Exception(f\"Failed to download and/or use the AutoModel class, trust_remote_code={TRUST_REMOTE_CODE} - Original Exception: {exception_msg}\")\nException: Failed to download and/or use the AutoModel class, trust_remote_code=True - Original Exception: Can't load the configuration of 'fernandosola/bluearara-7B'. If you were trying to load it from 'https://huggingface.co/models', make sure you don't have a local directory with the same name. Otherwise, make sure 'fernandosola/bluearara-7B' is the correct path to a directory containing a config.json file\n" }