File size: 4,388 Bytes
e796c34
 
 
 
 
 
 
 
 
baf9d9c
e796c34
 
 
8788cb7
baf9d9c
 
fb2444c
 
e796c34
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
{
    "model": "BAAI/Aquila2-34B",
    "base_model": "",
    "revision": "main",
    "private": false,
    "precision": "bfloat16",
    "params": 34.0,
    "architectures": "LlamaForCausalLM",
    "weight_type": "Original",
    "status": "FAILED",
    "submitted_time": "2024-02-05T23:10:17Z",
    "model_type": "🟢 : pretrained",
    "source": "script",
    "job_id": 341,
    "job_start_time": "2024-04-02T07-07-26.615695",
    "error_msg": "Consistency check failed: file should be of size 9814920915 but has size 7945050153 (pytorch_model-00002-of-00007.bin).\nWe are sorry for the inconvenience. Please retry download and pass `force_download=True, resume_download=False` as argument.\nIf the issue persists, please let us know by opening an issue on https://github.com/huggingface/huggingface_hub.",
    "traceback": "Traceback (most recent call last):\n  File \"/workspace/repos/llm_leaderboard/llm_leaderboard_eval_bot/evaluate_llms.py\", line 238, in wait_download_and_run_request\n    run_request(\n  File \"/workspace/repos/llm_leaderboard/llm_leaderboard_eval_bot/evaluate_llms.py\", line 64, in run_request\n    results = run_eval_on_model(\n              ^^^^^^^^^^^^^^^^^^\n  File \"/workspace/repos/llm_leaderboard/llm_leaderboard_eval_bot/run_eval.py\", line 55, in run_eval_on_model\n    result = evaluate(\n             ^^^^^^^^^\n  File \"/workspace/repos/llm_leaderboard/llm_leaderboard_eval_bot/lm_eval_util.py\", line 145, in evaluate\n    results = evaluator.simple_evaluate(\n              ^^^^^^^^^^^^^^^^^^^^^^^^^^\n  File \"/workspace/repos/llm_leaderboard/lm-evaluation-harness-pt/lm_eval/utils.py\", line 415, in _wrapper\n    return fn(*args, **kwargs)\n           ^^^^^^^^^^^^^^^^^^^\n  File \"/workspace/repos/llm_leaderboard/lm-evaluation-harness-pt/lm_eval/evaluator.py\", line 100, in simple_evaluate\n    lm = lm_eval.api.registry.get_model(model).create_from_arg_string(\n         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n  File \"/workspace/repos/llm_leaderboard/lm-evaluation-harness-pt/lm_eval/api/model.py\", line 134, in create_from_arg_string\n    return cls(**args, **args2)\n           ^^^^^^^^^^^^^^^^^^^^\n  File \"/workspace/repos/llm_leaderboard/lm-evaluation-harness-pt/lm_eval/models/huggingface.py\", line 297, in __init__\n    self._create_model(\n  File \"/workspace/repos/llm_leaderboard/lm-evaluation-harness-pt/lm_eval/models/huggingface.py\", line 608, in _create_model\n    self._model = self.AUTO_MODEL_CLASS.from_pretrained(\n                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n  File \"/root/miniconda3/envs/torch21/lib/python3.11/site-packages/transformers/models/auto/auto_factory.py\", line 561, in from_pretrained\n    return model_class.from_pretrained(\n           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n  File \"/root/miniconda3/envs/torch21/lib/python3.11/site-packages/transformers/modeling_utils.py\", line 3264, in from_pretrained\n    resolved_archive_file, sharded_metadata = get_checkpoint_shard_files(\n                                              ^^^^^^^^^^^^^^^^^^^^^^^^^^^\n  File \"/root/miniconda3/envs/torch21/lib/python3.11/site-packages/transformers/utils/hub.py\", line 1038, in get_checkpoint_shard_files\n    cached_filename = cached_file(\n                      ^^^^^^^^^^^^\n  File \"/root/miniconda3/envs/torch21/lib/python3.11/site-packages/transformers/utils/hub.py\", line 398, in cached_file\n    resolved_file = hf_hub_download(\n                    ^^^^^^^^^^^^^^^^\n  File \"/root/miniconda3/envs/torch21/lib/python3.11/site-packages/huggingface_hub/utils/_validators.py\", line 118, in _inner_fn\n    return fn(*args, **kwargs)\n           ^^^^^^^^^^^^^^^^^^^\n  File \"/root/miniconda3/envs/torch21/lib/python3.11/site-packages/huggingface_hub/file_download.py\", line 1457, in hf_hub_download\n    http_get(\n  File \"/root/miniconda3/envs/torch21/lib/python3.11/site-packages/huggingface_hub/file_download.py\", line 552, in http_get\n    raise EnvironmentError(\nOSError: Consistency check failed: file should be of size 9814920915 but has size 7945050153 (pytorch_model-00002-of-00007.bin).\nWe are sorry for the inconvenience. Please retry download and pass `force_download=True, resume_download=False` as argument.\nIf the issue persists, please let us know by opening an issue on https://github.com/huggingface/huggingface_hub.\n",
    "main_language": "?"
}