runtime error

/2.04G [00:04<00:02, 364MB/s] model-00001-of-00002.safetensors: 94%|█████████▍| 1.91G/2.04G [00:05<00:00, 473MB/s] model-00001-of-00002.safetensors: 100%|█████████▉| 2.04G/2.04G [00:05<00:00, 389MB/s] Downloading shards: 50%|█████ | 1/2 [00:05<00:05, 5.71s/it] model-00002-of-00002.safetensors: 0%| | 0.00/1.63G [00:00<?, ?B/s] model-00002-of-00002.safetensors: 4%|▍ | 73.4M/1.63G [00:01<00:21, 72.0MB/s] model-00002-of-00002.safetensors: 13%|█▎ | 220M/1.63G [00:02<00:13, 107MB/s]  model-00002-of-00002.safetensors: 48%|████▊ | 786M/1.63G [00:03<00:02, 304MB/s] model-00002-of-00002.safetensors: 67%|██████▋ | 1.10G/1.63G [00:04<00:02, 247MB/s] model-00002-of-00002.safetensors: 88%|████████▊ | 1.45G/1.63G [00:05<00:00, 274MB/s] model-00002-of-00002.safetensors: 100%|█████████▉| 1.63G/1.63G [00:07<00:00, 233MB/s] Downloading shards: 100%|██████████| 2/2 [00:13<00:00, 6.64s/it] Downloading shards: 100%|██████████| 2/2 [00:13<00:00, 6.50s/it] Traceback (most recent call last): File "/home/user/app/app.py", line 8, in <module> model = AutoModelForCausalLM.from_pretrained("Qwen/Qwen-1_8B-Chat", device_map="auto", trust_remote_code=True).eval() File "/home/user/.local/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 511, in from_pretrained return model_class.from_pretrained( File "/home/user/.local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 3091, in from_pretrained ) = cls._load_pretrained_model( File "/home/user/.local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 3408, in _load_pretrained_model offload_index = { File "/home/user/.local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 3411, in <dictcomp> if param_device_map[p] == "disk" KeyError: 'lm_head.weight'

Container logs:

Fetching error logs...