runtime error
Exit code: 1. Reason: οΏ½βββββ | 3.55G/4.92G [00:03<00:01, 1.28GB/s][A model-00003-of-00004.safetensors: 100%|ββββββββββ| 4.92G/4.92G [00:03<00:00, 1.31GB/s] Downloading shards: 75%|ββββββββ | 3/4 [00:12<00:04, 4.23s/it][A model-00004-of-00004.safetensors: 0%| | 0.00/1.17G [00:00<?, ?B/s][A model-00004-of-00004.safetensors: 48%|βββββ | 560M/1.17G [00:01<00:01, 559MB/s][A model-00004-of-00004.safetensors: 100%|ββββββββββ| 1.17G/1.17G [00:01<00:00, 842MB/s] Downloading shards: 100%|ββββββββββ| 4/4 [00:14<00:00, 3.22s/it][A Downloading shards: 100%|ββββββββββ| 4/4 [00:14<00:00, 3.62s/it] Loading checkpoint shards: 0%| | 0/4 [00:00<?, ?it/s][A Loading checkpoint shards: 100%|ββββββββββ| 4/4 [00:01<00:00, 3.40it/s][A Loading checkpoint shards: 100%|ββββββββββ| 4/4 [00:01<00:00, 3.40it/s] generation_config.json: 0%| | 0.00/184 [00:00<?, ?B/s][A generation_config.json: 100%|ββββββββββ| 184/184 [00:00<00:00, 1.17MB/s] Traceback (most recent call last): File "/home/user/app/app.py", line 43, in <module> model = AutoModelForCausalLM.from_pretrained( File "/usr/local/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 564, in from_pretrained return model_class.from_pretrained( File "/usr/local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 3977, in from_pretrained dispatch_model(model, **device_map_kwargs) File "/usr/local/lib/python3.10/site-packages/accelerate/big_modeling.py", line 501, in dispatch_model model.to(device) File "/usr/local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 2826, in to raise ValueError( ValueError: `.to` is not supported for `4-bit` or `8-bit` bitsandbytes models. Please use the model as it is, since the model has already been set to the correct devices and casted to the correct `dtype`.
Container logs:
Fetching error logs...