runtime error
Exit code: 1. Reason: 100%|ββββββββββ| 281/281 [00:00<00:00, 1.61MB/s] config.json: 0%| | 0.00/1.05k [00:00<?, ?B/s][A config.json: 100%|ββββββββββ| 1.05k/1.05k [00:00<00:00, 6.86MB/s] The installed version of bitsandbytes was compiled without GPU support. 8-bit optimizers, 8-bit multiplication, and GPU quantization are unavailable. CUDA is required but not available for bitsandbytes. Please consider installing the multi-platform enabled version of bitsandbytes, which is currently a work in progress. Please check currently supported platforms and installation instructions at https://huggingface.co/docs/bitsandbytes/main/en/installation#multi-backend Traceback (most recent call last): File "/home/user/app/app.py", line 18, in <module> model = AutoModelForCausalLM.from_pretrained( File "/usr/local/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 564, in from_pretrained return model_class.from_pretrained( File "/usr/local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 3669, in from_pretrained hf_quantizer.validate_environment( File "/usr/local/lib/python3.10/site-packages/transformers/quantizers/quantizer_bnb_4bit.py", line 82, in validate_environment validate_bnb_backend_availability(raise_exception=True) File "/usr/local/lib/python3.10/site-packages/transformers/integrations/bitsandbytes.py", line 558, in validate_bnb_backend_availability return _validate_bnb_cuda_backend_availability(raise_exception) File "/usr/local/lib/python3.10/site-packages/transformers/integrations/bitsandbytes.py", line 536, in _validate_bnb_cuda_backend_availability raise RuntimeError(log_msg) RuntimeError: CUDA is required but not available for bitsandbytes. Please consider installing the multi-platform enabled version of bitsandbytes, which is currently a work in progress. Please check currently supported platforms and installation instructions at https://huggingface.co/docs/bitsandbytes/main/en/installation#multi-backend
Container logs:
Fetching error logs...