runtime error

nloads always resume when possible. If you want to force a new download, use `force_download=True`. warnings.warn( CUDA extension not installed. CUDA extension not installed. /usr/local/lib/python3.10/site-packages/huggingface_hub/file_download.py:1132: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`. warnings.warn( /usr/local/lib/python3.10/site-packages/transformers/modeling_utils.py:4371: FutureWarning: `_is_quantized_training_enabled` is going to be deprecated in transformers 4.39.0. Please use `model.hf_*********.is_trainable` instead warnings.warn( Traceback (most recent call last): File "/home/user/app/app.py", line 15, in <module> model = AutoPeftModelForCausalLM.from_pretrained( File "/usr/local/lib/python3.10/site-packages/peft/auto.py", line 104, in from_pretrained base_model = target_class.from_pretrained(base_model_path, **kwargs) File "/usr/local/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 563, in from_pretrained return model_class.from_pretrained( File "/usr/local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 3738, in from_pretrained hf_*********.postprocess_model(model) File "/usr/local/lib/python3.10/site-packages/transformers/quantizers/base.py", line 195, in postprocess_model return self._process_model_after_weight_loading(model, **kwargs) File "/usr/local/lib/python3.10/site-packages/transformers/quantizers/quantizer_gptq.py", line 80, in _process_model_after_weight_loading model = self.optimum_quantizer.post_init_model(model) File "/usr/local/lib/python3.10/site-packages/optimum/gptq/quantizer.py", line 588, in post_init_model raise ValueError( ValueError: Found modules on cpu/disk. Using Exllama or Exllamav2 backend requires all the modules to be on GPU.You can deactivate exllama backend by setting `disable_exllama=True` in the quantization config object

Container logs:

Fetching error logs...