runtime error

config.json: 0%| | 0.00/2.37k [00:00<?, ?B/s] config.json: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 2.37k/2.37k [00:00<00:00, 13.4MB/s] model.bin: 0%| | 0.00/484M [00:00<?, ?B/s] model.bin: 13%|β–ˆβ–Ž | 64.1M/484M [00:01<00:06, 63.3MB/s] model.bin: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰| 484M/484M [00:01<00:00, 297MB/s] tokenizer.json: 0%| | 0.00/2.20M [00:00<?, ?B/s] tokenizer.json: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 2.20M/2.20M [00:00<00:00, 27.6MB/s] vocabulary.txt: 0%| | 0.00/460k [00:00<?, ?B/s] vocabulary.txt: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 460k/460k [00:00<00:00, 55.3MB/s] Traceback (most recent call last): File "/home/user/app/app.py", line 34, in <module> model = WhisperModel(model_size, device="cuda", compute_type="float16") File "/home/user/.local/lib/python3.10/site-packages/faster_whisper/transcribe.py", line 130, in __init__ self.model = ctranslate2.models.Whisper( RuntimeError: CUDA failed with error CUDA driver version is insufficient for CUDA runtime version

Container logs:

Fetching error logs...