runtime error

is container with GPU support; see https://docs.nvidia.com/datacenter/cloud-native/ . WARNING 03-14 08:33:37 config.py:179] gptq quantization is not fully optimized yet. The speed can be slower than non-quantized models. INFO 03-14 08:33:37 utils.py:175] CUDA_HOME is not found in the environment. Using /usr/local/cuda as CUDA_HOME. Traceback (most recent call last): File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main return _run_code(code, main_globals, None, File "/usr/lib/python3.10/runpy.py", line 86, in _run_code exec(code, run_globals) File "/home/user/.local/lib/python3.10/site-packages/aphrodite/endpoints/openai/api_server.py", line 859, in <module> engine = AsyncAphrodite.from_engine_args(engine_args) File "/home/user/.local/lib/python3.10/site-packages/aphrodite/engine/async_aphrodite.py", line 612, in from_engine_args engine_configs = engine_args.create_engine_configs() File "/home/user/.local/lib/python3.10/site-packages/aphrodite/engine/args_tools.py", line 290, in create_engine_configs cache_config = CacheConfig(self.block_size, File "/home/user/.local/lib/python3.10/site-packages/aphrodite/common/config.py", line 298, in __init__ self._verify_cache_dtype() File "/home/user/.local/lib/python3.10/site-packages/aphrodite/common/config.py", line 320, in _verify_cache_dtype device_name = torch.cuda.get_device_name() File "/home/user/.local/lib/python3.10/site-packages/torch/cuda/__init__.py", line 423, in get_device_name return get_device_properties(device).name File "/home/user/.local/lib/python3.10/site-packages/torch/cuda/__init__.py", line 453, in get_device_properties _lazy_init() # will define _get_device_properties File "/home/user/.local/lib/python3.10/site-packages/torch/cuda/__init__.py", line 302, in _lazy_init torch._C._cuda_init() RuntimeError: Found no NVIDIA driver on your system. Please check that you have an NVIDIA GPU and installed a driver from http://www.nvidia.com/Download/index.aspx

Container logs:

Fetching error logs...