runtime error

", line 375, in synchronize_async return fsspec.asyn.sync(fsspec.asyn.get_loop(), func, *args, **kwargs) File "/home/user/.local/lib/python3.8/site-packages/fsspec/asyn.py", line 103, in sync raise return_result File "/home/user/.local/lib/python3.8/site-packages/fsspec/asyn.py", line 56, in _runner result[0] = await coro File "/home/user/.local/lib/python3.8/site-packages/gradio/blocks.py", line 1015, in process_api result = await self.call_function(fn_index, inputs, iterator, request) File "/home/user/.local/lib/python3.8/site-packages/gradio/blocks.py", line 856, in call_function prediction = await anyio.to_thread.run_sync( File "/home/user/.local/lib/python3.8/site-packages/anyio/to_thread.py", line 56, in run_sync return await get_async_backend().run_sync_in_worker_thread( File "/home/user/.local/lib/python3.8/site-packages/anyio/_backends/_asyncio.py", line 2134, in run_sync_in_worker_thread return await future File "/home/user/.local/lib/python3.8/site-packages/anyio/_backends/_asyncio.py", line 851, in run result = context.run(func, *args) File "/home/user/.local/lib/python3.8/site-packages/gradio/external.py", line 282, in query_huggingface_api raise ValueError( ValueError: Could not complete request to HuggingFace API, Status Code: 500, Error: unknown error, Warnings: ['CUDA out of memory. Tried to allocate 20.00 MiB (GPU 0; 14.76 GiB total capacity; 2.05 GiB already allocated; 14.75 MiB free; 2.09 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF', 'There was an inference error: CUDA out of memory. Tried to allocate 20.00 MiB (GPU 0; 14.76 GiB total capacity; 2.05 GiB already allocated; 14.75 MiB free; 2.09 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF']

Container logs:

Fetching error logs...