runtime error

cache prediction = await Context.root_block.process_api( File "/home/user/.local/lib/python3.10/site-packages/gradio/blocks.py", line 1664, in process_api result = await self.call_function( File "/home/user/.local/lib/python3.10/site-packages/gradio/blocks.py", line 1205, in call_function prediction = await anyio.to_thread.run_sync( File "/home/user/.local/lib/python3.10/site-packages/anyio/to_thread.py", line 56, in run_sync return await get_async_backend().run_sync_in_worker_thread( File "/home/user/.local/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 2144, in run_sync_in_worker_thread return await future File "/home/user/.local/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 851, in run result = context.run(func, *args) File "/home/user/.local/lib/python3.10/site-packages/gradio/utils.py", line 690, in wrapper response = f(*args, **kwargs) File "/home/user/.local/lib/python3.10/site-packages/gradio/external.py", line 352, in query_huggingface_inference_endpoints data = fn(*data) # type: ignore File "/home/user/.local/lib/python3.10/site-packages/huggingface_hub/inference/_client.py", line 301, in audio_classification response = self.post(data=audio, model=model, task="audio-classification") File "/home/user/.local/lib/python3.10/site-packages/huggingface_hub/inference/_client.py", line 242, in post hf_raise_for_status(response) File "/home/user/.local/lib/python3.10/site-packages/huggingface_hub/utils/_errors.py", line 362, in hf_raise_for_status raise HfHubHTTPError(str(e), response=response) from e huggingface_hub.utils._errors.HfHubHTTPError: 500 Server Error: Internal Server Error for url: https://api-inference.huggingface.co/models/mispeech/ced-base (Request ID: pXefOk8hvX6udyl4ZbNyU) The checkpoint you are trying to load has model type `ced` but Transformers does not recognize this architecture. This could be because of an issue with the checkpoint, or because your version of Transformers is out of date.

Container logs:

Fetching error logs...