runtime error

Exit code: 1. Reason: b/python3.11/site-packages/chainlit/config.py", line 280, in load_module spec.loader.exec_module(module) File "<frozen importlib._bootstrap_external>", line 940, in exec_module File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed File "/home/user/app/app.py", line 132, in <module> print(hf_llm.invoke("What is Deep Learning?")) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/user/.local/lib/python3.11/site-packages/langchain_core/language_models/llms.py", line 276, in invoke self.generate_prompt( File "/home/user/.local/lib/python3.11/site-packages/langchain_core/language_models/llms.py", line 633, in generate_prompt return self.generate(prompt_strings, stop=stop, callbacks=callbacks, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/user/.local/lib/python3.11/site-packages/langchain_core/language_models/llms.py", line 803, in generate output = self._generate_helper( ^^^^^^^^^^^^^^^^^^^^^^ File "/home/user/.local/lib/python3.11/site-packages/langchain_core/language_models/llms.py", line 670, in _generate_helper raise e File "/home/user/.local/lib/python3.11/site-packages/langchain_core/language_models/llms.py", line 657, in _generate_helper self._generate( File "/home/user/.local/lib/python3.11/site-packages/langchain_core/language_models/llms.py", line 1317, in _generate self._call(prompt, stop=stop, run_manager=run_manager, **kwargs) File "/home/user/.local/lib/python3.11/site-packages/langchain_huggingface/llms/huggingface_endpoint.py", line 258, in _call response = self.client.post( ^^^^^^^^^^^^^^^^^ File "/home/user/.local/lib/python3.11/site-packages/huggingface_hub/inference/_client.py", line 283, in post raise InferenceTimeoutError( huggingface_hub.errors.InferenceTimeoutError: Model not loaded on the server: https://l929qnehi21k01ft.us-east-1.aws.endpoints.huggingface.cloud. Please retry with a higher timeout (current: 120).

Container logs:

Fetching error logs...