runtime error

/utils/hub.py", line 409, in cached_file resolved_file = hf_hub_download( File "/home/user/.local/lib/python3.8/site-packages/huggingface_hub/utils/_validators.py", line 124, in _inner_fn return fn(*args, **kwargs) File "/home/user/.local/lib/python3.8/site-packages/huggingface_hub/file_download.py", line 1212, in hf_hub_download raise LocalEntryNotFoundError( huggingface_hub.utils._errors.LocalEntryNotFoundError: Connection error, and we cannot find the requested files in the disk cache. Please try again or make sure your Internet connection is on. During handling of the above exception, another exception occurred: Traceback (most recent call last): File "app.py", line 332, in <module> clip_model = CLIPModel.from_pretrained(clip_model_id) File "/home/user/.local/lib/python3.8/site-packages/transformers/modeling_utils.py", line 2079, in from_pretrained config, model_kwargs = cls.config_class.from_pretrained( File "/home/user/.local/lib/python3.8/site-packages/transformers/configuration_utils.py", line 538, in from_pretrained config_dict, kwargs = cls.get_config_dict(pretrained_model_name_or_path, **kwargs) File "/home/user/.local/lib/python3.8/site-packages/transformers/configuration_utils.py", line 565, in get_config_dict config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) File "/home/user/.local/lib/python3.8/site-packages/transformers/configuration_utils.py", line 620, in _get_config_dict resolved_config_file = cached_file( File "/home/user/.local/lib/python3.8/site-packages/transformers/utils/hub.py", line 443, in cached_file raise EnvironmentError( OSError: We couldn't connect to 'https://huggingface.co' to load this file, couldn't find it in the cached files and it looks like openai/clip-vit-large-patch14 is not the path to a directory containing a file named config.json. Checkout your internet connection or see how to run the library in offline mode at 'https://huggingface.co/docs/transformers/installation#offline-mode'.

Container logs:

Fetching error logs...