runtime error
Exit code: 1. Reason: _chain.py", line 10, in __init__ self.vector_store = VectorStore() File "/home/user/app/modules/rag/vector_store.py", line 13, in __init__ self.embedding_function = SentenceTransformerEmbeddings(model_name="all-MiniLM-L6-v2") File "/usr/local/lib/python3.10/site-packages/langchain_core/_api/deprecation.py", line 223, in warn_if_direct_instance return wrapped(self, *args, **kwargs) File "/usr/local/lib/python3.10/site-packages/langchain_community/embeddings/huggingface.py", line 92, in __init__ self.client = sentence_transformers.SentenceTransformer( File "/usr/local/lib/python3.10/site-packages/sentence_transformers/SentenceTransformer.py", line 339, in __init__ modules = self._load_auto_model( File "/usr/local/lib/python3.10/site-packages/sentence_transformers/SentenceTransformer.py", line 2060, in _load_auto_model transformer_model = Transformer( File "/usr/local/lib/python3.10/site-packages/sentence_transformers/models/Transformer.py", line 86, in __init__ config, is_peft_model = self._load_config(model_name_or_path, cache_dir, backend, config_args) File "/usr/local/lib/python3.10/site-packages/sentence_transformers/models/Transformer.py", line 127, in _load_config find_adapter_config_file( File "/usr/local/lib/python3.10/site-packages/transformers/utils/peft_utils.py", line 88, in find_adapter_config_file adapter_cached_filename = cached_file( File "/usr/local/lib/python3.10/site-packages/transformers/utils/hub.py", line 321, in cached_file file = cached_files(path_or_repo_id=path_or_repo_id, filenames=[filename], **kwargs) File "/usr/local/lib/python3.10/site-packages/transformers/utils/hub.py", line 510, in cached_files raise OSError( OSError: sentence-transformers/all-MiniLM-L6-v2 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models' If this is a private repository, make sure to pass a token having permission to this repo either by logging in with `hf auth login` or by passing `token=<your_token>`
Container logs:
Fetching error logs...