Merge pull request #92 from joshuasundance-swca/models
Browse files
langchain-streamlit-demo/defaults.py
CHANGED
@@ -5,13 +5,17 @@ from collections import namedtuple
|
|
5 |
MODEL_DICT = {
|
6 |
"gpt-3.5-turbo": "OpenAI",
|
7 |
"gpt-4": "OpenAI",
|
8 |
-
"
|
|
|
|
|
9 |
"claude-2": "Anthropic",
|
|
|
10 |
"meta-llama/Llama-2-7b-chat-hf": "Anyscale Endpoints",
|
11 |
"meta-llama/Llama-2-13b-chat-hf": "Anyscale Endpoints",
|
12 |
"meta-llama/Llama-2-70b-chat-hf": "Anyscale Endpoints",
|
13 |
"codellama/CodeLlama-34b-Instruct-hf": "Anyscale Endpoints",
|
14 |
"mistralai/Mistral-7B-Instruct-v0.1": "Anyscale Endpoints",
|
|
|
15 |
"Azure OpenAI": "Azure OpenAI",
|
16 |
}
|
17 |
|
|
|
5 |
MODEL_DICT = {
|
6 |
"gpt-3.5-turbo": "OpenAI",
|
7 |
"gpt-4": "OpenAI",
|
8 |
+
"gpt-4-1106-preview": "OpenAI",
|
9 |
+
"claude-instant-1": "Anthropic",
|
10 |
+
"claude-instant-1.2": "Anthropic",
|
11 |
"claude-2": "Anthropic",
|
12 |
+
"claude-2.1": "Anthropic",
|
13 |
"meta-llama/Llama-2-7b-chat-hf": "Anyscale Endpoints",
|
14 |
"meta-llama/Llama-2-13b-chat-hf": "Anyscale Endpoints",
|
15 |
"meta-llama/Llama-2-70b-chat-hf": "Anyscale Endpoints",
|
16 |
"codellama/CodeLlama-34b-Instruct-hf": "Anyscale Endpoints",
|
17 |
"mistralai/Mistral-7B-Instruct-v0.1": "Anyscale Endpoints",
|
18 |
+
"HuggingFaceH4/zephyr-7b-beta": "Anyscale Endpoints",
|
19 |
"Azure OpenAI": "Azure OpenAI",
|
20 |
}
|
21 |
|
langchain-streamlit-demo/llm_resources.py
CHANGED
@@ -110,6 +110,7 @@ def get_llm(
|
|
110 |
|
111 |
return None
|
112 |
|
|
|
113 |
def get_texts_and_retriever(
|
114 |
uploaded_file_bytes: bytes,
|
115 |
openai_api_key: str,
|
@@ -145,7 +146,7 @@ def get_texts_and_retriever(
|
|
145 |
retrievers=[bm25_retriever, faiss_retriever],
|
146 |
weights=[0.5, 0.5],
|
147 |
)
|
148 |
-
|
149 |
return texts, ensemble_retriever
|
150 |
|
151 |
|
|
|
110 |
|
111 |
return None
|
112 |
|
113 |
+
|
114 |
def get_texts_and_retriever(
|
115 |
uploaded_file_bytes: bytes,
|
116 |
openai_api_key: str,
|
|
|
146 |
retrievers=[bm25_retriever, faiss_retriever],
|
147 |
weights=[0.5, 0.5],
|
148 |
)
|
149 |
+
|
150 |
return texts, ensemble_retriever
|
151 |
|
152 |
|
requirements.txt
CHANGED
@@ -1,15 +1,15 @@
|
|
1 |
-
anthropic==0.7.
|
2 |
faiss-cpu==1.7.4
|
3 |
-
langchain==0.0.
|
4 |
-
langsmith==0.0.
|
5 |
numpy>=1.22.2 # not directly required, pinned by Snyk to avoid a vulnerability
|
6 |
-
openai==1.3.
|
7 |
pillow>=10.0.1 # not directly required, pinned by Snyk to avoid a vulnerability
|
8 |
pyarrow>=14.0.1 # not directly required, pinned by Snyk to avoid a vulnerability
|
9 |
pypdf==3.17.1
|
10 |
rank_bm25==0.2.2
|
11 |
-
streamlit==1.28.
|
12 |
-
streamlit-feedback==0.1.
|
13 |
tiktoken==0.5.1
|
14 |
tornado>=6.3.3 # not directly required, pinned by Snyk to avoid a vulnerability
|
15 |
validators>=0.21.0 # not directly required, pinned by Snyk to avoid a vulnerability
|
|
|
1 |
+
anthropic==0.7.7
|
2 |
faiss-cpu==1.7.4
|
3 |
+
langchain==0.0.343
|
4 |
+
langsmith==0.0.67
|
5 |
numpy>=1.22.2 # not directly required, pinned by Snyk to avoid a vulnerability
|
6 |
+
openai==1.3.6
|
7 |
pillow>=10.0.1 # not directly required, pinned by Snyk to avoid a vulnerability
|
8 |
pyarrow>=14.0.1 # not directly required, pinned by Snyk to avoid a vulnerability
|
9 |
pypdf==3.17.1
|
10 |
rank_bm25==0.2.2
|
11 |
+
streamlit==1.28.2
|
12 |
+
streamlit-feedback==0.1.3
|
13 |
tiktoken==0.5.1
|
14 |
tornado>=6.3.3 # not directly required, pinned by Snyk to avoid a vulnerability
|
15 |
validators>=0.21.0 # not directly required, pinned by Snyk to avoid a vulnerability
|