fixed minor typos
Browse files- .gitignore +1 -1
- app_modules/llm_chat_chain.py +1 -1
.gitignore
CHANGED
@@ -1,4 +1,4 @@
|
|
1 |
-
pdfs
|
2 |
.vscode/
|
3 |
|
4 |
# Byte-compiled / optimized / DLL files
|
|
|
1 |
+
pdfs
|
2 |
.vscode/
|
3 |
|
4 |
# Byte-compiled / optimized / DLL files
|
app_modules/llm_chat_chain.py
CHANGED
@@ -13,7 +13,7 @@ def get_llama_2_prompt_template():
|
|
13 |
B_SYS, E_SYS = "<<SYS>>\n", "\n<</SYS>>\n\n"
|
14 |
|
15 |
instruction = "Chat History:\n\n{chat_history} \n\nUser: {question}"
|
16 |
-
system_prompt = "You are a helpful assistant, you always only answer for the assistant then you stop.
|
17 |
|
18 |
SYSTEM_PROMPT = B_SYS + system_prompt + E_SYS
|
19 |
prompt_template = B_INST + SYSTEM_PROMPT + instruction + E_INST
|
|
|
13 |
B_SYS, E_SYS = "<<SYS>>\n", "\n<</SYS>>\n\n"
|
14 |
|
15 |
instruction = "Chat History:\n\n{chat_history} \n\nUser: {question}"
|
16 |
+
system_prompt = "You are a helpful assistant, you always only answer for the assistant then you stop. Read the chat history to get context"
|
17 |
|
18 |
SYSTEM_PROMPT = B_SYS + system_prompt + E_SYS
|
19 |
prompt_template = B_INST + SYSTEM_PROMPT + instruction + E_INST
|