Update app.py
Browse files
app.py
CHANGED
|
@@ -81,8 +81,8 @@ all_splits = text_splitter.split_documents(data)
|
|
| 81 |
vectorstore = Chroma.from_documents(documents=all_splits, embedding=OpenAIEmbeddings())
|
| 82 |
|
| 83 |
# Chat Model
|
| 84 |
-
|
| 85 |
-
model = 'gpt-4'
|
| 86 |
|
| 87 |
llm = ChatOpenAI(model=model, temperature=0)
|
| 88 |
|
|
@@ -154,7 +154,7 @@ with gr.Blocks() as demo:
|
|
| 154 |
fn= predict, inputs=[msg, chatbot], outputs=[msg, chatbot],
|
| 155 |
concurrency_limit = 100).then(text_fn)
|
| 156 |
|
| 157 |
-
gr.
|
| 158 |
["Suggest a topic"],["Tell me more about that"],
|
| 159 |
["Can I have a reference to read more?"],["Please provide citations."]],
|
| 160 |
msg)
|
|
|
|
| 81 |
vectorstore = Chroma.from_documents(documents=all_splits, embedding=OpenAIEmbeddings())
|
| 82 |
|
| 83 |
# Chat Model
|
| 84 |
+
model = 'gpt-3.5-turbo'
|
| 85 |
+
# model = 'gpt-4'
|
| 86 |
|
| 87 |
llm = ChatOpenAI(model=model, temperature=0)
|
| 88 |
|
|
|
|
| 154 |
fn= predict, inputs=[msg, chatbot], outputs=[msg, chatbot],
|
| 155 |
concurrency_limit = 100).then(text_fn)
|
| 156 |
|
| 157 |
+
gr.Examples=([["What are the main topics?"],["Summarise the lesson"],["Explain this simply"],
|
| 158 |
["Suggest a topic"],["Tell me more about that"],
|
| 159 |
["Can I have a reference to read more?"],["Please provide citations."]],
|
| 160 |
msg)
|