Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -26,6 +26,9 @@ from llama_index.llms.openai import OpenAI
|
|
26 |
from datetime import datetime;
|
27 |
import socket;
|
28 |
|
|
|
|
|
|
|
29 |
# deprecated
|
30 |
storage_context = StorageContext.from_defaults(persist_dir='./')
|
31 |
# gpt-3.5_turbo is the current default model
|
@@ -39,7 +42,8 @@ class Chatbot:
|
|
39 |
self.index = index
|
40 |
openai.api_key = api_key
|
41 |
self.chat_history = []
|
42 |
-
self.history_file = f"zlmqi/index/chat_log.json"
|
|
|
43 |
|
44 |
def generate_response(self, user_input):
|
45 |
query_engine = index.as_query_engine(llm=llm_predictor)
|
|
|
26 |
from datetime import datetime;
|
27 |
import socket;
|
28 |
|
29 |
+
# access data stored in datasets
|
30 |
+
from datasets import load_dataset
|
31 |
+
|
32 |
# deprecated
|
33 |
storage_context = StorageContext.from_defaults(persist_dir='./')
|
34 |
# gpt-3.5_turbo is the current default model
|
|
|
42 |
self.index = index
|
43 |
openai.api_key = api_key
|
44 |
self.chat_history = []
|
45 |
+
#self.history_file = f"zlmqi/index/chat_log.json"
|
46 |
+
self.history_file = load_dataset("json", data_files="chat_log.json")
|
47 |
|
48 |
def generate_response(self, user_input):
|
49 |
query_engine = index.as_query_engine(llm=llm_predictor)
|