Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -21,11 +21,14 @@ import gradio as gr
|
|
21 |
from openai import OpenAI
|
22 |
|
23 |
# rebuild storage context and load knowledge index
|
24 |
-
|
25 |
-
from llama_index.
|
26 |
-
|
27 |
-
|
28 |
-
from llama_index.core import
|
|
|
|
|
|
|
29 |
|
30 |
# add datetime and ip to the log file
|
31 |
from datetime import datetime;
|
@@ -39,13 +42,15 @@ from uuid import uuid4
|
|
39 |
# generate an unique identifier for the session
|
40 |
session_id = uuid4()
|
41 |
|
42 |
-
# deprecated
|
43 |
storage_context = StorageContext.from_defaults(persist_dir='./')
|
44 |
# gpt-3.5_turbo is the current default model
|
45 |
-
|
46 |
-
|
47 |
-
|
48 |
-
index
|
|
|
|
|
49 |
|
50 |
class Chatbot:
|
51 |
def __init__(self, api_key, index):
|
@@ -68,7 +73,7 @@ class Chatbot:
|
|
68 |
)
|
69 |
|
70 |
def generate_response(self, user_input):
|
71 |
-
query_engine = index.as_query_engine(
|
72 |
response = query_engine.query(user_input)
|
73 |
|
74 |
# generate response
|
|
|
21 |
from openai import OpenAI
|
22 |
|
23 |
# rebuild storage context and load knowledge index
|
24 |
+
from llama_index import StorageContext, load_index_from_storage, LLMPredictor, ServiceContext
|
25 |
+
from llama_index.llms import OpenAI
|
26 |
+
|
27 |
+
# for llama-index 0.10
|
28 |
+
#from llama_index.core import StorageContext
|
29 |
+
#from llama_index.core import load_index_from_storage
|
30 |
+
#from llama_index.llms.openai import OpenAI
|
31 |
+
#from llama_index.core import Settings
|
32 |
|
33 |
# add datetime and ip to the log file
|
34 |
from datetime import datetime;
|
|
|
42 |
# generate an unique identifier for the session
|
43 |
session_id = uuid4()
|
44 |
|
45 |
+
# deprecated (llama-index 0.9)
|
46 |
storage_context = StorageContext.from_defaults(persist_dir='./')
|
47 |
# gpt-3.5_turbo is the current default model
|
48 |
+
llm = OpenAI(temperature=0.5, model="gpt-3.5_turbo")
|
49 |
+
service_context = ServiceContext.from_defaults(llm=llm)
|
50 |
+
index = load_index_from_storage(storage_context, service_context=service_context)
|
51 |
+
# for llama-index 0.10
|
52 |
+
#Settings.llm = OpenAI(temperature=0.5, model="gpt-3.5_turbo")
|
53 |
+
#index = load_index_from_storage(storage_context)
|
54 |
|
55 |
class Chatbot:
|
56 |
def __init__(self, api_key, index):
|
|
|
73 |
)
|
74 |
|
75 |
def generate_response(self, user_input):
|
76 |
+
query_engine = index.as_query_engine()
|
77 |
response = query_engine.query(user_input)
|
78 |
|
79 |
# generate response
|