Spaces:
Sleeping
Sleeping
Commit
·
8d8b9ba
1
Parent(s):
b80eb73
Update src/backend/chatbot.py
Browse files- src/backend/chatbot.py +7 -6
src/backend/chatbot.py
CHANGED
@@ -18,7 +18,7 @@ from llama_index.llms.llama_utils import (
|
|
18 |
messages_to_prompt,
|
19 |
completion_to_prompt,
|
20 |
)
|
21 |
-
import
|
22 |
|
23 |
# set version
|
24 |
# st.session_state.demo_lite = False
|
@@ -33,11 +33,12 @@ print("BP 4 ")
|
|
33 |
def init_llm(model, demo_lite):
|
34 |
# st.write("BP 4.1: model: ", model)
|
35 |
if demo_lite == False:
|
36 |
-
|
37 |
-
|
38 |
-
|
39 |
-
|
40 |
-
|
|
|
41 |
print("BP 5 : running full demo")
|
42 |
if model == "Llama2-7b_CPP":
|
43 |
model_path = "src/models/llama-2-7b-chat.Q4_K_M.gguf"
|
|
|
18 |
messages_to_prompt,
|
19 |
completion_to_prompt,
|
20 |
)
|
21 |
+
from huggingface_hub import HfApi, SpaceHardware
|
22 |
|
23 |
# set version
|
24 |
# st.session_state.demo_lite = False
|
|
|
33 |
def init_llm(model, demo_lite):
|
34 |
# st.write("BP 4.1: model: ", model)
|
35 |
if demo_lite == False:
|
36 |
+
api = HfApi(token=HF_TOKEN)
|
37 |
+
runtime = api.get_space_runtime(repo_id=TRAINING_SPACE_ID)
|
38 |
+
# Check if Space is loaded with a GPU.
|
39 |
+
if runtime.hardware == SpaceHardware.T4_SMALL:
|
40 |
+
print("using nvidia T4 small")
|
41 |
+
|
42 |
print("BP 5 : running full demo")
|
43 |
if model == "Llama2-7b_CPP":
|
44 |
model_path = "src/models/llama-2-7b-chat.Q4_K_M.gguf"
|