danidanidani commited on
Commit
8d8b9ba
1 Parent(s): b80eb73

Update src/backend/chatbot.py

Browse files
Files changed (1) hide show
  1. src/backend/chatbot.py +7 -6
src/backend/chatbot.py CHANGED
@@ -18,7 +18,7 @@ from llama_index.llms.llama_utils import (
18
  messages_to_prompt,
19
  completion_to_prompt,
20
  )
21
- import subprocess
22
 
23
  # set version
24
  # st.session_state.demo_lite = False
@@ -33,11 +33,12 @@ print("BP 4 ")
33
  def init_llm(model, demo_lite):
34
  # st.write("BP 4.1: model: ", model)
35
  if demo_lite == False:
36
-
37
- # Set environment variables
38
- subprocess.run(['powershell', '$env:FORCE_CMAKE=1'], shell=True)
39
- subprocess.run(['powershell', '$env:CMAKE_ARGS="-DLLAMA_CUBLAS=on"'], shell=True)
40
-
 
41
  print("BP 5 : running full demo")
42
  if model == "Llama2-7b_CPP":
43
  model_path = "src/models/llama-2-7b-chat.Q4_K_M.gguf"
 
18
  messages_to_prompt,
19
  completion_to_prompt,
20
  )
21
+ from huggingface_hub import HfApi, SpaceHardware
22
 
23
  # set version
24
  # st.session_state.demo_lite = False
 
33
  def init_llm(model, demo_lite):
34
  # st.write("BP 4.1: model: ", model)
35
  if demo_lite == False:
36
+ api = HfApi(token=HF_TOKEN)
37
+ runtime = api.get_space_runtime(repo_id=TRAINING_SPACE_ID)
38
+ # Check if Space is loaded with a GPU.
39
+ if runtime.hardware == SpaceHardware.T4_SMALL:
40
+ print("using nvidia T4 small")
41
+
42
  print("BP 5 : running full demo")
43
  if model == "Llama2-7b_CPP":
44
  model_path = "src/models/llama-2-7b-chat.Q4_K_M.gguf"