Benjamin Consolvo
commited on
Commit
·
97ccf9a
1
Parent(s):
0577ef7
markdown links
Browse files- app.py +3 -15
- model_info.json +0 -18
app.py
CHANGED
@@ -5,38 +5,26 @@ import os
|
|
5 |
from openai import OpenAI
|
6 |
import json
|
7 |
|
8 |
-
working_dir = os.path.dirname(os.path.abspath(__file__))
|
9 |
-
endpoint_data = json.load(open(f"{working_dir}/model_info.json"))
|
10 |
-
|
11 |
def clear_chat():
|
12 |
st.session_state.messages = []
|
13 |
|
14 |
st.title("Intel® AI for Enterprise Inference")
|
15 |
st.header("LLM chatbot")
|
16 |
|
17 |
-
# Extract the keys (model names) from the JSON data
|
18 |
-
# model_names = list(endpoint_data.keys())
|
19 |
-
|
20 |
-
|
21 |
with st.sidebar:
|
22 |
-
#Enter openai_api key under "Secrets " in HF settings
|
23 |
-
#Enter base_url under "Variables" in HF settings
|
24 |
-
api_key = st.session_state.api_key = st.secrets["openai_apikey"]
|
25 |
-
base_url = st.session_state.base_url = os.environ.get("base_url")
|
26 |
client = OpenAI(api_key=api_key, base_url=base_url)
|
27 |
models = client.models.list()
|
28 |
model_names = sorted([model.id for model in models]) # Extract 'id' from each model object
|
29 |
-
# Specify the default model name
|
30 |
default_model_name = "meta-llama/Llama-3.3-70B-Instruct" # Replace with your desired default model name
|
31 |
|
32 |
-
|
33 |
# Use st.session_state to persist the selected model
|
34 |
if "selected_model" not in st.session_state:
|
35 |
-
# st.session_state.selected_model = model_names[0] # Default to the first model
|
36 |
st.session_state.selected_model = default_model_name if default_model_name in model_names else model_names[0]
|
37 |
|
38 |
modelname = st.selectbox(
|
39 |
-
"Select LLM model (Running on Intel® Gaudi®) on Denvr Dataworks",
|
40 |
model_names,
|
41 |
index=model_names.index(st.session_state.selected_model) if st.session_state.selected_model in model_names else 0,
|
42 |
key="selected_model",
|
|
|
5 |
from openai import OpenAI
|
6 |
import json
|
7 |
|
|
|
|
|
|
|
8 |
def clear_chat():
|
9 |
st.session_state.messages = []
|
10 |
|
11 |
st.title("Intel® AI for Enterprise Inference")
|
12 |
st.header("LLM chatbot")
|
13 |
|
|
|
|
|
|
|
|
|
14 |
with st.sidebar:
|
15 |
+
api_key = st.session_state.api_key = st.secrets["openai_apikey"] #Enter openai_api key under "Secrets " in HF settings
|
16 |
+
base_url = st.session_state.base_url = os.environ.get("base_url") #Enter base_url under "Variables" in HF settings
|
|
|
|
|
17 |
client = OpenAI(api_key=api_key, base_url=base_url)
|
18 |
models = client.models.list()
|
19 |
model_names = sorted([model.id for model in models]) # Extract 'id' from each model object
|
|
|
20 |
default_model_name = "meta-llama/Llama-3.3-70B-Instruct" # Replace with your desired default model name
|
21 |
|
|
|
22 |
# Use st.session_state to persist the selected model
|
23 |
if "selected_model" not in st.session_state:
|
|
|
24 |
st.session_state.selected_model = default_model_name if default_model_name in model_names else model_names[0]
|
25 |
|
26 |
modelname = st.selectbox(
|
27 |
+
"Select an LLM model (Running on Intel® Gaudi®). Hosted on Denvr Dataworks",
|
28 |
model_names,
|
29 |
index=model_names.index(st.session_state.selected_model) if st.session_state.selected_model in model_names else 0,
|
30 |
key="selected_model",
|
model_info.json
DELETED
@@ -1,18 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"meta-llama/Llama-3.3-70B-Instruct": "https://api.inference.denvrdata.com/v1/",
|
3 |
-
"meta-llama/Meta-Llama-3.2-1B-Instruct": "https://api.inference.denvrdata.com/v1/",
|
4 |
-
"meta-llama/Meta-Llama-3.2-3B-Instruct": "https://api.inference.denvrdata.com/v1/",
|
5 |
-
"meta-llama/Meta-Llama-3.1-8B-Instruct": "https://api.inference.denvrdata.com/v1/",
|
6 |
-
"meta-llama/Meta-Llama-3.1-70B-Instruct": "https://api.inference.denvrdata.com/v1/",
|
7 |
-
"deepseek-ai/DeepSeek-R1-Distill-Llama-8B": "https://api.inference.denvrdata.com/v1/",
|
8 |
-
"deepseek-ai/DeepSeek-R1-Distill-Llama-70B": "https://api.inference.denvrdata.com/v1/",
|
9 |
-
"mistralai/Mistral-7B-Instruct-v0.2": "https://api.inference.denvrdata.com/v1/",
|
10 |
-
"mistralai/Mistral-7B-Instruct-v0.1": "https://api.inference.denvrdata.com/v1/",
|
11 |
-
"mistralai/Mixtral-8x7B-Instruct-v0.1": "https://api.inference.denvrdata.com/v1/",
|
12 |
-
"tiiuae/Falcon3-7B-Instruct": "https://api.inference.denvrdata.com/v1/",
|
13 |
-
"tiiuae/Falcon3-10B-Instruct": "https://api.inference.denvrdata.com/v1/",
|
14 |
-
"tiiuae/Falcon3.1-74B-Instruct": "https://api.inference.denvrdata.com/v1/",
|
15 |
-
"ALLaM/ALLaM-7B-Instruct-preview": "https://api.inference.denvrdata.com/v1/",
|
16 |
-
"inceptionai/jais-adapted-70b-chat": "https://api.inference.denvrdata.com/v1/",
|
17 |
-
"codellama/CodeLlama-34b-Instruct-hf": "https://api.inference.denvrdata.com/v1/"
|
18 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|