Spaces:
Paused
Paused
Update app.py
Browse files
app.py
CHANGED
|
@@ -32,6 +32,12 @@ if 'hf_token' not in st.session_state:
|
|
| 32 |
st.session_state.hf_token = None
|
| 33 |
if 'repo_name' not in st.session_state:
|
| 34 |
st.session_state.repo_name = None
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 35 |
|
| 36 |
# --- Agent Class ---
|
| 37 |
class AIAgent:
|
|
@@ -123,7 +129,7 @@ def chat_interface_with_agent(input_text, agent_name):
|
|
| 123 |
return f"Agent {agent_name} not found."
|
| 124 |
|
| 125 |
# Use a more powerful language model (GPT-3 or similar) for better chat experience
|
| 126 |
-
model_name = "text-davinci-003" #
|
| 127 |
try:
|
| 128 |
model = transformers_pipeline("text-generation", model=model_name)
|
| 129 |
except EnvironmentError as e:
|
|
@@ -143,7 +149,7 @@ def chat_interface_with_cluster(input_text, cluster_name):
|
|
| 143 |
return f"Cluster {cluster_name} not found."
|
| 144 |
|
| 145 |
# Use a more powerful language model (GPT-3 or similar) for better chat experience
|
| 146 |
-
model_name = "text-davinci-003" #
|
| 147 |
try:
|
| 148 |
model = transformers_pipeline("text-generation", model=model_name)
|
| 149 |
except EnvironmentError as e:
|
|
@@ -260,7 +266,7 @@ def translate_code(code, source_language, target_language):
|
|
| 260 |
|
| 261 |
def generate_code(idea):
|
| 262 |
"""Generates code based on a given idea using a Hugging Face model."""
|
| 263 |
-
model_name = "bigcode/starcoder" #
|
| 264 |
try:
|
| 265 |
model = AutoModelForCausalLM.from_pretrained(model_name)
|
| 266 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
|
@@ -488,7 +494,10 @@ with tabs[1]:
|
|
| 488 |
st.subheader("Chat with AI Agents")
|
| 489 |
selected_agent_or_cluster = st.selectbox("Select an AI agent or cluster", st.session_state.available_agents + st.session_state.available_clusters)
|
| 490 |
agent_chat_input = st.text_area("Enter your message:")
|
|
|
|
|
|
|
| 491 |
if st.button("Send"):
|
|
|
|
| 492 |
if selected_agent_or_cluster in st.session_state.available_agents:
|
| 493 |
st.session_state.current_agent = selected_agent_or_cluster
|
| 494 |
st.session_state.current_cluster = None
|
|
@@ -532,7 +541,10 @@ with tabs[2]:
|
|
| 532 |
st.subheader("Chat with AI Agents")
|
| 533 |
selected_agent_or_cluster = st.selectbox("Select an AI agent or cluster", st.session_state.available_agents + st.session_state.available_clusters)
|
| 534 |
agent_chat_input = st.text_area("Enter your message:")
|
|
|
|
|
|
|
| 535 |
if st.button("Send"):
|
|
|
|
| 536 |
if selected_agent_or_cluster in st.session_state.available_agents:
|
| 537 |
st.session_state.current_agent = selected_agent_or_cluster
|
| 538 |
st.session_state.current_cluster = None
|
|
@@ -599,7 +611,10 @@ with tabs[2]:
|
|
| 599 |
|
| 600 |
st.write("Code Generation:")
|
| 601 |
code_idea = st.text_input("Enter your code idea:")
|
|
|
|
|
|
|
| 602 |
if st.button("Generate"):
|
|
|
|
| 603 |
generated_code = generate_code(code_idea)
|
| 604 |
st.code(generated_code, language="python")
|
| 605 |
|
|
@@ -608,7 +623,6 @@ with tabs[2]:
|
|
| 608 |
if st.session_state.current_project:
|
| 609 |
st.write(f"Current project: {st.session_state.current_project}")
|
| 610 |
if st.button("Build"):
|
| 611 |
-
# Implement build logic here
|
| 612 |
build_dir = build_project(st.session_state.current_project)
|
| 613 |
st.write(f"Project built successfully! Build directory: {build_dir}")
|
| 614 |
|
|
@@ -620,11 +634,9 @@ with tabs[2]:
|
|
| 620 |
if st.button("Deploy to Hugging Face Spaces"):
|
| 621 |
st.session_state.hf_token = hf_token
|
| 622 |
st.session_state.repo_name = repo_name
|
| 623 |
-
# Implement Hugging Face Spaces deployment logic here
|
| 624 |
deploy_to_huggingface(build_dir, hf_token, repo_name)
|
| 625 |
elif deployment_target == "Local":
|
| 626 |
if st.button("Deploy Locally"):
|
| 627 |
-
# Implement local deployment logic here
|
| 628 |
deploy_locally(build_dir)
|
| 629 |
else:
|
| 630 |
st.warning("Please select a project first.")
|
|
|
|
| 32 |
st.session_state.hf_token = None
|
| 33 |
if 'repo_name' not in st.session_state:
|
| 34 |
st.session_state.repo_name = None
|
| 35 |
+
if 'selected_model' not in st.session_state:
|
| 36 |
+
st.session_state.selected_model = None
|
| 37 |
+
if 'selected_code_model' not in st.session_state:
|
| 38 |
+
st.session_state.selected_code_model = None
|
| 39 |
+
if 'selected_chat_model' not in st.session_state:
|
| 40 |
+
st.session_state.selected_chat_model = None
|
| 41 |
|
| 42 |
# --- Agent Class ---
|
| 43 |
class AIAgent:
|
|
|
|
| 129 |
return f"Agent {agent_name} not found."
|
| 130 |
|
| 131 |
# Use a more powerful language model (GPT-3 or similar) for better chat experience
|
| 132 |
+
model_name = st.session_state.selected_chat_model or "text-davinci-003" # Default to GPT-3 if not selected
|
| 133 |
try:
|
| 134 |
model = transformers_pipeline("text-generation", model=model_name)
|
| 135 |
except EnvironmentError as e:
|
|
|
|
| 149 |
return f"Cluster {cluster_name} not found."
|
| 150 |
|
| 151 |
# Use a more powerful language model (GPT-3 or similar) for better chat experience
|
| 152 |
+
model_name = st.session_state.selected_chat_model or "text-davinci-003" # Default to GPT-3 if not selected
|
| 153 |
try:
|
| 154 |
model = transformers_pipeline("text-generation", model=model_name)
|
| 155 |
except EnvironmentError as e:
|
|
|
|
| 266 |
|
| 267 |
def generate_code(idea):
|
| 268 |
"""Generates code based on a given idea using a Hugging Face model."""
|
| 269 |
+
model_name = st.session_state.selected_code_model or "bigcode/starcoder" # Default to Starcoder if not selected
|
| 270 |
try:
|
| 271 |
model = AutoModelForCausalLM.from_pretrained(model_name)
|
| 272 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
|
|
|
| 494 |
st.subheader("Chat with AI Agents")
|
| 495 |
selected_agent_or_cluster = st.selectbox("Select an AI agent or cluster", st.session_state.available_agents + st.session_state.available_clusters)
|
| 496 |
agent_chat_input = st.text_area("Enter your message:")
|
| 497 |
+
chat_model_options = ["text-davinci-003", "gpt-3.5-turbo"] # Add more chat models as needed
|
| 498 |
+
selected_chat_model = st.selectbox("Select a chat model", chat_model_options)
|
| 499 |
if st.button("Send"):
|
| 500 |
+
st.session_state.selected_chat_model = selected_chat_model
|
| 501 |
if selected_agent_or_cluster in st.session_state.available_agents:
|
| 502 |
st.session_state.current_agent = selected_agent_or_cluster
|
| 503 |
st.session_state.current_cluster = None
|
|
|
|
| 541 |
st.subheader("Chat with AI Agents")
|
| 542 |
selected_agent_or_cluster = st.selectbox("Select an AI agent or cluster", st.session_state.available_agents + st.session_state.available_clusters)
|
| 543 |
agent_chat_input = st.text_area("Enter your message:")
|
| 544 |
+
chat_model_options = ["text-davinci-003", "gpt-3.5-turbo"] # Add more chat models as needed
|
| 545 |
+
selected_chat_model = st.selectbox("Select a chat model", chat_model_options)
|
| 546 |
if st.button("Send"):
|
| 547 |
+
st.session_state.selected_chat_model = selected_chat_model
|
| 548 |
if selected_agent_or_cluster in st.session_state.available_agents:
|
| 549 |
st.session_state.current_agent = selected_agent_or_cluster
|
| 550 |
st.session_state.current_cluster = None
|
|
|
|
| 611 |
|
| 612 |
st.write("Code Generation:")
|
| 613 |
code_idea = st.text_input("Enter your code idea:")
|
| 614 |
+
code_model_options = ["bigcode/starcoder", "google/flan-t5-xl"] # Add more code models as needed
|
| 615 |
+
selected_code_model = st.selectbox("Select a code generation model", code_model_options)
|
| 616 |
if st.button("Generate"):
|
| 617 |
+
st.session_state.selected_code_model = selected_code_model
|
| 618 |
generated_code = generate_code(code_idea)
|
| 619 |
st.code(generated_code, language="python")
|
| 620 |
|
|
|
|
| 623 |
if st.session_state.current_project:
|
| 624 |
st.write(f"Current project: {st.session_state.current_project}")
|
| 625 |
if st.button("Build"):
|
|
|
|
| 626 |
build_dir = build_project(st.session_state.current_project)
|
| 627 |
st.write(f"Project built successfully! Build directory: {build_dir}")
|
| 628 |
|
|
|
|
| 634 |
if st.button("Deploy to Hugging Face Spaces"):
|
| 635 |
st.session_state.hf_token = hf_token
|
| 636 |
st.session_state.repo_name = repo_name
|
|
|
|
| 637 |
deploy_to_huggingface(build_dir, hf_token, repo_name)
|
| 638 |
elif deployment_target == "Local":
|
| 639 |
if st.button("Deploy Locally"):
|
|
|
|
| 640 |
deploy_locally(build_dir)
|
| 641 |
else:
|
| 642 |
st.warning("Please select a project first.")
|