Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -5,7 +5,7 @@ from typing import Generator
|
|
5 |
from groq import Groq
|
6 |
|
7 |
_ = load_dotenv(find_dotenv())
|
8 |
-
st.set_page_config(page_icon="
|
9 |
|
10 |
|
11 |
def icon(emoji: str):
|
@@ -16,7 +16,7 @@ def icon(emoji: str):
|
|
16 |
)
|
17 |
|
18 |
|
19 |
-
icon("⚡️")
|
20 |
|
21 |
st.subheader("Groq Chat with LLaMA3 App", divider="rainbow", anchor=False)
|
22 |
|
@@ -111,7 +111,7 @@ if prompt := st.chat_input("Enter your prompt here..."):
|
|
111 |
chat_responses_generator = generate_chat_responses(chat_completion)
|
112 |
full_response = st.write_stream(chat_responses_generator)
|
113 |
except Exception as e:
|
114 |
-
st.error(e, icon="
|
115 |
|
116 |
# Append the full response to session_state.messages
|
117 |
if isinstance(full_response, str):
|
|
|
5 |
from groq import Groq
|
6 |
|
7 |
_ = load_dotenv(find_dotenv())
|
8 |
+
st.set_page_config(page_icon="📃", layout="wide", page_title="Groq & LLaMA3 Chat Bot...")
|
9 |
|
10 |
|
11 |
def icon(emoji: str):
|
|
|
16 |
)
|
17 |
|
18 |
|
19 |
+
# icon("⚡️")
|
20 |
|
21 |
st.subheader("Groq Chat with LLaMA3 App", divider="rainbow", anchor=False)
|
22 |
|
|
|
111 |
chat_responses_generator = generate_chat_responses(chat_completion)
|
112 |
full_response = st.write_stream(chat_responses_generator)
|
113 |
except Exception as e:
|
114 |
+
st.error(e, icon="❌")
|
115 |
|
116 |
# Append the full response to session_state.messages
|
117 |
if isinstance(full_response, str):
|