Spaces:
Running
Running
Commit
·
a741634
1
Parent(s):
35cbe02
removed sliding window, as token size increased
Browse files
app.py
CHANGED
|
@@ -132,7 +132,7 @@ with st.sidebar:
|
|
| 132 |
label="Temperature", min_value=0.0, max_value=1.0, step=0.1, value=0.9)
|
| 133 |
|
| 134 |
st.session_state.max_tokens = st.slider(
|
| 135 |
-
label="New tokens to generate", min_value = 64, max_value=
|
| 136 |
)
|
| 137 |
|
| 138 |
st.session_state.repetion_penalty = st.slider(
|
|
@@ -203,8 +203,7 @@ if prompt := st.chat_input("Chat with Ikigai Docs..."):
|
|
| 203 |
len_response = (len(prompt.split()) + len(full_response.split())) * 1.25
|
| 204 |
st.session_state["tokens_used"] = len_response + st.session_state["tokens_used"]
|
| 205 |
|
| 206 |
-
|
| 207 |
-
st.session_state.history = st.session_state.history[-3:]
|
| 208 |
|
| 209 |
st.session_state.history.append([prompt, full_response])
|
| 210 |
st.session_state.history.append(identity_change)
|
|
|
|
| 132 |
label="Temperature", min_value=0.0, max_value=1.0, step=0.1, value=0.9)
|
| 133 |
|
| 134 |
st.session_state.max_tokens = st.slider(
|
| 135 |
+
label="New tokens to generate", min_value = 64, max_value=2048, step= 32, value=512
|
| 136 |
)
|
| 137 |
|
| 138 |
st.session_state.repetion_penalty = st.slider(
|
|
|
|
| 203 |
len_response = (len(prompt.split()) + len(full_response.split())) * 1.25
|
| 204 |
st.session_state["tokens_used"] = len_response + st.session_state["tokens_used"]
|
| 205 |
|
| 206 |
+
|
|
|
|
| 207 |
|
| 208 |
st.session_state.history.append([prompt, full_response])
|
| 209 |
st.session_state.history.append(identity_change)
|