Spaces:
Sleeping
Sleeping
Change input placement
Browse files
app.py
CHANGED
@@ -33,28 +33,22 @@ with st.expander("Last price update ($ currency): {}.".format(os.getenv("PRICE_U
|
|
33 |
|
34 |
st.subheader("Your project information:")
|
35 |
with st.expander("#### Average use of your LLM:", expanded=True):
|
36 |
-
col1, col2, col3 = st.columns(
|
37 |
|
38 |
with col1:
|
39 |
request_number = st.number_input("Average request number per hour", min_value=0, value=0, step=100)
|
40 |
|
41 |
with col2:
|
42 |
-
|
43 |
-
|
44 |
-
with col2_1:
|
45 |
-
request_size = st.number_input("Average input token size", min_value=0, value=0, step=1000)
|
46 |
-
|
47 |
-
with col2_2:
|
48 |
-
response_size = st.number_input("Average output token size", min_value=0, value=0, step=1000)
|
49 |
|
50 |
with col3:
|
51 |
-
|
52 |
|
53 |
-
|
54 |
-
|
55 |
|
56 |
-
|
57 |
-
|
58 |
|
59 |
with st.expander("#### LLM models to compare:", expanded=True):
|
60 |
llm_chat_list = st.multiselect("Select Chat LLM models", options=[llm.name for llm in llm_service.get_list_by_category("chat")])
|
|
|
33 |
|
34 |
st.subheader("Your project information:")
|
35 |
with st.expander("#### Average use of your LLM:", expanded=True):
|
36 |
+
col1, col2, col3, col4, col5 = st.columns(5)
|
37 |
|
38 |
with col1:
|
39 |
request_number = st.number_input("Average request number per hour", min_value=0, value=0, step=100)
|
40 |
|
41 |
with col2:
|
42 |
+
request_size = st.number_input("Average input token size", min_value=0, value=0, step=1000)
|
|
|
|
|
|
|
|
|
|
|
|
|
43 |
|
44 |
with col3:
|
45 |
+
response_size = st.number_input("Average output token size", min_value=0, value=0, step=1000)
|
46 |
|
47 |
+
with col4:
|
48 |
+
run_hours = st.number_input("Running hours per day", min_value=0, max_value=24, value=0, step=1)
|
49 |
|
50 |
+
with col5:
|
51 |
+
run_day = st.number_input("Running days per week", min_value=0, max_value=7, value=0, step=1)
|
52 |
|
53 |
with st.expander("#### LLM models to compare:", expanded=True):
|
54 |
llm_chat_list = st.multiselect("Select Chat LLM models", options=[llm.name for llm in llm_service.get_list_by_category("chat")])
|