Clémentine commited on
Commit
8dfa543
1 Parent(s): 97b27da

moved the submit to a tab since the results are becoming very long

Browse files
Files changed (1) hide show
  1. app.py +43 -41
app.py CHANGED
@@ -335,51 +335,40 @@ with demo:
335
  with gr.TabItem("About", elem_id="llm-benchmark-tab-table", id=2):
336
  gr.Markdown(LLM_BENCHMARKS_TEXT, elem_classes="markdown-text")
337
 
338
- with gr.Column():
339
- with gr.Row():
340
- gr.Markdown(EVALUATION_QUEUE_TEXT, elem_classes="markdown-text")
341
-
342
- with gr.Column():
343
- with gr.Accordion("✅ Finished Evaluations", open=False):
344
  with gr.Row():
345
- finished_eval_table = gr.components.Dataframe(
346
- value=finished_eval_queue_df,
347
- headers=EVAL_COLS,
348
- datatype=EVAL_TYPES,
349
- max_rows=5,
350
- )
351
- with gr.Accordion("🔄 Running Evaluation Queue", open=False):
352
- with gr.Row():
353
- running_eval_table = gr.components.Dataframe(
354
- value=running_eval_queue_df,
355
- headers=EVAL_COLS,
356
- datatype=EVAL_TYPES,
357
- max_rows=5,
358
- )
359
 
360
- with gr.Accordion("⏳ Pending Evaluation Queue", open=False):
361
- with gr.Row():
362
- pending_eval_table = gr.components.Dataframe(
363
- value=pending_eval_queue_df,
364
- headers=EVAL_COLS,
365
- datatype=EVAL_TYPES,
366
- max_rows=5,
367
- )
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
368
 
369
- with gr.Row():
370
- refresh_button = gr.Button("Refresh")
371
- refresh_button.click(
372
- refresh,
373
- inputs=[],
374
- outputs=[
375
- leaderboard_table,
376
- finished_eval_table,
377
- running_eval_table,
378
- pending_eval_table,
379
- ],
380
- )
381
- with gr.Accordion("Submit a new model for evaluation"):
382
  with gr.Row():
 
383
  with gr.Column():
384
  model_name_textbox = gr.Textbox(label="Model name")
385
  revision_name_textbox = gr.Textbox(
@@ -413,6 +402,19 @@ with demo:
413
  submission_result,
414
  )
415
 
 
 
 
 
 
 
 
 
 
 
 
 
 
416
  with gr.Row():
417
  with gr.Accordion("📙 Citation", open=False):
418
  citation_button = gr.Textbox(
 
335
  with gr.TabItem("About", elem_id="llm-benchmark-tab-table", id=2):
336
  gr.Markdown(LLM_BENCHMARKS_TEXT, elem_classes="markdown-text")
337
 
338
+ with gr.TabItem("✉️✨ Submit here! ", elem_id="llm-benchmark-tab-table", id=3):
339
+ with gr.Column():
 
 
 
 
340
  with gr.Row():
341
+ gr.Markdown(EVALUATION_QUEUE_TEXT, elem_classes="markdown-text")
 
 
 
 
 
 
 
 
 
 
 
 
 
342
 
343
+ with gr.Column():
344
+ with gr.Accordion("✅ Finished Evaluations", open=False):
345
+ with gr.Row():
346
+ finished_eval_table = gr.components.Dataframe(
347
+ value=finished_eval_queue_df,
348
+ headers=EVAL_COLS,
349
+ datatype=EVAL_TYPES,
350
+ max_rows=5,
351
+ )
352
+ with gr.Accordion("🔄 Running Evaluation Queue", open=False):
353
+ with gr.Row():
354
+ running_eval_table = gr.components.Dataframe(
355
+ value=running_eval_queue_df,
356
+ headers=EVAL_COLS,
357
+ datatype=EVAL_TYPES,
358
+ max_rows=5,
359
+ )
360
+
361
+ with gr.Accordion("⏳ Pending Evaluation Queue", open=False):
362
+ with gr.Row():
363
+ pending_eval_table = gr.components.Dataframe(
364
+ value=pending_eval_queue_df,
365
+ headers=EVAL_COLS,
366
+ datatype=EVAL_TYPES,
367
+ max_rows=5,
368
+ )
369
 
 
 
 
 
 
 
 
 
 
 
 
 
 
370
  with gr.Row():
371
+ gr.Markdown("# SUBMIT HERE", elem_classes="markdown-text")
372
  with gr.Column():
373
  model_name_textbox = gr.Textbox(label="Model name")
374
  revision_name_textbox = gr.Textbox(
 
402
  submission_result,
403
  )
404
 
405
+ with gr.Row():
406
+ refresh_button = gr.Button("Refresh")
407
+ refresh_button.click(
408
+ refresh,
409
+ inputs=[],
410
+ outputs=[
411
+ leaderboard_table,
412
+ finished_eval_table,
413
+ running_eval_table,
414
+ pending_eval_table,
415
+ ],
416
+ )
417
+
418
  with gr.Row():
419
  with gr.Accordion("📙 Citation", open=False):
420
  citation_button = gr.Textbox(