yzabc007 commited on
Commit
8d7f3c1
·
1 Parent(s): 550d12e

Update space

Browse files
Files changed (1) hide show
  1. app.py +14 -1
app.py CHANGED
@@ -57,10 +57,11 @@ LEADERBOARD_DF = get_leaderboard_df(EVAL_RESULTS_PATH, EVAL_REQUESTS_PATH, COLS,
57
  pending_eval_queue_df,
58
  ) = get_evaluation_queue_df(EVAL_REQUESTS_PATH, EVAL_COLS)
59
 
 
60
  def init_leaderboard(dataframe):
61
  if dataframe is None or dataframe.empty:
62
  raise ValueError("Leaderboard DataFrame is empty or None.")
63
- # dataframe['None'] = []
64
  return Leaderboard(
65
  value=dataframe,
66
  datatype=[c.type for c in fields(AutoEvalColumn)],
@@ -105,6 +106,18 @@ with demo:
105
  with gr.TabItem("🏅 LLM Benchmark", elem_id="llm-benchmark-tab-table", id=0):
106
  leaderboard = init_leaderboard(LEADERBOARD_DF)
107
 
 
 
 
 
 
 
 
 
 
 
 
 
108
  with gr.TabItem("📝 About", elem_id="llm-benchmark-tab-table", id=2):
109
  gr.Markdown(LLM_BENCHMARKS_TEXT, elem_classes="markdown-text")
110
 
 
57
  pending_eval_queue_df,
58
  ) = get_evaluation_queue_df(EVAL_REQUESTS_PATH, EVAL_COLS)
59
 
60
+
61
  def init_leaderboard(dataframe):
62
  if dataframe is None or dataframe.empty:
63
  raise ValueError("Leaderboard DataFrame is empty or None.")
64
+
65
  return Leaderboard(
66
  value=dataframe,
67
  datatype=[c.type for c in fields(AutoEvalColumn)],
 
106
  with gr.TabItem("🏅 LLM Benchmark", elem_id="llm-benchmark-tab-table", id=0):
107
  leaderboard = init_leaderboard(LEADERBOARD_DF)
108
 
109
+ with gr.TabItem("Math", elem_id="llm-benchmark-tab-table", id=0):
110
+ leaderboard = init_leaderboard(LEADERBOARD_DF)
111
+
112
+ with gr.TabItem("Reasoning", elem_id="llm-benchmark-tab-table", id=0):
113
+ leaderboard = init_leaderboard(LEADERBOARD_DF)
114
+
115
+ with gr.TabItem("Coding", elem_id="llm-benchmark-tab-table", id=0):
116
+ leaderboard = init_leaderboard(LEADERBOARD_DF)
117
+
118
+ with gr.TabItem("Science", elem_id="llm-benchmark-tab-table", id=0):
119
+ leaderboard = init_leaderboard(LEADERBOARD_DF)
120
+
121
  with gr.TabItem("📝 About", elem_id="llm-benchmark-tab-table", id=2):
122
  gr.Markdown(LLM_BENCHMARKS_TEXT, elem_classes="markdown-text")
123