wadood commited on
Commit
77c75da
1 Parent(s): 5e1b642

added metric selection functionality in comparision tab

Browse files
Files changed (1) hide show
  1. app.py +17 -6
app.py CHANGED
@@ -150,12 +150,20 @@ if __name__ == "__main__":
150
  ### EVALUATION METRICS COMPARISION ###
151
 
152
  st.subheader("Evaluation Metrics Comparision") # , divider='rainbow')
153
- metric_names = "\n".join(
154
- ["- " + evaluation_metric.name for evaluation_metric in EVALUATION_METRICS]
155
- )
156
  st.markdown(
157
- "The different evaluation metrics we have for the NER task are\n"
158
- f"{metric_names}"
 
 
 
 
 
 
 
 
159
  )
160
 
161
  with st.expander("View Predictions Details"):
@@ -175,4 +183,7 @@ if __name__ == "__main__":
175
 
176
  metrics_df = predictions_df.drop(["ner_spans"], axis=1)
177
 
178
- st.write(metrics_df.to_html(escape=False), unsafe_allow_html=True)
 
 
 
 
150
  ### EVALUATION METRICS COMPARISION ###
151
 
152
  st.subheader("Evaluation Metrics Comparision") # , divider='rainbow')
153
+ # metric_names = "\n".join(
154
+ # ["- " + evaluation_metric.name for evaluation_metric in EVALUATION_METRICS]
155
+ # )
156
  st.markdown(
157
+ "The different evaluation metrics we have for the NER task are shown below, select the metrics to compare.\n"
158
+ # f"{metric_names}"
159
+ )
160
+
161
+ metrics_selection = [
162
+ (st.checkbox(evaluation_metric.name, value=True), evaluation_metric.name)
163
+ for evaluation_metric in EVALUATION_METRICS
164
+ ]
165
+ metrics_to_show = list(
166
+ map(lambda x: x[1], filter(lambda x: x[0], metrics_selection))
167
  )
168
 
169
  with st.expander("View Predictions Details"):
 
183
 
184
  metrics_df = predictions_df.drop(["ner_spans"], axis=1)
185
 
186
+ st.write(
187
+ metrics_df[["Prediction"] + metrics_to_show].to_html(escape=False),
188
+ unsafe_allow_html=True,
189
+ )