Files changed (3) hide show
  1. app.py +5 -2
  2. requirements.txt +1 -1
  3. utilities.py +1 -0
app.py CHANGED
@@ -734,6 +734,8 @@ with gr.Blocks(theme='shivi/calm_seafoam', css_paths='style.css', js=js_func) as
734
  metrics_conc = target_df
735
  if 'valid_efficency_score' not in metrics_conc.columns:
736
  metrics_conc['valid_efficency_score'] = metrics_conc['VES']
 
 
737
  eval_text = generate_eval_text("End evaluation")
738
  yield gr.Markdown(eval_text, visible=True), gr.Image(), gr.Markdown(), gr.Markdown(), gr.Markdown(), metrics_conc, *[predictions_dict[model][columns_to_visulize] for model in model_list]
739
 
@@ -868,14 +870,15 @@ with gr.Blocks(theme='shivi/calm_seafoam', css_paths='style.css', js=js_func) as
868
  db_path_name="db_path"
869
  )
870
  else:
 
871
  metrics_df_model = us.evaluate_answer(predictions_dict[model])
872
  metrics_df_model['model'] = model
873
  metrics_conc = pd.concat([metrics_conc, metrics_df_model], ignore_index=True)
874
 
875
- if 'valid_efficency_score' not in metrics_conc.columns and 'VES' in metrics_conc.columns:
876
  metrics_conc['valid_efficency_score'] = metrics_conc['VES']
877
 
878
- if 'VES' not in metrics_conc.columns and 'valid_efficency_score' in metrics_conc.columns:
879
  metrics_conc['VES'] = metrics_conc['valid_efficency_score']
880
 
881
  eval_text = generate_eval_text("End evaluation")
 
734
  metrics_conc = target_df
735
  if 'valid_efficency_score' not in metrics_conc.columns:
736
  metrics_conc['valid_efficency_score'] = metrics_conc['VES']
737
+ if 'VES' not in metrics_conc.columns:
738
+ metrics_conc['VES'] = metrics_conc['valid_efficency_score']
739
  eval_text = generate_eval_text("End evaluation")
740
  yield gr.Markdown(eval_text, visible=True), gr.Image(), gr.Markdown(), gr.Markdown(), gr.Markdown(), metrics_conc, *[predictions_dict[model][columns_to_visulize] for model in model_list]
741
 
 
870
  db_path_name="db_path"
871
  )
872
  else:
873
+
874
  metrics_df_model = us.evaluate_answer(predictions_dict[model])
875
  metrics_df_model['model'] = model
876
  metrics_conc = pd.concat([metrics_conc, metrics_df_model], ignore_index=True)
877
 
878
+ if 'valid_efficency_score' not in metrics_conc.columns:
879
  metrics_conc['valid_efficency_score'] = metrics_conc['VES']
880
 
881
+ if 'VES' not in metrics_conc.columns:
882
  metrics_conc['VES'] = metrics_conc['valid_efficency_score']
883
 
884
  eval_text = generate_eval_text("End evaluation")
requirements.txt CHANGED
@@ -22,7 +22,7 @@ statistics>=1.0.3.5
22
  transformers>=4.49.0
23
  fsspec==2024.12.0
24
  pytest
25
-
26
  # Docs dependencies
27
  mkdocs>=1.5.3
28
  mkdocs-material>=9.4.7
 
22
  transformers>=4.49.0
23
  fsspec==2024.12.0
24
  pytest
25
+ json
26
  # Docs dependencies
27
  mkdocs>=1.5.3
28
  mkdocs-material>=9.4.7
utilities.py CHANGED
@@ -10,6 +10,7 @@ from qatch.evaluate_dataset.metrics_evaluators import CellPrecision, CellRecall,
10
  import qatch.evaluate_dataset.orchestrator_evaluator as eva
11
  import utils_get_db_tables_info
12
  #import tiktoken
 
13
  from transformers import AutoTokenizer
14
 
15
  def extract_tables(file_path):
 
10
  import qatch.evaluate_dataset.orchestrator_evaluator as eva
11
  import utils_get_db_tables_info
12
  #import tiktoken
13
+ import json
14
  from transformers import AutoTokenizer
15
 
16
  def extract_tables(file_path):