Spaces:
Sleeping
Sleeping
Fix requirements
#30
by
franceth
- opened
- app.py +5 -2
- requirements.txt +1 -1
- utilities.py +1 -0
app.py
CHANGED
@@ -734,6 +734,8 @@ with gr.Blocks(theme='shivi/calm_seafoam', css_paths='style.css', js=js_func) as
|
|
734 |
metrics_conc = target_df
|
735 |
if 'valid_efficency_score' not in metrics_conc.columns:
|
736 |
metrics_conc['valid_efficency_score'] = metrics_conc['VES']
|
|
|
|
|
737 |
eval_text = generate_eval_text("End evaluation")
|
738 |
yield gr.Markdown(eval_text, visible=True), gr.Image(), gr.Markdown(), gr.Markdown(), gr.Markdown(), metrics_conc, *[predictions_dict[model][columns_to_visulize] for model in model_list]
|
739 |
|
@@ -868,14 +870,15 @@ with gr.Blocks(theme='shivi/calm_seafoam', css_paths='style.css', js=js_func) as
|
|
868 |
db_path_name="db_path"
|
869 |
)
|
870 |
else:
|
|
|
871 |
metrics_df_model = us.evaluate_answer(predictions_dict[model])
|
872 |
metrics_df_model['model'] = model
|
873 |
metrics_conc = pd.concat([metrics_conc, metrics_df_model], ignore_index=True)
|
874 |
|
875 |
-
if 'valid_efficency_score' not in metrics_conc.columns
|
876 |
metrics_conc['valid_efficency_score'] = metrics_conc['VES']
|
877 |
|
878 |
-
if 'VES' not in metrics_conc.columns
|
879 |
metrics_conc['VES'] = metrics_conc['valid_efficency_score']
|
880 |
|
881 |
eval_text = generate_eval_text("End evaluation")
|
|
|
734 |
metrics_conc = target_df
|
735 |
if 'valid_efficency_score' not in metrics_conc.columns:
|
736 |
metrics_conc['valid_efficency_score'] = metrics_conc['VES']
|
737 |
+
if 'VES' not in metrics_conc.columns:
|
738 |
+
metrics_conc['VES'] = metrics_conc['valid_efficency_score']
|
739 |
eval_text = generate_eval_text("End evaluation")
|
740 |
yield gr.Markdown(eval_text, visible=True), gr.Image(), gr.Markdown(), gr.Markdown(), gr.Markdown(), metrics_conc, *[predictions_dict[model][columns_to_visulize] for model in model_list]
|
741 |
|
|
|
870 |
db_path_name="db_path"
|
871 |
)
|
872 |
else:
|
873 |
+
|
874 |
metrics_df_model = us.evaluate_answer(predictions_dict[model])
|
875 |
metrics_df_model['model'] = model
|
876 |
metrics_conc = pd.concat([metrics_conc, metrics_df_model], ignore_index=True)
|
877 |
|
878 |
+
if 'valid_efficency_score' not in metrics_conc.columns:
|
879 |
metrics_conc['valid_efficency_score'] = metrics_conc['VES']
|
880 |
|
881 |
+
if 'VES' not in metrics_conc.columns:
|
882 |
metrics_conc['VES'] = metrics_conc['valid_efficency_score']
|
883 |
|
884 |
eval_text = generate_eval_text("End evaluation")
|
requirements.txt
CHANGED
@@ -22,7 +22,7 @@ statistics>=1.0.3.5
|
|
22 |
transformers>=4.49.0
|
23 |
fsspec==2024.12.0
|
24 |
pytest
|
25 |
-
|
26 |
# Docs dependencies
|
27 |
mkdocs>=1.5.3
|
28 |
mkdocs-material>=9.4.7
|
|
|
22 |
transformers>=4.49.0
|
23 |
fsspec==2024.12.0
|
24 |
pytest
|
25 |
+
json
|
26 |
# Docs dependencies
|
27 |
mkdocs>=1.5.3
|
28 |
mkdocs-material>=9.4.7
|
utilities.py
CHANGED
@@ -10,6 +10,7 @@ from qatch.evaluate_dataset.metrics_evaluators import CellPrecision, CellRecall,
|
|
10 |
import qatch.evaluate_dataset.orchestrator_evaluator as eva
|
11 |
import utils_get_db_tables_info
|
12 |
#import tiktoken
|
|
|
13 |
from transformers import AutoTokenizer
|
14 |
|
15 |
def extract_tables(file_path):
|
|
|
10 |
import qatch.evaluate_dataset.orchestrator_evaluator as eva
|
11 |
import utils_get_db_tables_info
|
12 |
#import tiktoken
|
13 |
+
import json
|
14 |
from transformers import AutoTokenizer
|
15 |
|
16 |
def extract_tables(file_path):
|