Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
fix: fix bug in loading more data
Browse files- app.py +1 -1
- src/envs.py +1 -2
- src/leaderboard/read_evals.py +6 -2
app.py
CHANGED
@@ -11,7 +11,7 @@ from src.about import (
|
|
11 |
from src.display.css_html_js import custom_css
|
12 |
from src.leaderboard.read_evals import get_raw_eval_results, get_leaderboard_df
|
13 |
|
14 |
-
from src.envs import API,
|
15 |
from utils import update_table, update_metric, update_table_long_doc, upload_file, get_default_cols
|
16 |
from src.benchmarks import DOMAIN_COLS_QA, LANG_COLS_QA, DOMAIN_COLS_LONG_DOC, LANG_COLS_LONG_DOC, metric_list
|
17 |
|
|
|
11 |
from src.display.css_html_js import custom_css
|
12 |
from src.leaderboard.read_evals import get_raw_eval_results, get_leaderboard_df
|
13 |
|
14 |
+
from src.envs import API, EVAL_RESULTS_PATH, REPO_ID, RESULTS_REPO, TOKEN
|
15 |
from utils import update_table, update_metric, update_table_long_doc, upload_file, get_default_cols
|
16 |
from src.benchmarks import DOMAIN_COLS_QA, LANG_COLS_QA, DOMAIN_COLS_LONG_DOC, LANG_COLS_LONG_DOC, metric_list
|
17 |
|
src/envs.py
CHANGED
@@ -16,7 +16,6 @@ RESULTS_REPO = f"{OWNER}/results"
|
|
16 |
CACHE_PATH = os.getenv("HF_HOME", ".")
|
17 |
|
18 |
# Local caches
|
19 |
-
|
20 |
-
EVAL_RESULTS_PATH = "/Users/nanwang/Codes/huggingface/nan/leaderboard/toys/toydata/results" #os.path.join(CACHE_PATH, "eval-results")
|
21 |
|
22 |
API = HfApi(token=TOKEN)
|
|
|
16 |
CACHE_PATH = os.getenv("HF_HOME", ".")
|
17 |
|
18 |
# Local caches
|
19 |
+
EVAL_RESULTS_PATH = "/Users/nanwang/Codes/huggingface/nan/results/demo-leaderboard" #os.path.join(CACHE_PATH, "eval-results")
|
|
|
20 |
|
21 |
API = HfApi(token=TOKEN)
|
src/leaderboard/read_evals.py
CHANGED
@@ -92,7 +92,7 @@ class FullEvalResult:
|
|
92 |
results[eval_result.eval_name][COL_NAME_RETRIEVAL_MODEL] = self.retrieval_model
|
93 |
results[eval_result.eval_name][COL_NAME_RERANKING_MODEL] = self.reranking_model
|
94 |
|
95 |
-
print(f'result loaded: {eval_result.eval_name}')
|
96 |
for result in eval_result.results:
|
97 |
# add result for each domain, language, and dataset
|
98 |
domain = result["domain"]
|
@@ -127,7 +127,11 @@ def get_raw_eval_results(results_path: str) -> List[FullEvalResult]:
|
|
127 |
eval_results = {}
|
128 |
for model_result_filepath in model_result_filepaths:
|
129 |
# create evaluation results
|
130 |
-
|
|
|
|
|
|
|
|
|
131 |
print(f'file loaded: {model_result_filepath}')
|
132 |
eval_name = eval_result.eval_name
|
133 |
eval_results[eval_name] = eval_result
|
|
|
92 |
results[eval_result.eval_name][COL_NAME_RETRIEVAL_MODEL] = self.retrieval_model
|
93 |
results[eval_result.eval_name][COL_NAME_RERANKING_MODEL] = self.reranking_model
|
94 |
|
95 |
+
# print(f'result loaded: {eval_result.eval_name}')
|
96 |
for result in eval_result.results:
|
97 |
# add result for each domain, language, and dataset
|
98 |
domain = result["domain"]
|
|
|
127 |
eval_results = {}
|
128 |
for model_result_filepath in model_result_filepaths:
|
129 |
# create evaluation results
|
130 |
+
try:
|
131 |
+
eval_result = FullEvalResult.init_from_json_file(model_result_filepath)
|
132 |
+
except UnicodeDecodeError as e:
|
133 |
+
print(f"loading file failed. {model_result_filepath}")
|
134 |
+
continue
|
135 |
print(f'file loaded: {model_result_filepath}')
|
136 |
eval_name = eval_result.eval_name
|
137 |
eval_results[eval_name] = eval_result
|