Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
new tasks
Browse files
src/leaderboard/read_evals.py
CHANGED
@@ -442,7 +442,7 @@ def get_raw_eval_results(results_path: str, requests_path: str, metadata) -> lis
|
|
442 |
missing_results_for_task[task_name] = [f"{v.full_model}|{v.org_and_model}"]
|
443 |
if r[AutoEvalColumn.lang.name] is None or r[AutoEvalColumn.lang.name] == "?":
|
444 |
missing_metadata.append(f"{v.full_model}")
|
445 |
-
all_models.append((v.full_model, v.num_params))
|
446 |
|
447 |
# print('missing_results_for_task', missing_results_for_task)
|
448 |
for task, models in missing_results_for_task.items():
|
|
|
442 |
missing_results_for_task[task_name] = [f"{v.full_model}|{v.org_and_model}"]
|
443 |
if r[AutoEvalColumn.lang.name] is None or r[AutoEvalColumn.lang.name] == "?":
|
444 |
missing_metadata.append(f"{v.full_model}")
|
445 |
+
all_models.append((v.full_model, v.num_params, v.still_on_hub))
|
446 |
|
447 |
# print('missing_results_for_task', missing_results_for_task)
|
448 |
for task, models in missing_results_for_task.items():
|