Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
Rename
Browse files- app.py +4 -4
- src/display/utils.py +3 -3
- src/submission/check_validity.py +2 -2
- src/submission/submit.py +2 -2
app.py
CHANGED
@@ -30,10 +30,10 @@ from src.display.utils import (
|
|
30 |
TYPES,
|
31 |
AddSpecialTokens,
|
32 |
AutoEvalColumn,
|
|
|
33 |
ModelType,
|
34 |
NumFewShots,
|
35 |
Precision,
|
36 |
-
Version,
|
37 |
VllmVersion,
|
38 |
fields,
|
39 |
)
|
@@ -203,7 +203,7 @@ leaderboard_df = filter_models(
|
|
203 |
[i.value.name for i in Precision],
|
204 |
[i.value.name for i in AddSpecialTokens],
|
205 |
[i.value.name for i in NumFewShots],
|
206 |
-
[i.value.name for i in
|
207 |
[i.value.name for i in VllmVersion],
|
208 |
)
|
209 |
# Initialize columns
|
@@ -418,8 +418,8 @@ with gr.Blocks() as demo_leaderboard:
|
|
418 |
)
|
419 |
filter_columns_version = gr.CheckboxGroup(
|
420 |
label="llm-jp-eval version",
|
421 |
-
choices=[i.value.name for i in
|
422 |
-
value=[i.value.name for i in
|
423 |
elem_id="filter-columns-version",
|
424 |
)
|
425 |
filter_columns_vllm = gr.CheckboxGroup(
|
|
|
30 |
TYPES,
|
31 |
AddSpecialTokens,
|
32 |
AutoEvalColumn,
|
33 |
+
LLMJpEvalVersion,
|
34 |
ModelType,
|
35 |
NumFewShots,
|
36 |
Precision,
|
|
|
37 |
VllmVersion,
|
38 |
fields,
|
39 |
)
|
|
|
203 |
[i.value.name for i in Precision],
|
204 |
[i.value.name for i in AddSpecialTokens],
|
205 |
[i.value.name for i in NumFewShots],
|
206 |
+
[i.value.name for i in LLMJpEvalVersion],
|
207 |
[i.value.name for i in VllmVersion],
|
208 |
)
|
209 |
# Initialize columns
|
|
|
418 |
)
|
419 |
filter_columns_version = gr.CheckboxGroup(
|
420 |
label="llm-jp-eval version",
|
421 |
+
choices=[i.value.name for i in LLMJpEvalVersion],
|
422 |
+
value=[i.value.name for i in LLMJpEvalVersion],
|
423 |
elem_id="filter-columns-version",
|
424 |
)
|
425 |
filter_columns_vllm = gr.CheckboxGroup(
|
src/display/utils.py
CHANGED
@@ -148,15 +148,15 @@ class NumFewShots(Enum):
|
|
148 |
return NumFewShots.Unknown
|
149 |
|
150 |
|
151 |
-
class
|
152 |
current = ModelDetails("v1.4.1")
|
153 |
Unknown = ModelDetails("?")
|
154 |
|
155 |
def from_str(version):
|
156 |
if version == "1.4.1":
|
157 |
-
return
|
158 |
else:
|
159 |
-
return
|
160 |
|
161 |
|
162 |
class VllmVersion(Enum):
|
|
|
148 |
return NumFewShots.Unknown
|
149 |
|
150 |
|
151 |
+
class LLMJpEvalVersion(Enum):
|
152 |
current = ModelDetails("v1.4.1")
|
153 |
Unknown = ModelDetails("?")
|
154 |
|
155 |
def from_str(version):
|
156 |
if version == "1.4.1":
|
157 |
+
return LLMJpEvalVersion.current
|
158 |
else:
|
159 |
+
return LLMJpEvalVersion.Unknown
|
160 |
|
161 |
|
162 |
class VllmVersion(Enum):
|
src/submission/check_validity.py
CHANGED
@@ -9,7 +9,7 @@ from huggingface_hub.hf_api import ModelInfo
|
|
9 |
from transformers import AutoConfig
|
10 |
from transformers.models.auto.tokenization_auto import AutoTokenizer
|
11 |
|
12 |
-
from src.display.utils import
|
13 |
|
14 |
|
15 |
def check_model_card(repo_id: str) -> tuple[bool, str]:
|
@@ -110,7 +110,7 @@ def already_submitted_models(requested_models_dir: str) -> set[str]:
|
|
110 |
continue
|
111 |
|
112 |
# Version.v1_4_1.value.name γδ½Ώη¨γγ¦γγΌγΈγ§γ³ζ
ε ±γεεΎ
|
113 |
-
version = info.get("llm_jp_eval_version",
|
114 |
vllm_version = info.get("vllm_version", VllmVersion.current.value.name)
|
115 |
file_names.append(
|
116 |
f"{info['model']}_{info['precision']}_{info['add_special_tokens']}_{version}_{vllm_version}"
|
|
|
9 |
from transformers import AutoConfig
|
10 |
from transformers.models.auto.tokenization_auto import AutoTokenizer
|
11 |
|
12 |
+
from src.display.utils import LLMJpEvalVersion, VllmVersion
|
13 |
|
14 |
|
15 |
def check_model_card(repo_id: str) -> tuple[bool, str]:
|
|
|
110 |
continue
|
111 |
|
112 |
# Version.v1_4_1.value.name γδ½Ώη¨γγ¦γγΌγΈγ§γ³ζ
ε ±γεεΎ
|
113 |
+
version = info.get("llm_jp_eval_version", LLMJpEvalVersion.current.value.name)
|
114 |
vllm_version = info.get("vllm_version", VllmVersion.current.value.name)
|
115 |
file_names.append(
|
116 |
f"{info['model']}_{info['precision']}_{info['add_special_tokens']}_{version}_{vllm_version}"
|
src/submission/submit.py
CHANGED
@@ -3,7 +3,7 @@ import os
|
|
3 |
from datetime import datetime, timezone
|
4 |
|
5 |
from src.display.formatting import styled_error, styled_message, styled_warning
|
6 |
-
from src.display.utils import
|
7 |
from src.envs import API, EVAL_REQUESTS_PATH, QUEUE_REPO, TOKEN
|
8 |
from src.submission.check_validity import already_submitted_models, check_model_card, is_model_on_hub
|
9 |
|
@@ -23,7 +23,7 @@ def add_new_eval(
|
|
23 |
if not REQUESTED_MODELS:
|
24 |
REQUESTED_MODELS, USERS_TO_SUBMISSION_DATES = already_submitted_models(EVAL_REQUESTS_PATH)
|
25 |
|
26 |
-
current_version =
|
27 |
current_vllm_version = VllmVersion.current.value.name
|
28 |
|
29 |
# γγΌγΈγ§γ³ζ
ε ±γε«γγιθ€γγ§γγ―
|
|
|
3 |
from datetime import datetime, timezone
|
4 |
|
5 |
from src.display.formatting import styled_error, styled_message, styled_warning
|
6 |
+
from src.display.utils import LLMJpEvalVersion, VllmVersion
|
7 |
from src.envs import API, EVAL_REQUESTS_PATH, QUEUE_REPO, TOKEN
|
8 |
from src.submission.check_validity import already_submitted_models, check_model_card, is_model_on_hub
|
9 |
|
|
|
23 |
if not REQUESTED_MODELS:
|
24 |
REQUESTED_MODELS, USERS_TO_SUBMISSION_DATES = already_submitted_models(EVAL_REQUESTS_PATH)
|
25 |
|
26 |
+
current_version = LLMJpEvalVersion.current.value.name
|
27 |
current_vllm_version = VllmVersion.current.value.name
|
28 |
|
29 |
# γγΌγΈγ§γ³ζ
ε ±γε«γγιθ€γγ§γγ―
|