Clémentine commited on
Commit
5491f2d
1 Parent(s): ed118a6

updated GPTQ display!

Browse files
app.py CHANGED
@@ -253,14 +253,14 @@ def select_columns(df: pd.DataFrame, columns: list) -> pd.DataFrame:
253
 
254
 
255
  NUMERIC_INTERVALS = {
256
- "Unknown": pd.Interval(-1, 0, closed="right"),
257
- "0~1.5B": pd.Interval(0, 1.5, closed="right"),
258
- "1.5~3B": pd.Interval(1.5, 3, closed="right"),
259
- "3~7B": pd.Interval(3, 7, closed="right"),
260
- "7~13B": pd.Interval(7, 13, closed="right"),
261
- "13~35B": pd.Interval(13, 35, closed="right"),
262
- "35~60B": pd.Interval(35, 60, closed="right"),
263
- "60B+": pd.Interval(60, 10000, closed="right"),
264
  }
265
 
266
 
@@ -378,7 +378,7 @@ with demo:
378
  elem_id="filter-columns-precision",
379
  )
380
  filter_columns_size = gr.CheckboxGroup(
381
- label="Model sizes",
382
  choices=list(NUMERIC_INTERVALS.keys()),
383
  value=list(NUMERIC_INTERVALS.keys()),
384
  interactive=True,
 
253
 
254
 
255
  NUMERIC_INTERVALS = {
256
+ "?": pd.Interval(-1, 0, closed="right"),
257
+ "0~1.5": pd.Interval(0, 1.5, closed="right"),
258
+ "1.5~3": pd.Interval(1.5, 3, closed="right"),
259
+ "3~7": pd.Interval(3, 7, closed="right"),
260
+ "7~13": pd.Interval(7, 13, closed="right"),
261
+ "13~35": pd.Interval(13, 35, closed="right"),
262
+ "35~60": pd.Interval(35, 60, closed="right"),
263
+ "60+": pd.Interval(60, 10000, closed="right"),
264
  }
265
 
266
 
 
378
  elem_id="filter-columns-precision",
379
  )
380
  filter_columns_size = gr.CheckboxGroup(
381
+ label="Model sizes (in billions of parameters)",
382
  choices=list(NUMERIC_INTERVALS.keys()),
383
  value=list(NUMERIC_INTERVALS.keys()),
384
  interactive=True,
model_info_cache.pkl CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:94e311e2414e80b8eb5e50844c2e79daa4bd3bb6be516fc2448bd05242d125f9
3
- size 3656702
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f98e18ea0fcf11737e108f966f6d8b09120c6a2c231b70e9e242e9bba3145a47
3
+ size 3780284
model_size_cache.pkl CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4aedc91b51cf257cbe3e26a1fdd99e19250bacfa619a64dd85e67d4ff383130f
3
- size 75455
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ace7167a258f711fa7ffeaadddc6ebef8ccb92651dce8b805228c2f18c988958
3
+ size 75324
src/display_models/get_model_metadata.py CHANGED
@@ -45,13 +45,15 @@ def get_model_infos_from_hub(leaderboard_data: List[dict]):
45
  model_data[AutoEvalColumn.license.name] = None
46
  model_data[AutoEvalColumn.likes.name] = None
47
  if model_name not in model_size_cache:
48
- model_size_cache[model_name] = get_model_size(model_name, None)
 
49
  model_data[AutoEvalColumn.params.name] = model_size_cache[model_name]
50
 
51
  model_data[AutoEvalColumn.license.name] = get_model_license(model_info)
52
  model_data[AutoEvalColumn.likes.name] = get_model_likes(model_info)
53
  if model_name not in model_size_cache:
54
- model_size_cache[model_name] = get_model_size(model_name, model_info)
 
55
  model_data[AutoEvalColumn.params.name] = model_size_cache[model_name]
56
 
57
  # save cache to disk in pickle format
 
45
  model_data[AutoEvalColumn.license.name] = None
46
  model_data[AutoEvalColumn.likes.name] = None
47
  if model_name not in model_size_cache:
48
+ size_factor = 8 if model_data["Precision"] == "GPTQ" else 1
49
+ model_size_cache[model_name] = size_factor * get_model_size(model_name, None)
50
  model_data[AutoEvalColumn.params.name] = model_size_cache[model_name]
51
 
52
  model_data[AutoEvalColumn.license.name] = get_model_license(model_info)
53
  model_data[AutoEvalColumn.likes.name] = get_model_likes(model_info)
54
  if model_name not in model_size_cache:
55
+ size_factor = 8 if model_data["Precision"] == "GPTQ" else 1
56
+ model_size_cache[model_name] = size_factor * get_model_size(model_name, model_info)
57
  model_data[AutoEvalColumn.params.name] = model_size_cache[model_name]
58
 
59
  # save cache to disk in pickle format
src/display_models/read_results.py CHANGED
@@ -81,6 +81,8 @@ def parse_eval_result(json_filepath: str) -> Tuple[str, list[dict]]:
81
  model_split = model.split("/", 1)
82
 
83
  precision = config.get("model_dtype")
 
 
84
 
85
  model = model_split[-1]
86
 
 
81
  model_split = model.split("/", 1)
82
 
83
  precision = config.get("model_dtype")
84
+ if precision == "None":
85
+ precision = "GPTQ"
86
 
87
  model = model_split[-1]
88