Clémentine commited on
Commit
002172c
1 Parent(s): a163e5c

Fix requirements for mistral models - to change once transformers gets updated.

Browse files

We need hf_hub >= 0.18 for collections, transformers >= 4.34 for mistral, and both depend on uncompatible versions of tokenizers. Installing transformers from a fork for now

Files changed (2) hide show
  1. requirements.txt +4 -74
  2. src/leaderboard/read_evals.py +7 -4
requirements.txt CHANGED
@@ -1,90 +1,20 @@
1
- accelerate==0.23.0
2
- aiofiles==23.1.0
3
- aiohttp==3.8.4
4
- aiosignal==1.3.1
5
- altair==4.2.2
6
- anyio==3.7.1
7
  APScheduler==3.10.1
8
- async-timeout==4.0.2
9
- attrs==23.1.0
10
- beautifulsoup4==4.12.2
11
  black==23.11.0
12
- bs4==0.0.1
13
- certifi==2022.12.7
14
- charset-normalizer==3.1.0
15
  click==8.1.3
16
- contourpy==1.0.7
17
- cycler==0.11.0
18
  datasets==2.14.5
19
- debugpy==1.6.7
20
- decorator==5.1.1
21
- dill==0.3.6
22
- entrypoints==0.4
23
- fastapi==0.104.1
24
- ffmpy==0.3.0
25
- filelock==3.11.0
26
- fonttools==4.39.3
27
- frozenlist==1.3.3
28
- fsspec==2023.6.0
29
- gitdb==4.0.10
30
- GitPython==3.1.32
31
  gradio==4.3.0
32
  gradio_client==0.7.0
33
- h11==0.14.0
34
- httpcore==0.17.0
35
- httpx==0.24.0
36
- huggingface-hub==0.18.0
37
- idna==3.4
38
- Jinja2==3.1.2
39
- jsonschema==4.17.3
40
- kiwisolver==1.4.4
41
- linkify-it-py==2.0.0
42
  markdown-it-py==2.2.0
43
  MarkupSafe==2.1.2
44
  matplotlib==3.7.1
45
- mdit-py-plugins==0.3.3
46
- mdurl==0.1.2
47
- multidict==6.0.4
48
  numpy==1.24.2
49
- orjson==3.8.10
50
- packaging==23.1
51
  pandas==2.0.0
52
- Pillow==9.5.0
53
  plotly==5.14.1
54
- pyarrow==11.0.0
55
- pydantic==2.4.2
56
- pydub==0.25.1
57
- pyparsing==3.0.9
58
- pyrsistent==0.19.3
59
  python-dateutil==2.8.2
60
- python-multipart==0.0.6
61
- pytz==2023.3
62
- pytz-deprecation-shim==0.1.0.post0
63
- PyYAML==6.0
64
  requests==2.28.2
65
  semantic-version==2.10.0
66
- six==1.16.0
67
- sniffio==1.3.0
68
- soupsieve==2.4.1
69
- stack-data==0.6.2
70
- starlette==0.27.0
71
- sympy==1.12
72
- tenacity==8.2.3
73
- tiktoken==0.4.0
74
- tokenizers==0.13.3
75
- tomli==2.0.1
76
- tomlkit==0.12.0
77
- toolz==0.12.0
78
  tqdm==4.65.0
79
- traitlets==5.9.0
80
- transformers==4.33.1
81
- triton==2.0.0
82
- typer==0.9.0
83
- typing_extensions==4.8.0
84
- tzdata==2023.3
85
- tzlocal==4.3
86
- uc-micro-py==1.0.1
87
- urllib3==1.26.15
88
- uvicorn==0.21.1
89
- websockets==11.0.1
90
- yarl==1.8.2
 
 
 
 
 
 
 
1
  APScheduler==3.10.1
 
 
 
2
  black==23.11.0
 
 
 
3
  click==8.1.3
 
 
4
  datasets==2.14.5
 
 
 
 
 
 
 
 
 
 
 
 
5
  gradio==4.3.0
6
  gradio_client==0.7.0
7
+ huggingface-hub>=0.18.0
 
 
 
 
 
 
 
 
8
  markdown-it-py==2.2.0
9
  MarkupSafe==2.1.2
10
  matplotlib==3.7.1
 
 
 
11
  numpy==1.24.2
 
 
12
  pandas==2.0.0
 
13
  plotly==5.14.1
 
 
 
 
 
14
  python-dateutil==2.8.2
 
 
 
 
15
  requests==2.28.2
16
  semantic-version==2.10.0
 
 
 
 
 
 
 
 
 
 
 
 
17
  tqdm==4.65.0
18
+ git+https://github.com/clefourrier/transformers.git@req-fix#egg=transformers
19
+ #transformers==4.35.1
20
+ tokenizers>=0.15.0
 
 
 
 
 
 
 
 
 
src/leaderboard/read_evals.py CHANGED
@@ -56,10 +56,13 @@ class EvalResult:
56
  org = org_and_model[0]
57
  model = org_and_model[1]
58
  result_key = f"{org}_{model}_{precision}"
 
59
 
60
- still_on_hub = is_model_on_hub(
61
- "/".join(org_and_model), config.get("model_sha", "main"), trust_remote_code=True
62
- )[0]
 
 
63
 
64
  # Extract results available in this file (some results are split in several files)
65
  results = {}
@@ -91,7 +94,7 @@ class EvalResult:
91
 
92
  return self(
93
  eval_name=result_key,
94
- full_model="/".join(org_and_model),
95
  org=org,
96
  model=model,
97
  results=results,
 
56
  org = org_and_model[0]
57
  model = org_and_model[1]
58
  result_key = f"{org}_{model}_{precision}"
59
+ full_model = "/".join(org_and_model)
60
 
61
+ still_on_hub, error = is_model_on_hub(
62
+ full_model, config.get("model_sha", "main"), trust_remote_code=True
63
+ )
64
+ if not still_on_hub:
65
+ print(full_model, error)
66
 
67
  # Extract results available in this file (some results are split in several files)
68
  results = {}
 
94
 
95
  return self(
96
  eval_name=result_key,
97
+ full_model=full_model,
98
  org=org,
99
  model=model,
100
  results=results,