davanstrien HF staff commited on
Commit
9edb8b6
1 Parent(s): 02881ff

chore: Update requirements_cpu.txt with new dependencies

Browse files
notebooks/requirements_cpu.txt CHANGED
@@ -1,331 +1,4 @@
1
- # This file was autogenerated by uv via the following command:
2
- # uv pip compile notebooks/requirements_cpu.in -o notebooks/requirements_cpu.txt
3
- aiohttp==3.9.5
4
- # via
5
- # datasets
6
- # fsspec
7
- # llama-index-core
8
- # llama-index-legacy
9
- aiosignal==1.3.1
10
- # via aiohttp
11
- annotated-types==0.7.0
12
- # via pydantic
13
- anyio==4.4.0
14
- # via
15
- # httpx
16
- # openai
17
- attrs==23.2.0
18
- # via
19
- # aiohttp
20
- # jsonschema
21
- # referencing
22
- beautifulsoup4==4.12.3
23
- # via llama-index-readers-file
24
- certifi==2024.6.2
25
- # via
26
- # httpcore
27
- # httpx
28
- # requests
29
- charset-normalizer==3.3.2
30
- # via requests
31
- click==8.1.7
32
- # via nltk
33
- cloudpickle==3.0.0
34
- # via outlines
35
- dataclasses-json==0.6.7
36
- # via
37
- # llama-index-core
38
- # llama-index-legacy
39
- datasets==2.20.0
40
- # via
41
- # -r notebooks/requirements_cpu.in
42
- # outlines
43
- deprecated==1.2.14
44
- # via
45
- # llama-index-core
46
- # llama-index-legacy
47
- dill==0.3.8
48
- # via
49
- # datasets
50
- # multiprocess
51
- dirtyjson==1.0.8
52
- # via
53
- # llama-index-core
54
- # llama-index-legacy
55
- diskcache==5.6.3
56
- # via outlines
57
- distro==1.9.0
58
- # via openai
59
- filelock==3.15.1
60
- # via
61
- # datasets
62
- # huggingface-hub
63
- frozenlist==1.4.1
64
- # via
65
- # aiohttp
66
- # aiosignal
67
- fsspec==2024.5.0
68
- # via
69
- # datasets
70
- # huggingface-hub
71
- # llama-index-core
72
- # llama-index-legacy
73
- greenlet==3.0.3
74
- # via sqlalchemy
75
- h11==0.14.0
76
- # via httpcore
77
- httpcore==1.0.5
78
- # via httpx
79
- httpx==0.27.0
80
- # via
81
- # llama-index-core
82
- # llama-index-legacy
83
- # llamaindex-py-client
84
- # openai
85
- huggingface-hub==0.23.4
86
- # via datasets
87
- idna==3.7
88
- # via
89
- # anyio
90
- # httpx
91
- # requests
92
- # yarl
93
- interegular==0.3.3
94
- # via outlines
95
- jinja2==3.1.4
96
- # via outlines
97
- joblib==1.4.2
98
- # via nltk
99
- jsonschema==4.22.0
100
- # via outlines
101
- jsonschema-specifications==2023.12.1
102
- # via jsonschema
103
- lark==1.1.9
104
- # via outlines
105
- llama-index==0.10.46
106
- # via -r notebooks/requirements_cpu.in
107
- llama-index-agent-openai==0.2.7
108
- # via
109
- # llama-index
110
- # llama-index-program-openai
111
- llama-index-cli==0.1.12
112
- # via llama-index
113
- llama-index-core==0.10.46
114
- # via
115
- # llama-index
116
- # llama-index-agent-openai
117
- # llama-index-cli
118
- # llama-index-embeddings-openai
119
- # llama-index-indices-managed-llama-cloud
120
- # llama-index-llms-openai
121
- # llama-index-multi-modal-llms-openai
122
- # llama-index-program-openai
123
- # llama-index-question-gen-openai
124
- # llama-index-readers-file
125
- # llama-index-readers-llama-parse
126
- # llama-parse
127
- llama-index-embeddings-openai==0.1.10
128
- # via
129
- # llama-index
130
- # llama-index-cli
131
- llama-index-indices-managed-llama-cloud==0.1.6
132
- # via llama-index
133
- llama-index-legacy==0.9.48
134
- # via llama-index
135
- llama-index-llms-openai==0.1.22
136
- # via
137
- # llama-index
138
- # llama-index-agent-openai
139
- # llama-index-cli
140
- # llama-index-multi-modal-llms-openai
141
- # llama-index-program-openai
142
- # llama-index-question-gen-openai
143
- llama-index-multi-modal-llms-openai==0.1.6
144
- # via llama-index
145
- llama-index-program-openai==0.1.6
146
- # via
147
- # llama-index
148
- # llama-index-question-gen-openai
149
- llama-index-question-gen-openai==0.1.3
150
- # via llama-index
151
- llama-index-readers-file==0.1.25
152
- # via llama-index
153
- llama-index-readers-llama-parse==0.1.4
154
- # via llama-index
155
- llama-parse==0.4.4
156
- # via llama-index-readers-llama-parse
157
- llamaindex-py-client==0.1.19
158
- # via
159
- # llama-index-core
160
- # llama-index-indices-managed-llama-cloud
161
- llvmlite==0.43.0
162
- # via numba
163
- markdown-it-py==3.0.0
164
- # via rich
165
- markupsafe==2.1.5
166
- # via jinja2
167
- marshmallow==3.21.3
168
- # via dataclasses-json
169
- mdurl==0.1.2
170
- # via markdown-it-py
171
- multidict==6.0.5
172
- # via
173
- # aiohttp
174
- # yarl
175
- multiprocess==0.70.16
176
- # via datasets
177
- mypy-extensions==1.0.0
178
- # via typing-inspect
179
- nest-asyncio==1.6.0
180
- # via
181
- # llama-index-core
182
- # llama-index-legacy
183
- # outlines
184
- networkx==3.3
185
- # via
186
- # llama-index-core
187
- # llama-index-legacy
188
- nltk==3.8.1
189
- # via
190
- # llama-index-core
191
- # llama-index-legacy
192
- numba==0.60.0
193
- # via outlines
194
- numpy==1.26.4
195
- # via
196
- # datasets
197
- # llama-index-core
198
- # llama-index-legacy
199
- # numba
200
- # outlines
201
- # pandas
202
- # pyarrow
203
- openai==1.34.0
204
- # via
205
- # llama-index-agent-openai
206
- # llama-index-core
207
- # llama-index-legacy
208
- outlines==0.0.45
209
- # via -r notebooks/requirements_cpu.in
210
- packaging==24.1
211
- # via
212
- # datasets
213
- # huggingface-hub
214
- # marshmallow
215
- pandas==2.2.2
216
- # via
217
- # datasets
218
- # llama-index-core
219
- # llama-index-legacy
220
- pillow==10.3.0
221
- # via llama-index-core
222
- pyairports==2.1.1
223
- # via outlines
224
- pyarrow==16.1.0
225
- # via datasets
226
- pyarrow-hotfix==0.6
227
- # via datasets
228
- pycountry==24.6.1
229
- # via outlines
230
- pydantic==2.7.4
231
- # via
232
- # llamaindex-py-client
233
- # openai
234
- # outlines
235
- pydantic-core==2.18.4
236
- # via pydantic
237
- pygments==2.18.0
238
- # via rich
239
- pypdf==4.2.0
240
- # via llama-index-readers-file
241
- python-dateutil==2.9.0.post0
242
- # via pandas
243
- pytz==2024.1
244
- # via pandas
245
- pyyaml==6.0.1
246
- # via
247
- # datasets
248
- # huggingface-hub
249
- # llama-index-core
250
- referencing==0.35.1
251
- # via
252
- # jsonschema
253
- # jsonschema-specifications
254
- # outlines
255
- regex==2024.5.15
256
- # via
257
- # nltk
258
- # tiktoken
259
- requests==2.32.3
260
- # via
261
- # datasets
262
- # huggingface-hub
263
- # llama-index-core
264
- # llama-index-legacy
265
- # outlines
266
- # tiktoken
267
- rich==13.7.1
268
- # via -r notebooks/requirements_cpu.in
269
- rpds-py==0.18.1
270
- # via
271
- # jsonschema
272
- # referencing
273
- six==1.16.0
274
- # via python-dateutil
275
- sniffio==1.3.1
276
- # via
277
- # anyio
278
- # httpx
279
- # openai
280
- soupsieve==2.5
281
- # via beautifulsoup4
282
- sqlalchemy==2.0.31
283
- # via
284
- # llama-index-core
285
- # llama-index-legacy
286
- striprtf==0.0.26
287
- # via llama-index-readers-file
288
- tenacity==8.3.0
289
- # via
290
- # llama-index-core
291
- # llama-index-legacy
292
- tiktoken==0.7.0
293
- # via
294
- # llama-index-core
295
- # llama-index-legacy
296
- tqdm==4.66.4
297
- # via
298
- # datasets
299
- # huggingface-hub
300
- # llama-index-core
301
- # nltk
302
- # openai
303
- # outlines
304
- typing-extensions==4.12.2
305
- # via
306
- # huggingface-hub
307
- # llama-index-core
308
- # llama-index-legacy
309
- # openai
310
- # outlines
311
- # pydantic
312
- # pydantic-core
313
- # sqlalchemy
314
- # typing-inspect
315
- typing-inspect==0.9.0
316
- # via
317
- # dataclasses-json
318
- # llama-index-core
319
- # llama-index-legacy
320
- tzdata==2024.1
321
- # via pandas
322
- urllib3==2.2.2
323
- # via requests
324
- wrapt==1.16.0
325
- # via
326
- # deprecated
327
- # llama-index-core
328
- xxhash==3.4.1
329
- # via datasets
330
- yarl==1.9.4
331
- # via aiohttp
 
1
+ outlines
2
+ llama_index
3
+ datasets>=2.18.0
4
+ rich
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
notebooks/requirements_gpu.txt CHANGED
@@ -1,517 +1,5 @@
1
- # This file was autogenerated by uv via the following command:
2
- # uv pip compile requirements.in -o requirements.txt
3
- aiohttp==3.9.5
4
- # via
5
- # datasets
6
- # fsspec
7
- # llama-index-core
8
- # llama-index-legacy
9
- # vllm
10
- aiosignal==1.3.1
11
- # via
12
- # aiohttp
13
- # ray
14
- annotated-types==0.7.0
15
- # via pydantic
16
- anyio==4.4.0
17
- # via
18
- # httpx
19
- # openai
20
- # starlette
21
- # watchfiles
22
- async-timeout==4.0.3
23
- # via aiohttp
24
- attrs==23.2.0
25
- # via
26
- # aiohttp
27
- # jsonschema
28
- # referencing
29
- beautifulsoup4==4.12.3
30
- # via llama-index-readers-file
31
- certifi==2024.6.2
32
- # via
33
- # httpcore
34
- # httpx
35
- # requests
36
- charset-normalizer==3.3.2
37
- # via requests
38
- click==8.1.7
39
- # via
40
- # nltk
41
- # ray
42
- # typer
43
- # uvicorn
44
- cloudpickle==3.0.0
45
- # via outlines
46
- cmake==3.29.5.1
47
- # via vllm
48
- dataclasses-json==0.6.7
49
- # via
50
- # llama-index-core
51
- # llama-index-legacy
52
- datasets==2.19.2
53
- # via
54
- # -r requirements.in
55
- # outlines
56
- deprecated==1.2.14
57
- # via
58
- # llama-index-core
59
- # llama-index-legacy
60
- dill==0.3.7
61
- # via
62
- # datasets
63
- # multiprocess
64
- dirtyjson==1.0.8
65
- # via
66
- # llama-index-core
67
- # llama-index-legacy
68
- diskcache==5.6.3
69
- # via outlines
70
- distro==1.9.0
71
- # via openai
72
- dnspython==2.6.1
73
- # via email-validator
74
- email-validator==2.1.1
75
- # via fastapi
76
- exceptiongroup==1.2.1
77
- # via anyio
78
- fastapi==0.111.0
79
- # via vllm
80
- fastapi-cli==0.0.4
81
- # via fastapi
82
- filelock==3.15.1
83
- # via
84
- # datasets
85
- # huggingface-hub
86
- # ray
87
- # torch
88
- # transformers
89
- # triton
90
- # vllm
91
- frozenlist==1.4.1
92
- # via
93
- # aiohttp
94
- # aiosignal
95
- # ray
96
- fsspec==2024.3.1
97
- # via
98
- # datasets
99
- # huggingface-hub
100
- # llama-index-core
101
- # llama-index-legacy
102
- # torch
103
- greenlet==3.0.3
104
- # via sqlalchemy
105
- h11==0.14.0
106
- # via
107
- # httpcore
108
- # uvicorn
109
- httpcore==1.0.5
110
- # via httpx
111
- httptools==0.6.1
112
- # via uvicorn
113
- httpx==0.27.0
114
- # via
115
- # fastapi
116
- # llama-index-core
117
- # llama-index-legacy
118
- # llamaindex-py-client
119
- # openai
120
- huggingface-hub==0.23.3
121
- # via
122
- # datasets
123
- # tokenizers
124
- # transformers
125
- idna==3.7
126
- # via
127
- # anyio
128
- # email-validator
129
- # httpx
130
- # requests
131
- # yarl
132
- interegular==0.3.3
133
- # via
134
- # lm-format-enforcer
135
- # outlines
136
- jinja2==3.1.4
137
- # via
138
- # fastapi
139
- # outlines
140
- # torch
141
- joblib==1.4.2
142
- # via nltk
143
- jsonschema==4.22.0
144
- # via
145
- # outlines
146
- # ray
147
- jsonschema-specifications==2023.12.1
148
- # via jsonschema
149
- lark==1.1.9
150
- # via outlines
151
- llama-index==0.10.44
152
- # via -r requirements.in
153
- llama-index-agent-openai==0.2.7
154
- # via
155
- # llama-index
156
- # llama-index-program-openai
157
- llama-index-cli==0.1.12
158
- # via llama-index
159
- llama-index-core==0.10.44
160
- # via
161
- # llama-index
162
- # llama-index-agent-openai
163
- # llama-index-cli
164
- # llama-index-embeddings-openai
165
- # llama-index-indices-managed-llama-cloud
166
- # llama-index-llms-openai
167
- # llama-index-multi-modal-llms-openai
168
- # llama-index-program-openai
169
- # llama-index-question-gen-openai
170
- # llama-index-readers-file
171
- # llama-index-readers-llama-parse
172
- # llama-parse
173
- llama-index-embeddings-openai==0.1.10
174
- # via
175
- # llama-index
176
- # llama-index-cli
177
- llama-index-indices-managed-llama-cloud==0.1.6
178
- # via llama-index
179
- llama-index-legacy==0.9.48
180
- # via llama-index
181
- llama-index-llms-openai==0.1.22
182
- # via
183
- # llama-index
184
- # llama-index-agent-openai
185
- # llama-index-cli
186
- # llama-index-multi-modal-llms-openai
187
- # llama-index-program-openai
188
- # llama-index-question-gen-openai
189
- llama-index-multi-modal-llms-openai==0.1.6
190
- # via llama-index
191
- llama-index-program-openai==0.1.6
192
- # via
193
- # llama-index
194
- # llama-index-question-gen-openai
195
- llama-index-question-gen-openai==0.1.3
196
- # via llama-index
197
- llama-index-readers-file==0.1.25
198
- # via llama-index
199
- llama-index-readers-llama-parse==0.1.4
200
- # via llama-index
201
- llama-parse==0.4.4
202
- # via llama-index-readers-llama-parse
203
- llamaindex-py-client==0.1.19
204
- # via
205
- # llama-index-core
206
- # llama-index-indices-managed-llama-cloud
207
- llvmlite==0.42.0
208
- # via numba
209
- lm-format-enforcer==0.10.1
210
- # via vllm
211
- markdown-it-py==3.0.0
212
- # via rich
213
- markupsafe==2.1.5
214
- # via jinja2
215
- marshmallow==3.21.3
216
- # via dataclasses-json
217
- mdurl==0.1.2
218
- # via markdown-it-py
219
- mpmath==1.3.0
220
- # via sympy
221
- msgpack==1.0.8
222
- # via ray
223
- multidict==6.0.5
224
- # via
225
- # aiohttp
226
- # yarl
227
- multiprocess==0.70.15
228
- # via datasets
229
- mypy-extensions==1.0.0
230
- # via typing-inspect
231
- nest-asyncio==1.6.0
232
- # via
233
- # llama-index-core
234
- # llama-index-legacy
235
- # outlines
236
- networkx==3.2.1
237
- # via
238
- # llama-index-core
239
- # llama-index-legacy
240
- # torch
241
- ninja==1.11.1.1
242
- # via vllm
243
- nltk==3.8.1
244
- # via
245
- # llama-index-core
246
- # llama-index-legacy
247
- numba==0.59.1
248
- # via outlines
249
- numpy==1.26.4
250
- # via
251
- # datasets
252
- # llama-index-core
253
- # llama-index-legacy
254
- # numba
255
- # outlines
256
- # pandas
257
- # pyarrow
258
- # transformers
259
- # vllm
260
- # xformers
261
- nvidia-cublas-cu12==12.1.3.1
262
- # via
263
- # nvidia-cudnn-cu12
264
- # nvidia-cusolver-cu12
265
- # torch
266
- nvidia-cuda-cupti-cu12==12.1.105
267
- # via torch
268
- nvidia-cuda-nvrtc-cu12==12.1.105
269
- # via torch
270
- nvidia-cuda-runtime-cu12==12.1.105
271
- # via torch
272
- nvidia-cudnn-cu12==8.9.2.26
273
- # via torch
274
- nvidia-cufft-cu12==11.0.2.54
275
- # via torch
276
- nvidia-curand-cu12==10.3.2.106
277
- # via torch
278
- nvidia-cusolver-cu12==11.4.5.107
279
- # via torch
280
- nvidia-cusparse-cu12==12.1.0.106
281
- # via
282
- # nvidia-cusolver-cu12
283
- # torch
284
- nvidia-ml-py==12.555.43
285
- # via vllm
286
- nvidia-nccl-cu12==2.20.5
287
- # via torch
288
- nvidia-nvjitlink-cu12==12.5.40
289
- # via
290
- # nvidia-cusolver-cu12
291
- # nvidia-cusparse-cu12
292
- nvidia-nvtx-cu12==12.1.105
293
- # via torch
294
- openai==1.34.0
295
- # via
296
- # llama-index-agent-openai
297
- # llama-index-core
298
- # llama-index-legacy
299
- # vllm
300
- orjson==3.10.4
301
- # via fastapi
302
- outlines==0.0.43
303
- # via
304
- # -r requirements.in
305
- # vllm
306
- packaging==24.1
307
- # via
308
- # datasets
309
- # huggingface-hub
310
- # lm-format-enforcer
311
- # marshmallow
312
- # ray
313
- # transformers
314
- pandas==2.2.2
315
- # via
316
- # datasets
317
- # llama-index-core
318
- # llama-index-legacy
319
- pillow==10.3.0
320
- # via
321
- # llama-index-core
322
- # vllm
323
- prometheus-client==0.20.0
324
- # via
325
- # prometheus-fastapi-instrumentator
326
- # vllm
327
- prometheus-fastapi-instrumentator==7.0.0
328
- # via vllm
329
- protobuf==5.27.1
330
- # via ray
331
- psutil==5.9.8
332
- # via vllm
333
- py-cpuinfo==9.0.0
334
- # via vllm
335
- pyairports==2.1.1
336
- # via outlines
337
- pyarrow==16.1.0
338
- # via datasets
339
- pyarrow-hotfix==0.6
340
- # via datasets
341
- pycountry==24.6.1
342
- # via outlines
343
- pydantic==2.7.4
344
- # via
345
- # fastapi
346
- # llamaindex-py-client
347
- # lm-format-enforcer
348
- # openai
349
- # outlines
350
- # vllm
351
- pydantic-core==2.18.4
352
- # via pydantic
353
- pygments==2.18.0
354
- # via rich
355
- pypdf==4.2.0
356
- # via llama-index-readers-file
357
- python-dateutil==2.9.0.post0
358
- # via pandas
359
- python-dotenv==1.0.1
360
- # via uvicorn
361
- python-multipart==0.0.9
362
- # via fastapi
363
- pytz==2024.1
364
- # via pandas
365
- pyyaml==6.0.1
366
- # via
367
- # datasets
368
- # huggingface-hub
369
- # llama-index-core
370
- # lm-format-enforcer
371
- # ray
372
- # transformers
373
- # uvicorn
374
- ray==2.24.0
375
- # via vllm
376
- referencing==0.35.1
377
- # via
378
- # jsonschema
379
- # jsonschema-specifications
380
- # outlines
381
- regex==2024.5.15
382
- # via
383
- # nltk
384
- # tiktoken
385
- # transformers
386
- requests==2.32.3
387
- # via
388
- # datasets
389
- # huggingface-hub
390
- # llama-index-core
391
- # llama-index-legacy
392
- # outlines
393
- # ray
394
- # tiktoken
395
- # transformers
396
- # vllm
397
- rich==13.7.1
398
- # via
399
- # -r requirements.in
400
- # typer
401
- rpds-py==0.18.1
402
- # via
403
- # jsonschema
404
- # referencing
405
- safetensors==0.4.3
406
- # via transformers
407
- sentencepiece==0.2.0
408
- # via vllm
409
- shellingham==1.5.4
410
- # via typer
411
- six==1.16.0
412
- # via python-dateutil
413
- sniffio==1.3.1
414
- # via
415
- # anyio
416
- # httpx
417
- # openai
418
- soupsieve==2.5
419
- # via beautifulsoup4
420
- sqlalchemy==2.0.30
421
- # via
422
- # llama-index-core
423
- # llama-index-legacy
424
- starlette==0.37.2
425
- # via
426
- # fastapi
427
- # prometheus-fastapi-instrumentator
428
- striprtf==0.0.26
429
- # via llama-index-readers-file
430
- sympy==1.12.1
431
- # via torch
432
- tenacity==8.3.0
433
- # via
434
- # llama-index-core
435
- # llama-index-legacy
436
- tiktoken==0.7.0
437
- # via
438
- # llama-index-core
439
- # llama-index-legacy
440
- # vllm
441
- tokenizers==0.19.1
442
- # via
443
- # transformers
444
- # vllm
445
- torch==2.3.0
446
- # via
447
- # vllm
448
- # vllm-flash-attn
449
- # xformers
450
- tqdm==4.66.4
451
- # via
452
- # datasets
453
- # huggingface-hub
454
- # llama-index-core
455
- # nltk
456
- # openai
457
- # outlines
458
- # transformers
459
- transformers==4.41.2
460
- # via vllm
461
- triton==2.3.0
462
- # via torch
463
- typer==0.12.3
464
- # via fastapi-cli
465
- typing-extensions==4.12.2
466
- # via
467
- # anyio
468
- # fastapi
469
- # huggingface-hub
470
- # llama-index-core
471
- # llama-index-legacy
472
- # openai
473
- # pydantic
474
- # pydantic-core
475
- # pypdf
476
- # sqlalchemy
477
- # starlette
478
- # torch
479
- # typer
480
- # typing-inspect
481
- # uvicorn
482
- # vllm
483
- typing-inspect==0.9.0
484
- # via
485
- # dataclasses-json
486
- # llama-index-core
487
- # llama-index-legacy
488
- tzdata==2024.1
489
- # via pandas
490
- ujson==5.10.0
491
- # via fastapi
492
- urllib3==2.2.1
493
- # via requests
494
- uvicorn==0.30.1
495
- # via
496
- # fastapi
497
- # vllm
498
- uvloop==0.19.0
499
- # via uvicorn
500
- vllm==0.5.0
501
- # via -r requirements.in
502
- vllm-flash-attn==2.5.9
503
- # via vllm
504
- watchfiles==0.22.0
505
- # via uvicorn
506
- websockets==12.0
507
- # via uvicorn
508
- wrapt==1.16.0
509
- # via
510
- # deprecated
511
- # llama-index-core
512
- xformers==0.0.26.post1
513
- # via vllm
514
- xxhash==3.4.1
515
- # via datasets
516
- yarl==1.9.4
517
- # via aiohttp
 
1
+ outlines
2
+ llama_index
3
+ datasets>=2.18.0
4
+ rich
5
+ vllm