Adding carbon footprint information
Browse filesCalculated the carbon footprint of different stages of IDEFICS training + experimentation based on information provided by
@VictorSanh
and
@stas
!
README.md
CHANGED
@@ -317,6 +317,32 @@ The IDEFICS models were trained on an AWS SageMaker cluster with 8x80GB A100 GPU
|
|
317 |
The training software is built on top of HuggingFace Transformers + Accelerate, and [DeepSpeed ZeRO-3](https://github.com/microsoft/DeepSpeed) for training, and [WebDataset](https://github.com/webdataset/webdataset) for data loading.
|
318 |
|
319 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
320 |
# Bias, Risks, and Limitations
|
321 |
|
322 |
Significant research has explored bias and fairness issues with language models (see, e.g., [Sheng et al. (2021)](https://aclanthology.org/2021.acl-long.330.pdf) and [Bender et al. (2021)](https://dl.acm.org/doi/pdf/10.1145/3442188.3445922)).
|
|
|
317 |
The training software is built on top of HuggingFace Transformers + Accelerate, and [DeepSpeed ZeRO-3](https://github.com/microsoft/DeepSpeed) for training, and [WebDataset](https://github.com/webdataset/webdataset) for data loading.
|
318 |
|
319 |
|
320 |
+
## Environmental Impact
|
321 |
+
|
322 |
+
We distinguish the 3 phases of the creation of IDEFICS and report our carbon emissions separately for each one of them:
|
323 |
+
|
324 |
+
*Preliminary experimentation*
|
325 |
+
- **Hardware Type:** Intel Cascade Lake CPUs, NVIDIA V100 and A100 GPUs
|
326 |
+
- **Hours used:** 460,000 CPU hours, 385,000 V100 GPU hours, and 300,000 A100 GPU hours
|
327 |
+
- **Cloud Provider:** N/A (Jean Zay cluster)
|
328 |
+
- **Compute Region:** France (57g CO2eq/kWh)
|
329 |
+
- **Carbon Emitted:** 16,714 kgs of CO2eq
|
330 |
+
|
331 |
+
*IDEFICS-80B pretraining*
|
332 |
+
- **Hardware Type:** 512 NVIDIA A100 GPUs
|
333 |
+
- **Hours used:** 672 hours (28 days)
|
334 |
+
- **Cloud Provider:** AWS
|
335 |
+
- **Compute Region:** US-West 2 (288g CO2eq/kWh)
|
336 |
+
- **Carbon Emitted:** 39,498 kgs of CO2eq
|
337 |
+
|
338 |
+
*IDEFICS-80b-instruct finetuning*
|
339 |
+
- **Hardware Type:** 384 NVIDIA A100 GPUs
|
340 |
+
- **Hours used:** 72 hours (3 days)
|
341 |
+
- **Cloud Provider:** AWS
|
342 |
+
- **Compute Region:** US-West 2 (288g CO2eq/kWh)
|
343 |
+
- **Carbon Emitted:** 3,174 kgs of CO2eq
|
344 |
+
|
345 |
+
|
346 |
# Bias, Risks, and Limitations
|
347 |
|
348 |
Significant research has explored bias and fairness issues with language models (see, e.g., [Sheng et al. (2021)](https://aclanthology.org/2021.acl-long.330.pdf) and [Bender et al. (2021)](https://dl.acm.org/doi/pdf/10.1145/3442188.3445922)).
|