sasha HF staff commited on
Commit
a2b0422
1 Parent(s): 4e645cd

Adding carbon footprint information

Browse files

Calculated the carbon footprint of different stages of IDEFICS training + experimentation based on information provided by

@VictorSanh

and

@stas

!

Files changed (1) hide show
  1. README.md +26 -0
README.md CHANGED
@@ -317,6 +317,32 @@ The IDEFICS models were trained on an AWS SageMaker cluster with 8x80GB A100 GPU
317
  The training software is built on top of HuggingFace Transformers + Accelerate, and [DeepSpeed ZeRO-3](https://github.com/microsoft/DeepSpeed) for training, and [WebDataset](https://github.com/webdataset/webdataset) for data loading.
318
 
319
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
320
  # Bias, Risks, and Limitations
321
 
322
  Significant research has explored bias and fairness issues with language models (see, e.g., [Sheng et al. (2021)](https://aclanthology.org/2021.acl-long.330.pdf) and [Bender et al. (2021)](https://dl.acm.org/doi/pdf/10.1145/3442188.3445922)).
 
317
  The training software is built on top of HuggingFace Transformers + Accelerate, and [DeepSpeed ZeRO-3](https://github.com/microsoft/DeepSpeed) for training, and [WebDataset](https://github.com/webdataset/webdataset) for data loading.
318
 
319
 
320
+ ## Environmental Impact
321
+
322
+ We distinguish the 3 phases of the creation of IDEFICS and report our carbon emissions separately for each one of them:
323
+
324
+ *Preliminary experimentation*
325
+ - **Hardware Type:** Intel Cascade Lake CPUs, NVIDIA V100 and A100 GPUs
326
+ - **Hours used:** 460,000 CPU hours, 385,000 V100 GPU hours, and 300,000 A100 GPU hours
327
+ - **Cloud Provider:** N/A (Jean Zay cluster)
328
+ - **Compute Region:** France (57g CO2eq/kWh)
329
+ - **Carbon Emitted:** 16,714 kgs of CO2eq
330
+
331
+ *IDEFICS-80B pretraining*
332
+ - **Hardware Type:** 512 NVIDIA A100 GPUs
333
+ - **Hours used:** 672 hours (28 days)
334
+ - **Cloud Provider:** AWS
335
+ - **Compute Region:** US-West 2 (288g CO2eq/kWh)
336
+ - **Carbon Emitted:** 39,498 kgs of CO2eq
337
+
338
+ *IDEFICS-80b-instruct finetuning*
339
+ - **Hardware Type:** 384 NVIDIA A100 GPUs
340
+ - **Hours used:** 72 hours (3 days)
341
+ - **Cloud Provider:** AWS
342
+ - **Compute Region:** US-West 2 (288g CO2eq/kWh)
343
+ - **Carbon Emitted:** 3,174 kgs of CO2eq
344
+
345
+
346
  # Bias, Risks, and Limitations
347
 
348
  Significant research has explored bias and fairness issues with language models (see, e.g., [Sheng et al. (2021)](https://aclanthology.org/2021.acl-long.330.pdf) and [Bender et al. (2021)](https://dl.acm.org/doi/pdf/10.1145/3442188.3445922)).