rwightman HF staff commited on
Commit
d5be347
1 Parent(s): 3e473ec

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +26 -4
README.md CHANGED
@@ -83,7 +83,7 @@ This model was trained with one of (see table in intro):
83
 
84
  All models were trained with a global batch size of 81920 for 64 checkpoint intervals of 203.7M samples for a total of ~13B samples seen over training.
85
 
86
- For 256x256 models, a slurm script w/ srun below was used on 20 8-GPU nodes (Stability), switching to 40 4-GPU nodes for time on JUWELS.
87
 
88
  ```
89
  /opt/slurm/sbin/srun --cpu_bind=v --accel-bind=gn python -m training.main \
@@ -129,7 +129,7 @@ The models achieve between 70.8 and 71.7 zero-shot top-1 accuracy on ImageNet-1k
129
 
130
  An initial round of benchmarks have been performed on a wider range of datasets, to be viewable at https://github.com/LAION-AI/CLIP_benchmark/blob/main/benchmark/results.ipynb
131
 
132
- As part of exploring increased augmentation + regularization, more analysis is required but early tests indicate the `augreg` models evaluate well over a wider range of resolutions than the non augreg models. Especially the 320x320 LAION-A model, where the augreg disappointed at 320x320 w/ 71.3, but passes the non augreg 71.7 w/ a 72.2 when evaluated at 384x384 (non augreg drops to 71.0 at 384x384).
133
 
134
  # Acknowledgements
135
 
@@ -139,8 +139,30 @@ Acknowledging [stability.ai](https://stability.ai/) and the Gauss Centre for Sup
139
 
140
  **BibTeX:**
141
 
142
- In addition to forthcoming LAION-5B (https://laion.ai/blog/laion-5b/) paper, please cite:
143
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
144
 
145
  OpenCLIP software
146
  ```bibtex
 
83
 
84
  All models were trained with a global batch size of 81920 for 64 checkpoint intervals of 203.7M samples for a total of ~13B samples seen over training.
85
 
86
+ For 256x256 models, a slurm script w/ srun below was used on 20 8-GPU (A100 40GB) nodes (Stability), switching to 40 4-GPU nodes for time on JUWELS.
87
 
88
  ```
89
  /opt/slurm/sbin/srun --cpu_bind=v --accel-bind=gn python -m training.main \
 
129
 
130
  An initial round of benchmarks have been performed on a wider range of datasets, to be viewable at https://github.com/LAION-AI/CLIP_benchmark/blob/main/benchmark/results.ipynb
131
 
132
+ As part of exploring increased augmentation + regularization, early evalations suggest that `augreg` trained models evaluate well over a wider range of resolutions. This is especially true for the 320x320 LAION-A model, where the augreg run was lower than the non-augreg when evaluated at the train resolution of 320x320 (71.3 vs 71.7), but improves to 72.2 when evaluated at 384x384 (the non-augreg drops to 71.0 at 384x384).
133
 
134
  # Acknowledgements
135
 
 
139
 
140
  **BibTeX:**
141
 
142
+ ```bibtex
143
+ @inproceedings{schuhmann2022laionb,
144
+ title={{LAION}-5B: An open large-scale dataset for training next generation image-text models},
145
+ author={Christoph Schuhmann and
146
+ Romain Beaumont and
147
+ Richard Vencu and
148
+ Cade W Gordon and
149
+ Ross Wightman and
150
+ Mehdi Cherti and
151
+ Theo Coombes and
152
+ Aarush Katta and
153
+ Clayton Mullis and
154
+ Mitchell Wortsman and
155
+ Patrick Schramowski and
156
+ Srivatsa R Kundurthy and
157
+ Katherine Crowson and
158
+ Ludwig Schmidt and
159
+ Robert Kaczmarczyk and
160
+ Jenia Jitsev},
161
+ booktitle={Thirty-sixth Conference on Neural Information Processing Systems Datasets and Benchmarks Track},
162
+ year={2022},
163
+ url={https://openreview.net/forum?id=M3Y74vmsMcY}
164
+ }
165
+ ```
166
 
167
  OpenCLIP software
168
  ```bibtex