Update README.md
Browse files
README.md
CHANGED
@@ -90,7 +90,7 @@ You can fine-tune the model on tile-level downstream tasks.
|
|
90 |
- **Training length:** 100,000 iterations with a batch size of 4,096
|
91 |
- **Model architecture:** ViT-Large (0.3B params): Patch size 16, embedding dimension 1024, 16 heads, MLP FFN
|
92 |
- **Hardware used:** 32x4 Nvidia V100 32GB
|
93 |
-
- **Hours trained
|
94 |
- **Platform**: French supercluster Jean-Zay
|
95 |
|
96 |
## Software Dependencies
|
|
|
90 |
- **Training length:** 100,000 iterations with a batch size of 4,096
|
91 |
- **Model architecture:** ViT-Large (0.3B params): Patch size 16, embedding dimension 1024, 16 heads, MLP FFN
|
92 |
- **Hardware used:** 32x4 Nvidia V100 32GB
|
93 |
+
- **Hours trained:** Approx 4,300 GPU hours (33 hours total)
|
94 |
- **Platform**: French supercluster Jean-Zay
|
95 |
|
96 |
## Software Dependencies
|