Transformers
PyTorch
English
bridgetower
gaudi
Inference Endpoints
anahita-b commited on
Commit
7ac04f4
·
1 Parent(s): 4460437

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -101,7 +101,7 @@ TODO
101
 
102
  ### Pretraining
103
 
104
- The model was pre-trained for 100k steps on 8 NVIDIA A100 GPUs with a batch size of 4096.
105
  The optimizer used was AdamW with a learning rate of 1e-5. No data augmentation was used except for center-crop. The image resolution in pre-training is set to 288 x 288.
106
 
107
  ## Evaluation results
 
101
 
102
  ### Pretraining
103
 
104
+ The model was pre-trained for ___ steps on an "Intel AI supercomputing cluster" using 512 Gaudis and 128 Xeons with a batch size of 4096.
105
  The optimizer used was AdamW with a learning rate of 1e-5. No data augmentation was used except for center-crop. The image resolution in pre-training is set to 288 x 288.
106
 
107
  ## Evaluation results