LH-Tech-AI commited on
Commit
a131673
·
verified ·
1 Parent(s): e8463a9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -34,7 +34,7 @@ We used the first 100 million tokens of the 10BT Sample of Fineweb-Edu to train
34
  - Batch Size: 32
35
  - Gradient Accumulation Steps: 4
36
  - Compile model: False
37
- - Device Type: float16 - CUDA on Kaggle T4 16GB GPU
38
 
39
  ## Training code
40
  As in all of our models, you can find the full training code in this repo in the files `train.py`, `model.py`, `configurator.py` and `prepare.py`.
 
34
  - Batch Size: 32
35
  - Gradient Accumulation Steps: 4
36
  - Compile model: False
37
+ - Device Type: float16 - CUDA on Kaggle T4 16GB GPU (train time: ~71min)
38
 
39
  ## Training code
40
  As in all of our models, you can find the full training code in this repo in the files `train.py`, `model.py`, `configurator.py` and `prepare.py`.