yongzx commited on
Commit
48a3427
1 Parent(s): 74a8c6d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -2
README.md CHANGED
@@ -28,5 +28,6 @@ We finetuned the `wte` and `wpe` layers of GPT-2 (while freezing the parameters
28
  - max_eval_samples: 5000
29
  ```
30
 
31
- Setup: 8 RTX-3090 GPUs, trained for seven days (total training steps: 110500, effective train batch size: 64, tokens per batch: 1024)
32
- Final checkpoint: checkpoint-111500
 
 
28
  - max_eval_samples: 5000
29
  ```
30
 
31
+ **Setup**: 8 RTX-3090 GPUs, trained for seven days (total training steps: 110500, effective train batch size: 64, tokens per batch: 1024)
32
+
33
+ **Final checkpoint**: checkpoint-111500