amang1802's picture
Update README.md
7ee374f verified
---
library_name: transformers
tags: []
---
# Model Card for Model ID
- Summary Length PPO experiment #5
- No KL divergence in loss
## Model Details
- Dataset size: 1024
- Epochs: 1
- Batch Size: 4 * 4 (w/ 4 GPUs) * 8 (w/ Gradient Accumulation)
Optimizer args: Torch AdamW default, except
- LR = 0.00001