amang1802's picture
Update README.md
b16491b verified
metadata
library_name: transformers
tags: []

Model Card for Model ID

  • Summary Length PPO experiment #7
  • No KL divergence in loss

Model Details

  • Dataset size: 16384
  • Epochs: 1
  • Batch Size: 16 * 4 (w/ 4 GPUs)

Optimizer args: Torch AdamW default, except

  • LR = 0.00001