--- library_name: transformers tags: [] --- # Model Card for Model ID - Summary Length PPO experiment #7 - No KL divergence in loss ## Model Details - Dataset size: 16384 - Epochs: 1 - Batch Size: 16 * 4 (w/ 4 GPUs) Optimizer args: Torch AdamW default, except - LR = 0.00001