YAML Metadata
Warning:
empty or missing yaml metadata in repo card
(https://huggingface.co/docs/hub/model-cards#model-card-metadata)
Quantization made by Richard Erkhov.
Llama3.2-1B-summary-length-exp7 - AWQ
- Model creator: https://huggingface.co/amang1802/
- Original model: https://huggingface.co/amang1802/Llama3.2-1B-summary-length-exp7/
Original model description:
library_name: transformers tags: []
Model Card for Model ID
- Summary Length PPO experiment #7
- No KL divergence in loss
Model Details
- Dataset size: 16384
- Epochs: 1
- Batch Size: 16 * 4 (w/ 4 GPUs)
Optimizer args: Torch AdamW default, except
- LR = 0.00001
- Downloads last month
- 4