RichardErkhov's picture
uploaded readme
76369b4 verified

Quantization made by Richard Erkhov.

Github

Discord

Request more models

Llama3.2-1B-summary-length-exp7 - AWQ

Original model description:

library_name: transformers tags: []

Model Card for Model ID

  • Summary Length PPO experiment #7
  • No KL divergence in loss

Model Details

  • Dataset size: 16384
  • Epochs: 1
  • Batch Size: 16 * 4 (w/ 4 GPUs)

Optimizer args: Torch AdamW default, except

  • LR = 0.00001