Qingyun Li
Update README.md
ddfc382 verified
metadata
base_model: unsloth/qwen2.5-14b-bnb-4bit
tags:
  - text-generation-inference
  - transformers
  - unsloth
  - qwen2
  - trl
license: apache-2.0
language:
  - en

Uploaded model

  • Developed by: qingy2019
  • License: apache-2.0
  • Finetuned from model : unsloth/qwen2.5-14b-bnb-4bit

Huge thanks to Unsloth and the Huggingface TRL library.

This model is Qwen 2.5 14B fine tuned for a full epoch on the high quality garage-bAInd/Open-Platypus dataset for STEM reasoning.

Training Detail Value
Epochs 1
Steps 2077
Loss 0.4218
Batch size 4
Gradient Acc. Steps 3
Learning Rate 2e-4
LR Scheduler cosine
Rank 32
Rank-Stabilized LoRA Yes
Warm up steps 5
Weight Decay 0.01
Seed 3407