metadata
base_model: unsloth/qwen2.5-14b-bnb-4bit
tags:
- text-generation-inference
- transformers
- unsloth
- qwen2
- trl
license: apache-2.0
language:
- en
Uploaded model
- Developed by: qingy2019
- License: apache-2.0
- Finetuned from model : unsloth/qwen2.5-14b-bnb-4bit
Huge thanks to Unsloth and the Huggingface TRL library.
This model is Qwen 2.5 14B fine tuned for a full epoch on the high quality garage-bAInd/Open-Platypus dataset for STEM reasoning.
Training Detail | Value |
---|---|
Epochs | 1 |
Steps | 2077 |
Loss | 0.4218 |
Batch size | 4 |
Gradient Acc. Steps | 3 |
Learning Rate | 2e-4 |
LR Scheduler | cosine |
Rank | 32 |
Rank-Stabilized LoRA | Yes |
Warm up steps | 5 |
Weight Decay | 0.01 |
Seed | 3407 |