Qingyun Li
Update README.md
ddfc382 verified
---
base_model: unsloth/qwen2.5-14b-bnb-4bit
tags:
- text-generation-inference
- transformers
- unsloth
- qwen2
- trl
license: apache-2.0
language:
- en
---
# Uploaded model
- **Developed by:** qingy2019
- **License:** apache-2.0
- **Finetuned from model :** unsloth/qwen2.5-14b-bnb-4bit
Huge thanks to Unsloth and the Huggingface TRL library.
This model is Qwen 2.5 14B fine tuned for a full epoch on the high quality [garage-bAInd/Open-Platypus](https://huggingface.co/datasets/garage-bAInd/Open-Platypus) dataset for STEM reasoning.
| Training Detail | Value |
|-------------------------|----------|
| Epochs | 1 |
| Steps | 2077 |
| Loss | 0.4218 |
| Batch size | 4 |
| Gradient Acc. Steps | 3 |
| Learning Rate | 2e-4 |
| LR Scheduler | cosine |
| Rank | 32 |
| Rank-Stabilized LoRA | Yes |
| Warm up steps | 5 |
| Weight Decay | 0.01 |
| Seed | 3407 |
<img src="https://raw.githubusercontent.com/qingy1337/Storage/refs/heads/main/loss.svg" height="500"/>
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)