--- base_model: unsloth/qwen2.5-14b-bnb-4bit tags: - text-generation-inference - transformers - unsloth - qwen2 - trl license: apache-2.0 language: - en --- # Uploaded model - **Developed by:** qingy2019 - **License:** apache-2.0 - **Finetuned from model :** unsloth/qwen2.5-14b-bnb-4bit Huge thanks to Unsloth and the Huggingface TRL library. This model is Qwen 2.5 14B fine tuned for a full epoch on the high quality [garage-bAInd/Open-Platypus](https://huggingface.co/datasets/garage-bAInd/Open-Platypus) dataset for STEM reasoning. | Training Detail | Value | |-------------------------|----------| | Epochs | 1 | | Steps | 2077 | | Loss | 0.4218 | | Batch size | 4 | | Gradient Acc. Steps | 3 | | Learning Rate | 2e-4 | | LR Scheduler | cosine | | Rank | 32 | | Rank-Stabilized LoRA | Yes | | Warm up steps | 5 | | Weight Decay | 0.01 | | Seed | 3407 | [](https://github.com/unslothai/unsloth)