Fine-Tuned Model: Qwen/Qwen2.5-Coder-14B-Instruct on GAIR/LIMO

Overview

This model is a fine-tuned version of the base model Qwen/Qwen2.5-Coder-14B-Instruct. It was trained on a subset of problems from the GAIR/LIMO dataset, specifically focusing on 611 problems over 11 training epochs.

After testing more I found that the model does not always include reasoning, I will update with more epochs.

Warning! The model often goes into an endless chain of reasoning.

Training Details

  • Base Model: Qwen/Qwen2.5-Coder-14B-Instruct
  • Dataset: GAIR/LIMO (subset of 611 problems)
  • Epochs: 11
  • Training Limitations: The training was constrained by the computational resources available on my machine, which means I haven't yet conducted a thorough evaluation of the model's performance improvements.

Key Observations

During testing, the fine-tuned model demonstrated significant improvements in reasoning ability compared to the base model. It began to provide more coherent and accurate responses, avoiding the mistakes observed in the base model during my initial tests.

Next Steps

While preliminary results are promising, further evaluation is needed to assess the overall improvement in model quality. I encourage the community to test the model and share their findings. Your feedback will be invaluable in understanding the extent of the improvements.

Acknowledgments

Special thanks to the creators of the Qwen/Qwen2.5-Coder-14B-Instruct and GAIR/LIMO datasets for providing the foundational resources.

Downloads last month
57
Safetensors
Model size
14.8B params
Tensor type
FP16
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Model tree for evilyesh/Qwen2.5-Coder-14B-Instruct-Thinking

Base model

Qwen/Qwen2.5-14B
Finetuned
(13)
this model
Quantizations
2 models

Dataset used to train evilyesh/Qwen2.5-Coder-14B-Instruct-Thinking