--- license: mit license_link: https://huggingface.co/OpenPipe/Deductive-Reasoning-Qwen-14B/blob/main/LICENSE language: - en pipeline_tag: text-generation base_model: - Qwen/Qwen2.5-1.5B-Instruct tags: - chat library_name: transformers --- # Deductive-Reasoning-Qwen-14B ![image/png](https://cdn-uploads.huggingface.co/production/uploads/674a1d102c0f27a385772cfe/JauBmEQM0FpOdShBMSfst.png) Deductive Reasoning Qwen 14B is a reinforcement fine-tune of [Qwen 2.5 14B Instruct](https://huggingface.co/Qwen/Qwen2.5-14B-Instruct) to solve challenging deduction problems from the [Temporal Clue](https://github.com/bradhilton/temporal-clue) dataset, trained by [OpenPipe](https://openpipe.ai)! Here are some additional resources to check out: - [Blog Post](https://openpipe.ai/blog/using-grpo-to-beat-o1-o3-mini-and-r1-on-temporal-clue) - [Training Recipe](https://github.com/openpipe/deductive-reasoning) - [RL Experiments](https://github.com/openpipe/rl-experiments) - [Deductive Reasoning Qwen 32B](https://huggingface.co/OpenPipe/Deductive-Reasoning-Qwen-32B) If you're interested in training your own models with reinforcement learning or just chatting, feel free to [reach out](https://openpipe.ai/contact) or email Kyle directly at kyle@openpipe.ai!