Phi4 Turn R1Distill LoRA Adapters

Overview

Hey! These LoRA adapters are trained using different reasoning datasets that utilize Thought and Solution for reasoning responses.
I hope these help jumpstart your project! These adapters have been trained on an A800 GPU and should provide a solid base for fine-tuning or merging.

Everything on my page is left public for Open Source use.

Available LoRA Adapters

Here are the links to the available adapters as of January 30, 2025:

Usage

These adapters can be loaded and used with peft and transformers. Here’s a quick example:

from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel

base_model = "microsoft/Phi-4"
lora_adapter = "Quazim0t0/Phi4.Turn.R1Distill-Lora1"

tokenizer = AutoTokenizer.from_pretrained(base_model)
model = AutoModelForCausalLM.from_pretrained(base_model)
model = PeftModel.from_pretrained(model, lora_adapter)

model.eval()
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no pipeline_tag.

Model tree for Quazim0t0/Phi4.Turn.R1Distill-Lora8

Base model

microsoft/phi-4
Finetuned
(142)
this model

Collection including Quazim0t0/Phi4.Turn.R1Distill-Lora8