berketonyukuk/TRdpoQwen
Viewer
•
Updated
•
10.8k
•
189
thx
I applied DPO fine-tuning to the Qwen VL-2.5 7B model using PEFT LoRA. When I loaded the model with the adapter, inference was 6 times slower compared to the base model. What could be the reason for this?