model_backup

This is a merged fine-tuned model based on Qwen/Qwen3-4B. The LoRA adapters have been merged into the base model, creating a standalone fine-tuned model.

Model Description

This language model has been fine-tuned using LLaMA-Factory and then merged with the base model. It specializes in email search and related tasks.

Model Details

  • Base Model: Qwen/Qwen3-4B
  • Model Size: ~4B parameters
  • Architecture: Qwen3
  • Training Method: LoRA fine-tuning + model merging
  • Dataset: tags_and_summary
  • Use Case: Email search and analysis
Downloads last month
6
Safetensors
Model size
4.02B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for trl-algo/model_backup

Base model

Qwen/Qwen3-4B-Base
Finetuned
Qwen/Qwen3-4B
Finetuned
(278)
this model