distill-1.7B-4bit-MLX

4-bit quantized version of distill-1.7B — an Expert Language Model for CLI output compression. Zero accuracy loss vs fp16.

Format Size Accuracy
fp16 3.2 GB 95%
4-bit 1.0 GB 95%

Built for distill. Full collection

Downloads last month
87
Safetensors
Model size
0.3B params
Tensor type
BF16
·
U32
·
MLX
Hardware compatibility
Log In to add your hardware

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for samuelfaj/distill-1.7B-4bit-MLX

Finetuned
Qwen/Qwen3-1.7B
Quantized
(1)
this model
Finetunes
2 models

Collection including samuelfaj/distill-1.7B-4bit-MLX