Logo

πŸš€ bloomvn-0.5b-ppo-GGUF

Optimized quantized models for efficient inference

πŸ“‹ Overview

A collection of optimized GGUF quantized models derived from BlossomsAI/BloomVN-0.5B-ppo, providing various performance-quality tradeoffs.

πŸ’Ž Model Variants

Variant Description Size Download
base Base model in FP16 format 948.1MB πŸ“₯
q2_k Basic text completion tasks 322.9MB πŸ“₯
q3_k_m Memory-efficient quality operations 339.0MB πŸ“₯

🀝 Contributors

Developed with ❀️ by Anhnv-sharedmodels


Star ⭐️ this repo if you find it valuable!
Downloads last month
31
GGUF
Model size
494M params
Architecture
qwen2

2-bit

3-bit

Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no pipeline_tag.

Model tree for Anhnv-sharedmodels/bloomvn-0.5b-ppo-gguf

Base model

Qwen/Qwen2.5-0.5B
Quantized
(2)
this model

Dataset used to train Anhnv-sharedmodels/bloomvn-0.5b-ppo-gguf