YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

Quantization made by Richard Erkhov.

Github

Discord

Request more models

Replete-LLM-V2.5-Qwen-3b - GGUF

Original model description:

library_name: transformers base_model: - Qwen/Qwen2.5-3B-Instruct license: other license_name: qwen-research license_link: https://huggingface.co/Qwen/Qwen2.5-3B-Instruct/blob/main/LICENSE

Replete-LLM-V2.5-Qwen-3b

image/png

Replete-LLM-V2.5-Qwen-3b is a continues finetuned version of Qwen2.5-3B. I noticed recently that the Qwen team did not learn from my methods of continuous finetuning, the great benefits, and no downsides of it. So I took it upon myself to merge the instruct model with the base model myself using the Ties merge method

This version of the model shows higher performance than the original instruct and base models.

Quants:

GGUF: https://huggingface.co/bartowski/Replete-LLM-V2.5-Qwen-3b-GGUF

Benchmarks: (Coming soon)

Downloads last month
47
GGUF
Model size
3.4B params
Architecture
qwen2

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.