Schaapje logo

Schaapje-2B-Chat-V1.0-GGUF

Introduction

This is a collection of GGUF files created from Schaapje-2B-Chat-V1.0

It contains the files in the following quantization formats:

Q5_0, Q5_K_M, Q6_K, Q8_0

Requirements

Before you can use the GGUF files you need to clone llama.cpp repository and install it following the official guide.

Recommendation

Experimenting with the llama.cpp parameters can have a big impact on the quality of the generated text. It is therefore recommended to do your own experimentation with different settings. In my own experiments it looks like quantization 'Q5_0' or better gives good quality.

Downloads last month
34
GGUF
Model size
2.53B params
Architecture
granite

5-bit

6-bit

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Collection including robinsmits/Schaapje-2B-Chat-V1.0-GGUF