--- license: mit language: - nl tags: - gguf --- This repository contains quantized versions of [BramVanroy/fietje-2b](https://huggingface.co/BramVanroy/fietje-2b): - `-f16` (5.6GB): best quality, but largest and slowest (recommended if you have the capacity, otherwise q8_0) - `-q8_0` (3.0GB): minimal quality loss, smaller - `-q5_k_m` (2.0GB): users have reported considerable quality loss in the chat `q5_k_m` version so you may want to avoid it Also available on ollama: ```sh # defaults to f16 ollama run bramvanroy/fietje-2b ollama run bramvanroy/fietje-2b:f16 ollama run bramvanroy/fietje-2b:q8_0 ollama run bramvanroy/fietje-2b:q5_k_m ```