Full-text search
+ 1,000 results

gguf-org / flux-dev-gguf
README.md
model
11 matches
tags:
gguf, gguf-node, text-to-image, en, base_model:Comfy-Org/flux1-dev, base_model:quantized:Comfy-Org/flux1-dev, license:other, region:us
33
34
35
36
37
# **gguf quantized version of flux-dev (incl. gguf encoder and gguf vae)**
- drag **flux-dev** to > `./ComfyUI/models/diffusion_models`
- drag **clip-l, t5xxl** to > `./ComfyUI/models/text_encoders`
- drag **vae** to > `./ComfyUI/models/vae`
- drag demo picture (below) to > your browser for workflow

AetherArchitectural / GGUF-Quantization-Script
README.md
model
16 matches
tags:
gguf, quantized, text-generation-inference, text-generation, license:cc-by-nc-4.0, region:us
13
14
15
16
17
ural/GGUF-Quantization-Script**](https://huggingface.co/AetherArchitectural/GGUF-Quantization-Script).
>
> **Credits:** <br>
> Made with love by [**@Aetherarchio**](https://huggingface.co/aetherarchio)/[**@FantasiaFoundry**](https://huggingface.co/FantasiaFoundry)/[**@Lewdiculous**](https://huggingface.co/Lewdiculous) with the generous contributions by [**@SolidSnacke**](https://huggingface.co/SolidSnacke) and [**@Virt-io**](https://huggingface.co/Virt-io). <br>
> If this proves useful for you, feel free to credit and share the repository and authors.

Cebtenzzre / gguf-misc
model
2 matches
npvinHnivqn / GGUF-openchat
README.md
model
3 matches
npvinHnivqn / GGUF-metamath-llemma
README.md
model
3 matches
huodon / gguf-models
model
2 matches

Felladrin / gguf-TinyMistral-248M-SFT-v4
README.md
model
3 matches

Felladrin / gguf-llama-160m
README.md
model
3 matches

Felladrin / gguf-Llama-160M-Chat-v1
README.md
model
3 matches

hesha / gguf-collection
model
2 matches

IndiaBuild / GGUF_Navarna_v0_1_OpenHermes_Hindi
README.md
model
3 matches
gguf / AlphaMonarch-7B-GGUF
model
2 matches
gguf / Mistral-7B-Instruct-v0.2-GGUF
model
2 matches
gguf / BioMistral-7B-GGUF
model
2 matches
gguf / cosmo-1b-GGUF
model
2 matches
gguf / Nous-Hermes-2-Mistral-7B-DPO-GGUF
model
2 matches
gguf / Smaug-34B-v0.1-GGUF
model
2 matches

Felladrin / gguf-Minueza-32M-Base
README.md
model
5 matches
tags:
gguf, base_model:Felladrin/Minueza-32M-Base, base_model:quantized:Felladrin/Minueza-32M-Base, license:apache-2.0, endpoints_compatible, region:us, conversational
6
7
8
9
10
GGUF version of [Felladrin/Minueza-32M-Base](https://huggingface.co/Felladrin/Minueza-32M-Base).
It was not possible to quantize the model, so only the F16 and F32 GGUF files are available.
## Try it with [llama.cpp](https://github.com/ggerganov/llama.cpp)

Felladrin / gguf-Minueza-32M-Chat
README.md
model
4 matches
tags:
gguf, base_model:Felladrin/Minueza-32M-Chat, base_model:quantized:Felladrin/Minueza-32M-Chat, license:apache-2.0, endpoints_compatible, region:us, conversational
6
7
8
9
10
GGUF version of [Felladrin/Minueza-32M-Chat](https://huggingface.co/Felladrin/Minueza-32M-Chat).
It was not possible to quantize the model after converting it to F16/F32 GGUF, so only those versions are available, being F32 the recommended one for having better precision.
## Recommended Inference Parameters