Edit model card
YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

GGUF for FuseChat-7B-VaRM using Capybara-Binarized for Imatrix at 8k context with Q8 model (Only got 1500 chunks through the dataset, I got tired of waiting :| cries in 6GB vram)

Sillytavern template inside presets folder (unsure if it's correct)

I wouldn't go lower than IQ4_XS, IQ3_XXS and IQ3_XS work but they're a little dumb

IQ1_S is unusable too dumb and has a repetition problem

Imatrix was also used for Q4_K_M and Q5_K_M

Downloads last month
131
GGUF
Model size
7.24B params
Architecture
llama

1-bit

2-bit

3-bit

4-bit

5-bit

Inference API
Unable to determine this model's library. Check the docs .