Edit model card

xriminact/TarsChattyBasev0.1 GGUF models converted using llama.cpp

Name Quant Method Bits Max RAM Required Usecase
TarsChattyBasev0.1-Q4_K_M.gguf Q4_K_M 4 ~6.5 GB balanced quality
TarsChattyBasev0.1-Q5_K_M.gguf Q5_K_M 5 ~7.5 GB large, very low quality loss
TarsChattyBasev0.1-Q8_0.gguf Q8_0 8 ~10.5 GB very large, extremely low quality loss
Downloads last month
38
GGUF
Model size
7.24B params
Architecture
llama

4-bit

5-bit

8-bit

16-bit

Inference API
Unable to determine this model’s pipeline type. Check the docs .