tuantran1632001's picture
Update README.md
b9980b3 verified
|
raw
history blame
1.98 kB
---
license: other
license_name: microsoft-research-license
tags:
- GGUF
- KoboldAI/LLaMA2-13B-Psyfighter2
- microsoft/Orca-2-13b
model_type: llama
model_name: Psyfighter2-Orca2-13B-ties
quantized_by: tuantran1632001
base_model: tuantran1632001/Psyfighter2-Orca2-13B-ties
---
This is the GGUF quantize of the merged model [tuantran1632001/Psyfighter2-Orca2-13B-ties](https://huggingface.co/tuantran1632001/Psyfighter2-Orca2-13B-ties).
| File | Quantize | Size |
|------|----------|------|
| [Psyfighter2-Orca2-13B-ties-fp16.gguf](Psyfighter2-Orca2-13B-ties-fp16.gguf) | fp16 | 25GiB |
| [Psyfighter2-Orca2-13B-ties-Q2_K.gguf](Psyfighter2-Orca2-13B-ties-Q2_K.gguf) | Q2_K | 5.1GiB |
| [Psyfighter2-Orca2-13B-ties-Q3_K_L.gguf](Psyfighter2-Orca2-13B-ties-Q3_K_L.gguf) | Q3_K_L | 6.5GiB |
| [Psyfighter2-Orca2-13B-ties-Q3_K_M.gguf](Psyfighter2-Orca2-13B-ties-Q3_K_M.gguf) | Q3_K_M | 6.0GiB |
| [Psyfighter2-Orca2-13B-ties-Q3_K_S.gguf](Psyfighter2-Orca2-13B-ties-Q3_K_S.gguf) | Q3_K_S | 5.3GiB |
| [Psyfighter2-Orca2-13B-ties-Q4_0.gguf](Psyfighter2-Orca2-13B-ties-Q4_0.gguf) | Q4_0 | 6.9GiB |
| [Psyfighter2-Orca2-13B-ties-Q4_1.gguf](Psyfighter2-Orca2-13B-ties-Q4_1.gguf) | Q4_1 | 6.8GiB |
| [Psyfighter2-Orca2-13B-ties-Q4_K_M.gguf](Psyfighter2-Orca2-13B-ties-Q4_K_M.gguf) | Q4_K_M | 7.7GiB |
| [Psyfighter2-Orca2-13B-ties-Q4_K_S.gguf](Psyfighter2-Orca2-13B-ties-Q4_K_S.gguf) | Q4_K_S | 7.0GiB |
| [Psyfighter2-Orca2-13B-ties-Q5_0.gguf](Psyfighter2-Orca2-13B-ties-Q5_0.gguf) | Q5_0 | 8.4GiB |
| [Psyfighter2-Orca2-13B-ties-Q5_1.gguf](Psyfighter2-Orca2-13B-ties-Q5_1.gguf) | Q5_1 | 9.2GiB |
| [Psyfighter2-Orca2-13B-ties-Q5_K_M.gguf](Psyfighter2-Orca2-13B-ties-Q5_K_M.gguf) | Q5_K_M | 8.6GiB |
| [Psyfighter2-Orca2-13B-ties-Q5_K_S.gguf](Psyfighter2-Orca2-13B-ties-Q5_K_S.gguf) | Q5_K_S | 8.4GiB |
| [Psyfighter2-Orca2-13B-ties-Q6_K.gguf](Psyfighter2-Orca2-13B-ties-Q6_K.gguf) | Q6_K | 10GiB |
| [Psyfighter2-Orca2-13B-ties-Q8_0.gguf](Psyfighter2-Orca2-13B-ties-Q8_0.gguf) | Q8_0 | 13GiB |