--- license: apache-2.0 tags: - merge - mergekit - lazymergekit --- # Llama-3-SPPO-turbcat-RP-v0.1-alpha **These are GGUF quants of the Llama-3-SPPO-turbcat-RP-v0.1-alpha. For transformers version check another repo.** The following GGUF quants are currently available: - Q5_K_M (no imatrix) - Q8_0 (no imatrix) Llama-3-SPPO-turbcat-RP-v0.1-alpha is a merge of the following models using [mergekit](https://github.com/cg123/mergekit): ## 🧩 Configuration ```yaml models: - model: Casual-Autopsy/L3-Umbral-Mind-RP-v3.0-8B - model: UCLA-AGI/Llama-3-Instruct-8B-SPPO-Iter3+grimjim/Llama-3-Instruct-abliteration-LoRA-8B - model: turboderp/llama3-turbcat-instruct-8b merge_method: model_stock base_model: failspy/Meta-Llama-3-8B-Instruct-abliterated-v3 dtype: bfloat16 ```