--- datasets: - Oniichat/bluemoon_roleplay_chat_data_300k_messages inference: false language: - en license: llama2 model_creator: PygmalionAI model_link: https://huggingface.co/PygmalionAI/mythalion-13b model_name: mythalion-13b model_type: llama pipeline_tag: text-generation quantized_by: Eigeen tags: - text generation - instruct thumbnail: null --- # Mythalion 13B - ExLlamaV2 Original model: [mythalion-13b](https://huggingface.co/PygmalionAI/mythalion-13b) # Description This is my trial of quantization. I use only RP dataset for calibration, it may cause the model to not perform as well in other situations. But people who use Mythalion basically use it for RP, I guess? Anyway, it works on RP. I haven't tested it's performance in other situations. ExLlamaV2 is great. 2.30 bpw is designed for 8GB VRAM. It is more extreme and can only up to 2048 context. If your VRAM is occupied by other program or system, lower the allowed context. I wouldn't use it though because its performance so poor compared to 4 and 6bpw. It's just can work. I shared it and maybe someone needed it.