Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
datasets:
|
3 |
+
- Oniichat/bluemoon_roleplay_chat_data_300k_messages
|
4 |
+
inference: false
|
5 |
+
language:
|
6 |
+
- en
|
7 |
+
license: llama2
|
8 |
+
model_creator: nRuaif
|
9 |
+
model_link: https://huggingface.co/nRuaif/Mythalion-Kimiko-v2
|
10 |
+
model_name: Mythalion-Kimiko-v2
|
11 |
+
model_type: llama
|
12 |
+
pipeline_tag: text-generation
|
13 |
+
quantized_by: Eigeen
|
14 |
+
tags:
|
15 |
+
- text generation
|
16 |
+
- instruct
|
17 |
+
thumbnail: null
|
18 |
+
---
|
19 |
+
|
20 |
+
# Mythalion 13B Kimiko-v2 - ExLlamaV2
|
21 |
+
|
22 |
+
Original model: [Mythalion-Kimiko-v2](https://huggingface.co/nRuaif/Mythalion-Kimiko-v2)
|
23 |
+
|
24 |
+
# Description
|
25 |
+
|
26 |
+
This is my trial of quantization. I use only RP dataset for calibration, it may cause the model to not perform as well in other situations. But people who use Mythalion basically use it for RP, I guess?
|
27 |
+
|
28 |
+
Anyway, it works well on RP. I haven't tested it's performance in other situations. ExLlamaV2 is great.
|
29 |
+
|
30 |
+
6.05 bpw is designed for 16GB VRAM. If you have 24GB VRAM, you can expand the context to at least 8192. I did not calculate the exact values.
|