Eigeen commited on
Commit
73c9917
1 Parent(s): 575bd88

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +30 -0
README.md ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ datasets:
3
+ - Oniichat/bluemoon_roleplay_chat_data_300k_messages
4
+ inference: false
5
+ language:
6
+ - en
7
+ license: llama2
8
+ model_creator: nRuaif
9
+ model_link: https://huggingface.co/nRuaif/Mythalion-Kimiko-v2
10
+ model_name: Mythalion-Kimiko-v2
11
+ model_type: llama
12
+ pipeline_tag: text-generation
13
+ quantized_by: Eigeen
14
+ tags:
15
+ - text generation
16
+ - instruct
17
+ thumbnail: null
18
+ ---
19
+
20
+ # Mythalion 13B Kimiko-v2 - ExLlamaV2
21
+
22
+ Original model: [Mythalion-Kimiko-v2](https://huggingface.co/nRuaif/Mythalion-Kimiko-v2)
23
+
24
+ # Description
25
+
26
+ This is my trial of quantization. I use only RP dataset for calibration, it may cause the model to not perform as well in other situations. But people who use Mythalion basically use it for RP, I guess?
27
+
28
+ Anyway, it works well on RP. I haven't tested it's performance in other situations. ExLlamaV2 is great.
29
+
30
+ 6.05 bpw is designed for 16GB VRAM. If you have 24GB VRAM, you can expand the context to at least 8192. I did not calculate the exact values.