File size: 1,198 Bytes
8ac6eef
 
 
 
 
 
 
 
 
ac6ed86
8ac6eef
 
 
 
 
 
 
5f6db29
 
8ac6eef
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9349e70
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
---
language:
- en
license: cc-by-nc-4.0
---


NyakuraV2 - A Multi-Turn / Instruct Mix Fine-tuned Model.

Compute is thanks to my wallet, and runpod community pod 8x a40s for roughly 2 hours? Yeah that's it. 

Trained in ShareGPT dataset format due to multi-turn capabilities. 

use Vicuna 1.1 prompt format. Alpaca may work fine too, that format is like universal, may give sub-par results though..

I had much better results with the fp16 instead of GGUF, GGUF quants seem fucked? I don't know. May be on my side. Had so much nonsense input, had to wrangle settings until I had it coherent, it was working ***Really*** good. Fuck Yi models are a pain to work with.

You need better care with samplers when handling Yi Models. I have tested that Nyakura works well from 0 to 8k+ context, so I know it is not the model's fault.

Meow.

```
(Optional) System: <Prompt>

User: <Input>

Assistant:
```

*Example Prompt:*

```
User: Are you stand proud you're strong because you're nah I'd win, or are you nah I'd win because you're stand proud you're strong?

Assistant: You really should go to the doctor, I think you need help.
```

![Proud](https://i.gyazo.com/f8f3804ca59f74372c03c8ffe3ad6f43.png)