Sao10K's picture
Update README.md
ac6ed86
---
language:
- en
license: cc-by-nc-4.0
---
NyakuraV2 - A Multi-Turn / Instruct Mix Fine-tuned Model.
Compute is thanks to my wallet, and runpod community pod 8x a40s for roughly 2 hours? Yeah that's it.
Trained in ShareGPT dataset format due to multi-turn capabilities.
use Vicuna 1.1 prompt format. Alpaca may work fine too, that format is like universal, may give sub-par results though..
I had much better results with the fp16 instead of GGUF, GGUF quants seem fucked? I don't know. May be on my side. Had so much nonsense input, had to wrangle settings until I had it coherent, it was working ***Really*** good. Fuck Yi models are a pain to work with.
You need better care with samplers when handling Yi Models. I have tested that Nyakura works well from 0 to 8k+ context, so I know it is not the model's fault.
Meow.
```
(Optional) System: <Prompt>
User: <Input>
Assistant:
```
*Example Prompt:*
```
User: Are you stand proud you're strong because you're nah I'd win, or are you nah I'd win because you're stand proud you're strong?
Assistant: You really should go to the doctor, I think you need help.
```
![Proud](https://i.gyazo.com/f8f3804ca59f74372c03c8ffe3ad6f43.png)