--- language: - en license: cc-by-nc-4.0 --- NyakuraV2 - A Multi-Turn / Instruct Mix Fine-tuned Model. Compute is thanks to my wallet, and runpod community pod 8x a40s for roughly 2 hours? Yeah that's it. Trained in ShareGPT dataset format due to multi-turn capabilities. use Vicuna 1.1 prompt format. Alpaca may work fine too, that format is like universal, may give sub-par results though.. I had much better results with the fp16 instead of GGUF, GGUF quants seem fucked? I don't know. May be on my side. Had so much nonsense input, had to wrangle settings until I had it coherent, it was working ***Really*** good. Fuck Yi models are a pain to work with. You need better care with samplers when handling Yi Models. I have tested that Nyakura works well from 0 to 8k+ context, so I know it is not the model's fault. Meow. ``` (Optional) System: User: Assistant: ``` *Example Prompt:* ``` User: Are you stand proud you're strong because you're nah I'd win, or are you nah I'd win because you're stand proud you're strong? Assistant: You really should go to the doctor, I think you need help. ``` ![Proud](https://i.gyazo.com/f8f3804ca59f74372c03c8ffe3ad6f43.png)