--- license: cc-by-2.0 --- Finetune of miqu-70b-sf dequant of miqudev's leak of Mistral-70B (allegedly an early mistral medium). My diffs are available under CC-0 (That is the Senku-70B repo, full includes the merge), this is a merge with the leaked model, you can use the other repository to save bandwidth. EQ-Bench: 84.89 GSM8k: 77.18 Hellaswag: 87.67 Edit: Upon further testing a score of 85.09 was achieved using ChatLM instead of Mistral's prompt. I recommend using the ChatML format instead, I will run more benchmarks. This also fixes the bug with Miqu dequant failing to provide a stop. <|im_start|>system Provide some context and/or instructions to the model. <|im_end|> <|im_start|>user The user’s message goes here <|im_end|> <|im_start|>assistant <|im_end|> Credit to https://twitter.com/hu_yifei for providing GSM & Hellaswag. It is the first open weight model to dethrone GPT-4 on EQ bench,