This model is finetuned on the distillation data from GPT-4. The base model is mistralai/Mistral-7B-Instruct-v0.2
The model has a query format as in mistral.
<s>[INST] {query} [/INST] {response}</s> [INST] {second_query} [/INST]"
-