This model is test version, alignment-tuned model.
We utilize state-of-the-art instruction fine-tuning methods including direct preference optimization (DPO).