--- license: apache-2.0 datasets: - davanstrien/haiku_dpo language: - en tags: - dpo - poetry base_model: - teknium/OpenHermes-2.5-Mistral-7B --- # Model Card for HaikuHermes-0.1-7B This is a very early model which uses the [davanstrien/haiku_dpo](https://huggingface.co/datasets/davanstrien/haiku_dpo) dataset to train [teknium/OpenHermes-2.5-Mistral-7B](https://huggingface.co/teknium/OpenHermes-2.5-Mistral-7B) using Direct Preference Optimization. The eventual goal of this model is for it to write "technically correct" haiku.