--- license: apache-2.0 library_name: transformers tags: - 4-bit - AWQ - text-generation - autotrain_compatible - endpoints_compatible pipeline_tag: text-generation inference: false quantized_by: Suparious --- # amazingvince/openhermes-7b-dpo AWQ - Model creator: [amazingvince](https://huggingface.co/amazingvince) - Original model: [openhermes-7b-dpo](https://huggingface.co/amazingvince/openhermes-7b-dpo) ## Model Summary OpenHermes 2.5 Mistral 7B is a state of the art Mistral Fine-tune, a continuation of OpenHermes 2 model, which trained on additional code datasets. Potentially the most interesting finding from training on a good ratio (est. of around 7-14% of the total dataset) of code instruction was that it has boosted several non-code benchmarks, including TruthfulQA, AGIEval, and GPT4All suite. It did however reduce BigBench benchmark score, but the net gain overall is significant. Here, we are finetuning openheremes using DPO with various data meant to improve its abilities.