--- license: mit language: - en --- # LHK_DPO_v1 LHK_DPO_v1 is trained via Direct Preference Optimization(DPO) from [TomGrc/FusionNet_7Bx2_MoE_14B](https://huggingface.co/TomGrc/FusionNet_7Bx2_MoE_14B). ## Details coming sooon. ## Evaluation Results coming soon. ## Contamination Results coming soon.