--- license: apache-2.0 tags: - moe - merge - mergekit - lazymergekit - cognitivecomputations/dolphin-2_6-phi-2 - lxuechen/phi-2-dpo --- ![](https://i.imgur.com/UOb2fvh.jpg) # phixtral-2x2.8 phixtral-2x2.8 is a Mixure of Experts (MoE) made with the following models using a custom version of mergekit: * [cognitivecomputations/dolphin-2_6-phi-2](https://huggingface.co/cognitivecomputations/dolphin-2_6-phi-2) * [lxuechen/phi-2-dpo](https://huggingface.co/lxuechen/phi-2-dpo) ## 🧩 Configuration ```yaml base_model: cognitivecomputations/dolphin-2_6-phi-2 gate_mode: cheap_embed experts: - source_model: cognitivecomputations/dolphin-2_6-phi-2 positive_prompts: [""] - source_model: lxuechen/phi-2-dpo positive_prompts: [""] ``` ## 💻 Usage This architecture is not compatible with the transformers library. I'm working on hacking something to run it. Contact me if you're interested!