--- license: mit language: - en pipeline_tag: text-generation tags: - moe --- # FusionNet_7Bx2_MoE_v0.1 Fine-tuned model on English language using MoE method. The improved version from FusionNet_7Bx2_MoE_14B. ## Model description The FusionNet_7Bx2_MoE_v0.1 is a model to experiment with the MoE method, which could significantly increase the performance of the original model. The FusionNet has 12.9B parameters, and this model is fine-tuned. Enjoy!