|
--- |
|
license: apache-2.0 |
|
pipeline_tag: image-text-to-text |
|
--- |
|
|
|
<h2 align="center"> <a href="https://arxiv.org/abs/2405.14297">Dynamic Mixture of Experts: An Auto-Tuning Approach for Efficient Transformer Models</a></h2> |
|
<h5 align="center"> If our project helps you, please give us a star β on <a href="https://github.com/LINs-lab/DynMoE">GitHub</a> and cite our paper!</h2> |
|
<h5 align="center"> |
|
|
|
## π° News |
|
|
|
- **[2024.05.31]** π₯ Our [code](https://github.com/LINs-lab/DynMoE/) is released! |
|
- **[2024.05.25]** π₯ Our **checkpoints** are available now! |
|
- **[2024.05.23]** π₯ Our [paper](https://arxiv.org/abs/2405.14297) is released! |
|
|
|
## π What's Interesting? |
|
|
|
**Dynamic Mixture of Experts (DynMoE)** incorporates (1) a novel gating method that enables each token to automatically determine the number of experts to activate. (2) An adaptive process automatically adjusts the number of experts during training. |
|
|
|
### Top-Any Gating |
|
|
|
<video controls src="https://i.imgur.com/bLgNaoH.mp4" title="Top-Any Gating"></video> |
|
|
|
### Adaptive Training Process |
|
|
|
![](https://cdn.jsdelivr.net/gh/QAQdev/Pics@master/uPic/adaptive.png) |
|
|
|
## π‘ Model Details |
|
|
|
- π€ DynMoE-Qwen is a MoE model with **dynamic top-k gating**, finetuned on [LanguageBind/MoE-LLaVA-Qwen-Stage2](https://huggingface.co/LanguageBind/MoE-LLaVA-Qwen-Stage2). |
|
- π Our DynMoE-Qwen-1.8B has totally 3.1B parameters, but **only 2.2B are activated!** (average top-k = 1.86) |
|
- β With the DynMoE tuning stage, we can complete training on 8 A100 GPUs **within 40 hours.** |
|
|
|
## π Acknowledgement |
|
|
|
We are grateful for the following awesome projects: |
|
|
|
- [tutel](https://github.com/microsoft/tutel) |
|
- [DeepSpeed](https://github.com/microsoft/DeepSpeed) |
|
- [GMoE](https://github.com/Luodian/Generalizable-Mixture-of-Experts) |
|
- [EMoE](https://github.com/qiuzh20/EMoE) |
|
- [MoE-LLaVA](https://github.com/PKU-YuanGroup/MoE-LLaVA) |
|
- [GLUE-X](https://github.com/YangLinyi/GLUE-X) |
|
|
|
## π License |
|
|
|
This project is released under the Apache-2.0 license as found in the [LICENSE](https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/mit.md) file. |
|
|
|
## βοΈ Citation |
|
|
|
```tex |
|
@misc{guo2024dynamic, |
|
title={Dynamic Mixture of Experts: An Auto-Tuning Approach for Efficient Transformer Models}, |
|
author={Yongxin Guo and Zhenglin Cheng and Xiaoying Tang and Tao Lin}, |
|
year={2024}, |
|
eprint={2405.14297}, |
|
archivePrefix={arXiv}, |
|
primaryClass={cs.LG} |
|
} |
|
``` |