Update README.md
Browse files
README.md
CHANGED
@@ -71,11 +71,11 @@ Training Data: 200B tokens from [SlimPajama](https://www.cerebras.net/blog/slimp
|
|
71 |
## 📃 Citation
|
72 |
|
73 |
```bibtex
|
74 |
-
@
|
75 |
title={LLaMA-MoE: Building Mixture-of-Experts from LLaMA with Continual Pre-training},
|
76 |
-
author={
|
77 |
-
|
78 |
-
|
79 |
-
url={https://
|
80 |
}
|
81 |
```
|
|
|
71 |
## 📃 Citation
|
72 |
|
73 |
```bibtex
|
74 |
+
@article{llama-moe,
|
75 |
title={LLaMA-MoE: Building Mixture-of-Experts from LLaMA with Continual Pre-training},
|
76 |
+
author={Tong Zhu and Xiaoye Qu and Daize Dong and Jiacheng Ruan and Jingqi Tong and Conghui He and Yu Cheng},
|
77 |
+
journal={arXiv preprint arXiv:2406.16554},
|
78 |
+
year={2024},
|
79 |
+
url={https://arxiv.org/abs/2406.16554},
|
80 |
}
|
81 |
```
|