Update README.md
Browse files
README.md
CHANGED
@@ -73,10 +73,9 @@ Training Data: 200B tokens from [SlimPajama](https://www.cerebras.net/blog/slimp
|
|
73 |
```bibtex
|
74 |
@article{llama-moe,
|
75 |
title={LLaMA-MoE: Building Mixture-of-Experts from LLaMA with Continual Pre-training},
|
76 |
-
author={
|
77 |
-
journal={arXiv},
|
78 |
-
year={
|
79 |
-
|
80 |
-
url={https://arxiv.org}
|
81 |
}
|
82 |
```
|
|
|
73 |
```bibtex
|
74 |
@article{llama-moe,
|
75 |
title={LLaMA-MoE: Building Mixture-of-Experts from LLaMA with Continual Pre-training},
|
76 |
+
author={Tong Zhu and Xiaoye Qu and Daize Dong and Jiacheng Ruan and Jingqi Tong and Conghui He and Yu Cheng},
|
77 |
+
journal={arXiv preprint arXiv:2406.16554},
|
78 |
+
year={2024},
|
79 |
+
url={https://arxiv.org/abs/2406.16554},
|
|
|
80 |
}
|
81 |
```
|