Update README.md
Browse files
README.md
CHANGED
@@ -41,7 +41,7 @@ Benchmark scores, including our models, can be found at https://rinnakk.github.i
|
|
41 |
|
42 |
* **Training**
|
43 |
|
44 |
-
The model was trained on approximately 19,000 hours of following Japanese speech corpus ReazonSpeech.
|
45 |
Note that speech samples longer than 16 seconds were excluded before training.
|
46 |
- [ReazonSpeech](https://huggingface.co/datasets/reazon-research/reazonspeech)
|
47 |
|
@@ -128,49 +128,62 @@ The model uses the same sentencepiece-based tokenizer as [japanese-gpt-neox-3.6b
|
|
128 |
|
129 |
# How to cite
|
130 |
```bibtex
|
131 |
-
@
|
132 |
-
title={
|
133 |
-
author={Hono, Yukiya and Mitsuda, Koh and Zhao, Tianyu and Mitsui, Kentaro and Wakatsuki, Toshiaki and Sawada, Kei},
|
134 |
-
|
135 |
-
year={
|
136 |
}
|
137 |
|
138 |
@misc{rinna-nue-asr,
|
139 |
-
title={rinna/nue-asr},
|
140 |
-
author={Hono, Yukiya and Mitsuda, Koh and Zhao, Tianyu and Mitsui, Kentaro and Wakatsuki, Toshiaki and Sawada, Kei},
|
141 |
-
url={https://huggingface.co/rinna/nue-asr}
|
142 |
}
|
143 |
```
|
144 |
---
|
145 |
|
146 |
# References
|
147 |
```bibtex
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
148 |
@article{hsu2021hubert,
|
149 |
-
title={HuBERT: Self-Supervised Speech Representation Learning by Masked Prediction of Hidden Units},
|
150 |
-
author={Hsu, Wei-Ning and Bolte, Benjamin and Tsai, Yao-Hung Hubert and Lakhotia, Kushal and Salakhutdinov, Ruslan and Mohamed, Abdelrahman},
|
151 |
-
journal={IEEE/ACM Transactions on Audio, Speech, and Language Processing},
|
152 |
-
|
153 |
-
|
154 |
-
|
155 |
-
|
|
|
156 |
}
|
157 |
|
158 |
@software{andoniangpt2021gpt,
|
159 |
-
title={{GPT-
|
160 |
-
author={Andonian, Alex and Anthony, Quentin and Biderman, Stella and Black, Sid and Gali, Preetham and Gao, Leo and Hallahan, Eric and Levy-Kramer, Josh and Leahy, Connor and Nestler, Lucas and Parker, Kip and Pieler, Michael and Purohit, Shivanshu and Songz, Tri and Phil, Wang and Weinbach, Samuel},
|
161 |
-
|
162 |
-
|
163 |
-
|
164 |
-
|
165 |
-
|
166 |
}
|
167 |
|
168 |
@inproceedings{aminabadi2022deepspeed,
|
169 |
-
title={{DeepSpeed-Inference}: enabling efficient inference of transformer models at unprecedented scale},
|
170 |
-
author={Aminabadi, Reza Yazdani and Rajbhandari, Samyam and Awan, Ammar Ahmad and Li, Cheng and Li, Du and Zheng, Elton and Ruwase, Olatunji and Smith, Shaden and Zhang, Minjia and Rasley, Jeff and others},
|
171 |
-
booktitle={SC22: International Conference for High Performance Computing, Networking, Storage and Analysis},
|
172 |
-
pages={1--15},
|
173 |
-
year={2022}
|
|
|
174 |
}
|
175 |
```
|
176 |
---
|
|
|
41 |
|
42 |
* **Training**
|
43 |
|
44 |
+
The model was trained on approximately 19,000 hours of following Japanese speech corpus ReazonSpeech v1.
|
45 |
Note that speech samples longer than 16 seconds were excluded before training.
|
46 |
- [ReazonSpeech](https://huggingface.co/datasets/reazon-research/reazonspeech)
|
47 |
|
|
|
128 |
|
129 |
# How to cite
|
130 |
```bibtex
|
131 |
+
@inproceedings{hono2024integrating,
|
132 |
+
title = {Integrating Pre-Trained Speech and Language Models for End-to-End Speech Recognition},
|
133 |
+
author = {Hono, Yukiya and Mitsuda, Koh and Zhao, Tianyu and Mitsui, Kentaro and Wakatsuki, Toshiaki and Sawada, Kei},
|
134 |
+
booktitle = {Findings of the Association for Computational Linguistics: ACL 2024},
|
135 |
+
year = {2024}
|
136 |
}
|
137 |
|
138 |
@misc{rinna-nue-asr,
|
139 |
+
title = {rinna/nue-asr},
|
140 |
+
author = {Hono, Yukiya and Mitsuda, Koh and Zhao, Tianyu and Mitsui, Kentaro and Wakatsuki, Toshiaki and Sawada, Kei},
|
141 |
+
url = {https://huggingface.co/rinna/nue-asr}
|
142 |
}
|
143 |
```
|
144 |
---
|
145 |
|
146 |
# References
|
147 |
```bibtex
|
148 |
+
@inproceedings{sawada2024release,
|
149 |
+
title = {Release of Pre-Trained Models for the {J}apanese Language},
|
150 |
+
author = {Sawada, Kei and Zhao, Tianyu and Shing, Makoto and Mitsui, Kentaro and Kaga, Akio and Hono, Yukiya and Wakatsuki, Toshiaki and Mitsuda, Koh},
|
151 |
+
booktitle = {Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)},
|
152 |
+
month = {5},
|
153 |
+
year = {2024},
|
154 |
+
pages = {13898--13905},
|
155 |
+
url = {https://aclanthology.org/2024.lrec-main.1213},
|
156 |
+
note = {\url{https://arxiv.org/abs/2404.01657}}
|
157 |
+
}
|
158 |
+
|
159 |
@article{hsu2021hubert,
|
160 |
+
title = {{HuBERT}: Self-Supervised Speech Representation Learning by Masked Prediction of Hidden Units},
|
161 |
+
author = {Hsu, Wei-Ning and Bolte, Benjamin and Tsai, Yao-Hung Hubert and Lakhotia, Kushal and Salakhutdinov, Ruslan and Mohamed, Abdelrahman},
|
162 |
+
journal = {IEEE/ACM Transactions on Audio, Speech, and Language Processing},
|
163 |
+
month = {10},
|
164 |
+
year = {2021},
|
165 |
+
volume = {29},
|
166 |
+
pages = {3451-3460},
|
167 |
+
doi = {10.1109/TASLP.2021.3122291}
|
168 |
}
|
169 |
|
170 |
@software{andoniangpt2021gpt,
|
171 |
+
title = {{GPT}-{N}eo{X}: Large Scale Autoregressive Language Modeling in {P}y{T}orch},
|
172 |
+
author = {Andonian, Alex and Anthony, Quentin and Biderman, Stella and Black, Sid and Gali, Preetham and Gao, Leo and Hallahan, Eric and Levy-Kramer, Josh and Leahy, Connor and Nestler, Lucas and Parker, Kip and Pieler, Michael and Purohit, Shivanshu and Songz, Tri and Phil, Wang and Weinbach, Samuel},
|
173 |
+
month = {8},
|
174 |
+
year = {2021},
|
175 |
+
version = {0.0.1},
|
176 |
+
doi = {10.5281/zenodo.5879544},
|
177 |
+
url = {https://www.github.com/eleutherai/gpt-neox}
|
178 |
}
|
179 |
|
180 |
@inproceedings{aminabadi2022deepspeed,
|
181 |
+
title = {{DeepSpeed-Inference}: enabling efficient inference of transformer models at unprecedented scale},
|
182 |
+
author = {Aminabadi, Reza Yazdani and Rajbhandari, Samyam and Awan, Ammar Ahmad and Li, Cheng and Li, Du and Zheng, Elton and Ruwase, Olatunji and Smith, Shaden and Zhang, Minjia and Rasley, Jeff and others},
|
183 |
+
booktitle = {SC22: International Conference for High Performance Computing, Networking, Storage and Analysis},
|
184 |
+
pages = {1--15},
|
185 |
+
year = {2022},
|
186 |
+
doi = {10.1109/SC41404.2022.00051}
|
187 |
}
|
188 |
```
|
189 |
---
|