yky-h commited on
Commit
b6684c8
1 Parent(s): 2199018

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +41 -28
README.md CHANGED
@@ -41,7 +41,7 @@ Benchmark scores, including our models, can be found at https://rinnakk.github.i
41
 
42
  * **Training**
43
 
44
- The model was trained on approximately 19,000 hours of following Japanese speech corpus ReazonSpeech.
45
  Note that speech samples longer than 16 seconds were excluded before training.
46
  - [ReazonSpeech](https://huggingface.co/datasets/reazon-research/reazonspeech)
47
 
@@ -128,49 +128,62 @@ The model uses the same sentencepiece-based tokenizer as [japanese-gpt-neox-3.6b
128
 
129
  # How to cite
130
  ```bibtex
131
- @article{hono2023integration,
132
- title={An Integration of Pre-Trained Speech and Language Models for End-to-End Speech Recognition},
133
- author={Hono, Yukiya and Mitsuda, Koh and Zhao, Tianyu and Mitsui, Kentaro and Wakatsuki, Toshiaki and Sawada, Kei},
134
- journal={arXiv preprint arXiv:2312.03668},
135
- year={2023}
136
  }
137
 
138
  @misc{rinna-nue-asr,
139
- title={rinna/nue-asr},
140
- author={Hono, Yukiya and Mitsuda, Koh and Zhao, Tianyu and Mitsui, Kentaro and Wakatsuki, Toshiaki and Sawada, Kei},
141
- url={https://huggingface.co/rinna/nue-asr}
142
  }
143
  ```
144
  ---
145
 
146
  # References
147
  ```bibtex
 
 
 
 
 
 
 
 
 
 
 
148
  @article{hsu2021hubert,
149
- title={HuBERT: Self-Supervised Speech Representation Learning by Masked Prediction of Hidden Units},
150
- author={Hsu, Wei-Ning and Bolte, Benjamin and Tsai, Yao-Hung Hubert and Lakhotia, Kushal and Salakhutdinov, Ruslan and Mohamed, Abdelrahman},
151
- journal={IEEE/ACM Transactions on Audio, Speech, and Language Processing},
152
- year={2021},
153
- volume={29},
154
- pages={3451-3460},
155
- doi={10.1109/TASLP.2021.3122291}
 
156
  }
157
 
158
  @software{andoniangpt2021gpt,
159
- title={{GPT-NeoX: Large Scale Autoregressive Language Modeling in PyTorch}},
160
- author={Andonian, Alex and Anthony, Quentin and Biderman, Stella and Black, Sid and Gali, Preetham and Gao, Leo and Hallahan, Eric and Levy-Kramer, Josh and Leahy, Connor and Nestler, Lucas and Parker, Kip and Pieler, Michael and Purohit, Shivanshu and Songz, Tri and Phil, Wang and Weinbach, Samuel},
161
- url={https://www.github.com/eleutherai/gpt-neox},
162
- doi={10.5281/zenodo.5879544},
163
- month={8},
164
- year={2021},
165
- version={0.0.1},
166
  }
167
 
168
  @inproceedings{aminabadi2022deepspeed,
169
- title={{DeepSpeed-Inference}: enabling efficient inference of transformer models at unprecedented scale},
170
- author={Aminabadi, Reza Yazdani and Rajbhandari, Samyam and Awan, Ammar Ahmad and Li, Cheng and Li, Du and Zheng, Elton and Ruwase, Olatunji and Smith, Shaden and Zhang, Minjia and Rasley, Jeff and others},
171
- booktitle={SC22: International Conference for High Performance Computing, Networking, Storage and Analysis},
172
- pages={1--15},
173
- year={2022}
 
174
  }
175
  ```
176
  ---
 
41
 
42
  * **Training**
43
 
44
+ The model was trained on approximately 19,000 hours of following Japanese speech corpus ReazonSpeech v1.
45
  Note that speech samples longer than 16 seconds were excluded before training.
46
  - [ReazonSpeech](https://huggingface.co/datasets/reazon-research/reazonspeech)
47
 
 
128
 
129
  # How to cite
130
  ```bibtex
131
+ @inproceedings{hono2024integrating,
132
+ title = {Integrating Pre-Trained Speech and Language Models for End-to-End Speech Recognition},
133
+ author = {Hono, Yukiya and Mitsuda, Koh and Zhao, Tianyu and Mitsui, Kentaro and Wakatsuki, Toshiaki and Sawada, Kei},
134
+ booktitle = {Findings of the Association for Computational Linguistics: ACL 2024},
135
+ year = {2024}
136
  }
137
 
138
  @misc{rinna-nue-asr,
139
+ title = {rinna/nue-asr},
140
+ author = {Hono, Yukiya and Mitsuda, Koh and Zhao, Tianyu and Mitsui, Kentaro and Wakatsuki, Toshiaki and Sawada, Kei},
141
+ url = {https://huggingface.co/rinna/nue-asr}
142
  }
143
  ```
144
  ---
145
 
146
  # References
147
  ```bibtex
148
+ @inproceedings{sawada2024release,
149
+ title = {Release of Pre-Trained Models for the {J}apanese Language},
150
+ author = {Sawada, Kei and Zhao, Tianyu and Shing, Makoto and Mitsui, Kentaro and Kaga, Akio and Hono, Yukiya and Wakatsuki, Toshiaki and Mitsuda, Koh},
151
+ booktitle = {Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)},
152
+ month = {5},
153
+ year = {2024},
154
+ pages = {13898--13905},
155
+ url = {https://aclanthology.org/2024.lrec-main.1213},
156
+ note = {\url{https://arxiv.org/abs/2404.01657}}
157
+ }
158
+
159
  @article{hsu2021hubert,
160
+ title = {{HuBERT}: Self-Supervised Speech Representation Learning by Masked Prediction of Hidden Units},
161
+ author = {Hsu, Wei-Ning and Bolte, Benjamin and Tsai, Yao-Hung Hubert and Lakhotia, Kushal and Salakhutdinov, Ruslan and Mohamed, Abdelrahman},
162
+ journal = {IEEE/ACM Transactions on Audio, Speech, and Language Processing},
163
+ month = {10},
164
+ year = {2021},
165
+ volume = {29},
166
+ pages = {3451-3460},
167
+ doi = {10.1109/TASLP.2021.3122291}
168
  }
169
 
170
  @software{andoniangpt2021gpt,
171
+ title = {{GPT}-{N}eo{X}: Large Scale Autoregressive Language Modeling in {P}y{T}orch},
172
+ author = {Andonian, Alex and Anthony, Quentin and Biderman, Stella and Black, Sid and Gali, Preetham and Gao, Leo and Hallahan, Eric and Levy-Kramer, Josh and Leahy, Connor and Nestler, Lucas and Parker, Kip and Pieler, Michael and Purohit, Shivanshu and Songz, Tri and Phil, Wang and Weinbach, Samuel},
173
+ month = {8},
174
+ year = {2021},
175
+ version = {0.0.1},
176
+ doi = {10.5281/zenodo.5879544},
177
+ url = {https://www.github.com/eleutherai/gpt-neox}
178
  }
179
 
180
  @inproceedings{aminabadi2022deepspeed,
181
+ title = {{DeepSpeed-Inference}: enabling efficient inference of transformer models at unprecedented scale},
182
+ author = {Aminabadi, Reza Yazdani and Rajbhandari, Samyam and Awan, Ammar Ahmad and Li, Cheng and Li, Du and Zheng, Elton and Ruwase, Olatunji and Smith, Shaden and Zhang, Minjia and Rasley, Jeff and others},
183
+ booktitle = {SC22: International Conference for High Performance Computing, Networking, Storage and Analysis},
184
+ pages = {1--15},
185
+ year = {2022},
186
+ doi = {10.1109/SC41404.2022.00051}
187
  }
188
  ```
189
  ---