MooER: LLM-based Speech Recognition and Translation Models from Moore Threads
Abstract
In this paper, we present MooER, a LLM-based large-scale automatic speech recognition (ASR) / automatic speech translation (AST) model of Moore Threads. A 5000h pseudo labeled dataset containing open source and self collected speech data is used for training. We achieve performance comparable to other open source models trained with up to hundreds of thousands of hours of labeled speech data. Meanwhile, experiments conducted on Covost2 Zh2en testset suggest that our model outperforms other open source Speech LLMs. A BLEU score of 25.2 can be obtained. The main contributions of this paper are summarized as follows. First, this paper presents a training strategy for encoders and LLMs on speech related tasks (including ASR and AST) using a small size of pseudo labeled data without any extra manual annotation and selection. Second, we release our ASR and AST models and plan to open-source our training code and strategy in the near future. Moreover, a model trained on 8wh scale training data is planned to be released later on.
Community
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- BESTOW: Efficient and Streamable Speech Language Model with the Best of Two Worlds in GPT and T5 (2024)
- Transferable speech-to-text large language model alignment module (2024)
- Investigating Decoder-only Large Language Models for Speech-to-text Translation (2024)
- A Comprehensive Solution to Connect Speech Encoder and Large Language Model for ASR (2024)
- Pronunciation Assessment with Multi-modal Large Language Models (2024)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Thank you for the contribution, Moore Threads Team. What exactly is the novelty of this paper? Is it training only adapters instead of the full parameterized model?
Models citing this paper 2
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper