YingxuHe commited on
Commit
ea45b1a
·
verified ·
1 Parent(s): 735299c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -4
README.md CHANGED
@@ -66,10 +66,8 @@ MERaLiON-AudioLLM is trained to mainly address 6 tasks, namely `Automatic Speech
66
  `Spoken Dialogue Summarization` (SDS), `Speech Instruction` (SI), and `Paralinguistics` (PARA).
67
 
68
  We benchmark MERaLiON-AudioLLM with a series of test sets from the [AudioBench benchmark](https://github.com/AudioLLMs/AudioBench)
69
- against three well-known AudioLLMs: `Qwen2-Audio 7B`, `WavLLM`, and `SALMONN`. We also compared with a cascaded model,
70
- which feeds the transcriptions recognized by Whisper-large-v2 and the instruction prompts to a Gemma2 9B CPT SEA-LIONv3 Instruct model to
71
- get the responses. We tuned its hyperparameters and prompt template to optimise performance across
72
- various speech-to-text tasks. As is shown in the following table, MERaLiON-AudioLLM performs better in the Singapore local context,
73
  as evidenced by evaluation results on Singapore's [Multitask National Speech Corpus](https://huggingface.co/datasets/MERaLiON/Multitask-National-Speech-Corpus-v1) (MNSC) datasets.
74
 
75
  > [!NOTE]
 
66
  `Spoken Dialogue Summarization` (SDS), `Speech Instruction` (SI), and `Paralinguistics` (PARA).
67
 
68
  We benchmark MERaLiON-AudioLLM with a series of test sets from the [AudioBench benchmark](https://github.com/AudioLLMs/AudioBench)
69
+ against three well-known AudioLLMs: `Qwen2-Audio 7B`, `WavLLM`, `SALMONN`, and a cascaded model.
70
+ As is shown in the following table, MERaLiON-AudioLLM performs better in the Singapore local context,
 
 
71
  as evidenced by evaluation results on Singapore's [Multitask National Speech Corpus](https://huggingface.co/datasets/MERaLiON/Multitask-National-Speech-Corpus-v1) (MNSC) datasets.
72
 
73
  > [!NOTE]