SalahZa's picture
Update README.md
221e46e
|
raw
history blame
No virus
3.69 kB
# Overview
This project aims to create an Automatic Speech Recognition (ASR) model dedicated for the Tunisian Arabic dialect. The goal is to improve speech recognition technology for underrepresented linguistic communities by transcribing Tunisian dialect speech into written text.
## Dataset
Part of the audio and text data (The ones we collected) used to train and test the model has been provided to encourage and support research within the community. Please find the dataset [here](https://zenodo.org/record/8370566). This Zenodo record contains labeled and unlabeled Tunisian Arabic audio data, along with textual data for language modelling.
The folder also contains a 4-gram language model trained with KenLM on data released within the Zenodo record. The .arpa file is called "outdomain.arpa".
## Performance
The following table summarizes the performance of the model on various considered test sets :
| Dataset | CER | WER |
| :-------- | :------- | :------------------------- |
| `TARIC` | `6.22%` | `10.55%` |
| `IWSLT` | `21.18%` | `39.53%` |
| `TunSwitch TO` | `9.67%` | `25.54%` |
More details about the test sets, and the conditions leading to this performance in the paper.
## Team
Here are the team members who have contributed to this project
* [Salah Zaiem](https://fr.linkedin.com/in/salah-zaiem)
* [Ahmed Amine Ben Aballah](https://www.linkedin.com/in/aabenz/)
* [Ata Kaboudi](https://www.linkedin.com/in/ata-kaboudi-63365b1a8)
* [Amir Kanoun](https://tn.linkedin.com/in/ahmed-amir-kanoun)
## Paper
More in-depth details and insights are available in a released preprint. Please find the paper [here](https://arxiv.org/abs/2309.11327).
If you use or refer to this model, please cite :
```
@misc{abdallah2023leveraging,
title={Leveraging Data Collection and Unsupervised Learning for Code-switched Tunisian Arabic Automatic Speech Recognition},
author={Ahmed Amine Ben Abdallah and Ata Kabboudi and Amir Kanoun and Salah Zaiem},
year={2023},
eprint={2309.11327},
archivePrefix={arXiv},
primaryClass={eess.AS}
}
```
## Datasets
This ASR model was trained on :
* TARIC : The corpus, named TARIC (Tunisian Arabic Railway Interaction Corpus) has a collection of audio recordings and transcriptions from dialogues in the Tunisian Railway Transport Network. - [Taric Corpus](https://aclanthology.org/L14-1385/) -
* IWSLT : A Tunisian conversational speech - [IWSLT Corpus](https://iwslt.org/2022/dialect)-
* TunSwitch : Our crowd-collected dataset described in the paper presented above.
## Demo
Here is a working live demo : [LINK](https://huggingface.co/spaces/SalahZa/Tunisian-Speech-Recognition)
## Inference
### 1. Create a CSV test file
First, you have to create a csv file that follows SpeechBrain's format which contain 4 columns:
* ID: contain ID to identify each audio sample in the dataset
* wav: contain the path to the audio file
* wrd: contain the text transcription of the spoken content in the audio file if you have it and use your set for evaluation. Put anything if you don't have transcriptions. An example is provided in this folder, the file is called : taric_test.csv
* duration: the duration of the audio in seconds
### 2. Adjust the hyperparams.yaml file
Adjust the path of **test_csv** parameter to your csv file path
To run this recipe, do the following:
```
> python train_with_wavlm.py semi_wavlm_large_tunisian_ctc/1234/hyperparams.yaml --test_csv = path_to_csv
```
If you want to infer on single files, the space demo offers proper easy-to-use inference code.
## Contact :
If you have questions, you can send an email to : zaiemsalah@gmail.com