Edit model card

This model is a fine-tuned version of facebook/wav2vec2-xls-r-300m on the OPENSLR_SLR53 - bengali dataset. It achieves the following results on the evaluation set.

Without language model :

  • WER: 0.21726385291857586
  • CER: 0.04725010353701041

With 5 gram language model trained on 30M sentences randomly chosen from AI4Bharat IndicCorp dataset :

  • WER: 0.15322879016421437
  • CER: 0.03413696666806267

Note : 5% of a total 10935 samples have been used for evaluation. Evaluation set has 10935 examples which was not part of training training was done on first 95% and eval was done on last 5%. Training was stopped after 180k steps. Output predictions are available under files section.

Training hyperparameters

The following hyperparameters were used during training:

  • dataset_name="openslr"
  • model_name_or_path="facebook/wav2vec2-xls-r-300m"
  • dataset_config_name="SLR53"
  • output_dir="./wav2vec2-xls-r-300m-bengali"
  • overwrite_output_dir
  • num_train_epochs="50"
  • per_device_train_batch_size="32"
  • per_device_eval_batch_size="32"
  • gradient_accumulation_steps="1"
  • learning_rate="7.5e-5"
  • warmup_steps="2000"
  • length_column_name="input_length"
  • evaluation_strategy="steps"
  • text_column_name="sentence"
  • chars_to_ignore , ? . ! - ; : " “ % ‘ ” � — ’ … –
  • save_steps="2000"
  • eval_steps="3000"
  • logging_steps="100"
  • layerdrop="0.0"
  • activation_dropout="0.1"
  • save_total_limit="3"
  • freeze_feature_encoder
  • feat_proj_dropout="0.0"
  • mask_time_prob="0.75"
  • mask_time_length="10"
  • mask_feature_prob="0.25"
  • mask_feature_length="64"
  • preprocessing_num_workers 32

Framework versions

  • Transformers 4.16.0.dev0
  • Pytorch 1.10.1+cu102
  • Datasets 1.17.1.dev0
  • Tokenizers 0.11.0

Notes

Downloads last month
100

Dataset used to train arijitx/wav2vec2-xls-r-300m-bengali

Space using arijitx/wav2vec2-xls-r-300m-bengali 1

Evaluation results