Automatic Speech Recognition
Transformers
4 languages
whisper
whisper-event
Generated from Trainer
Inference Endpoints
marinone94 commited on
Commit
b9d77dc
1 Parent(s): fa4a655

fix language hard coding

Browse files
run_speech_recognition_seq2seq_streaming.py CHANGED
@@ -49,7 +49,7 @@ from transformers import (
49
  set_seed,
50
  )
51
  from transformers.models.whisper.english_normalizer import BasicTextNormalizer
52
- from transformers.models.whisper.tokenization_whisper import TO_LANGUAGE_CODE
53
  from transformers.trainer_pt_utils import IterableDatasetShard
54
  from transformers.trainer_utils import get_last_checkpoint, is_main_process
55
  from transformers.utils import check_min_version, send_example_telemetry
@@ -843,8 +843,9 @@ def main():
843
  pred_labels = [f"Prediction: {pred}\nLabel: {label}\n" for pred, label in zip(preds, labels)]
844
  logger.info("Before setting language and task")
845
  logger.info(f"{pred_labels}")
 
846
  trainer.model.config.forced_decoder_ids = \
847
- tokenizer.get_decoder_prompt_ids(language=data_args.language_eval, task=data_args.task, no_timestamps=True)
848
  preds = tokenizer.batch_decode(predictions.predictions)
849
  labels = tokenizer.batch_decode(predictions.label_ids)
850
  pred_labels = [f"Prediction: {pred}\nLabel: {label}\n" for pred, label in zip(preds, labels)]
 
49
  set_seed,
50
  )
51
  from transformers.models.whisper.english_normalizer import BasicTextNormalizer
52
+ from transformers.models.whisper.tokenization_whisper import TO_LANGUAGE_CODE, LANGUAGES
53
  from transformers.trainer_pt_utils import IterableDatasetShard
54
  from transformers.trainer_utils import get_last_checkpoint, is_main_process
55
  from transformers.utils import check_min_version, send_example_telemetry
 
843
  pred_labels = [f"Prediction: {pred}\nLabel: {label}\n" for pred, label in zip(preds, labels)]
844
  logger.info("Before setting language and task")
845
  logger.info(f"{pred_labels}")
846
+ language_name = LANGUAGES[data_args.language_eval]
847
  trainer.model.config.forced_decoder_ids = \
848
+ tokenizer.get_decoder_prompt_ids(language_name, task=data_args.task, no_timestamps=True)
849
  preds = tokenizer.batch_decode(predictions.predictions)
850
  labels = tokenizer.batch_decode(predictions.label_ids)
851
  pred_labels = [f"Prediction: {pred}\nLabel: {label}\n" for pred, label in zip(preds, labels)]
test_run_nordic.sh CHANGED
@@ -9,19 +9,17 @@ python $1run_speech_recognition_seq2seq_streaming.py \
9
  --language_eval="sv" \
10
  --eval_split_name="test" \
11
  --model_index_name="Whisper Tiny Nordic" \
12
- --max_train_samples="64" \
13
- --max_eval_samples="32" \
14
- --max_steps="1" \
15
  --output_dir="./" \
16
- --per_device_train_batch_size="1" \
17
- --per_device_eval_batch_size="1" \
18
  --logging_steps="25" \
19
  --learning_rate="1e-5" \
20
- --warmup_steps="500" \
21
  --evaluation_strategy="steps" \
22
- --eval_steps="1000" \
23
  --save_strategy="steps" \
24
- --save_steps="1000" \
25
  --generation_max_length="225" \
26
  --length_column_name="input_length" \
27
  --max_duration_in_seconds="30" \
 
9
  --language_eval="sv" \
10
  --eval_split_name="test" \
11
  --model_index_name="Whisper Tiny Nordic" \
12
+ --max_steps="500" \
 
 
13
  --output_dir="./" \
14
+ --per_device_train_batch_size="128" \
15
+ --per_device_eval_batch_size="64" \
16
  --logging_steps="25" \
17
  --learning_rate="1e-5" \
18
+ --warmup_steps="50" \
19
  --evaluation_strategy="steps" \
20
+ --eval_steps="100" \
21
  --save_strategy="steps" \
22
+ --save_steps="100" \
23
  --generation_max_length="225" \
24
  --length_column_name="input_length" \
25
  --max_duration_in_seconds="30" \