marinone94
commited on
Commit
•
b9d77dc
1
Parent(s):
fa4a655
fix language hard coding
Browse files
run_speech_recognition_seq2seq_streaming.py
CHANGED
@@ -49,7 +49,7 @@ from transformers import (
|
|
49 |
set_seed,
|
50 |
)
|
51 |
from transformers.models.whisper.english_normalizer import BasicTextNormalizer
|
52 |
-
from transformers.models.whisper.tokenization_whisper import TO_LANGUAGE_CODE
|
53 |
from transformers.trainer_pt_utils import IterableDatasetShard
|
54 |
from transformers.trainer_utils import get_last_checkpoint, is_main_process
|
55 |
from transformers.utils import check_min_version, send_example_telemetry
|
@@ -843,8 +843,9 @@ def main():
|
|
843 |
pred_labels = [f"Prediction: {pred}\nLabel: {label}\n" for pred, label in zip(preds, labels)]
|
844 |
logger.info("Before setting language and task")
|
845 |
logger.info(f"{pred_labels}")
|
|
|
846 |
trainer.model.config.forced_decoder_ids = \
|
847 |
-
tokenizer.get_decoder_prompt_ids(
|
848 |
preds = tokenizer.batch_decode(predictions.predictions)
|
849 |
labels = tokenizer.batch_decode(predictions.label_ids)
|
850 |
pred_labels = [f"Prediction: {pred}\nLabel: {label}\n" for pred, label in zip(preds, labels)]
|
|
|
49 |
set_seed,
|
50 |
)
|
51 |
from transformers.models.whisper.english_normalizer import BasicTextNormalizer
|
52 |
+
from transformers.models.whisper.tokenization_whisper import TO_LANGUAGE_CODE, LANGUAGES
|
53 |
from transformers.trainer_pt_utils import IterableDatasetShard
|
54 |
from transformers.trainer_utils import get_last_checkpoint, is_main_process
|
55 |
from transformers.utils import check_min_version, send_example_telemetry
|
|
|
843 |
pred_labels = [f"Prediction: {pred}\nLabel: {label}\n" for pred, label in zip(preds, labels)]
|
844 |
logger.info("Before setting language and task")
|
845 |
logger.info(f"{pred_labels}")
|
846 |
+
language_name = LANGUAGES[data_args.language_eval]
|
847 |
trainer.model.config.forced_decoder_ids = \
|
848 |
+
tokenizer.get_decoder_prompt_ids(language_name, task=data_args.task, no_timestamps=True)
|
849 |
preds = tokenizer.batch_decode(predictions.predictions)
|
850 |
labels = tokenizer.batch_decode(predictions.label_ids)
|
851 |
pred_labels = [f"Prediction: {pred}\nLabel: {label}\n" for pred, label in zip(preds, labels)]
|
test_run_nordic.sh
CHANGED
@@ -9,19 +9,17 @@ python $1run_speech_recognition_seq2seq_streaming.py \
|
|
9 |
--language_eval="sv" \
|
10 |
--eval_split_name="test" \
|
11 |
--model_index_name="Whisper Tiny Nordic" \
|
12 |
-
--
|
13 |
-
--max_eval_samples="32" \
|
14 |
-
--max_steps="1" \
|
15 |
--output_dir="./" \
|
16 |
-
--per_device_train_batch_size="
|
17 |
-
--per_device_eval_batch_size="
|
18 |
--logging_steps="25" \
|
19 |
--learning_rate="1e-5" \
|
20 |
-
--warmup_steps="
|
21 |
--evaluation_strategy="steps" \
|
22 |
-
--eval_steps="
|
23 |
--save_strategy="steps" \
|
24 |
-
--save_steps="
|
25 |
--generation_max_length="225" \
|
26 |
--length_column_name="input_length" \
|
27 |
--max_duration_in_seconds="30" \
|
|
|
9 |
--language_eval="sv" \
|
10 |
--eval_split_name="test" \
|
11 |
--model_index_name="Whisper Tiny Nordic" \
|
12 |
+
--max_steps="500" \
|
|
|
|
|
13 |
--output_dir="./" \
|
14 |
+
--per_device_train_batch_size="128" \
|
15 |
+
--per_device_eval_batch_size="64" \
|
16 |
--logging_steps="25" \
|
17 |
--learning_rate="1e-5" \
|
18 |
+
--warmup_steps="50" \
|
19 |
--evaluation_strategy="steps" \
|
20 |
+
--eval_steps="100" \
|
21 |
--save_strategy="steps" \
|
22 |
+
--save_steps="100" \
|
23 |
--generation_max_length="225" \
|
24 |
--length_column_name="input_length" \
|
25 |
--max_duration_in_seconds="30" \
|