Update README.md
Browse files
README.md
CHANGED
@@ -64,7 +64,39 @@ Common Voice 8 mt dataset has been used for the model
|
|
64 |
## Training procedure
|
65 |
|
66 |
### Training hyperparameters
|
67 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
68 |
|
69 |
### Framework versions
|
70 |
- Transformers 4.16.0.dev0
|
|
|
64 |
## Training procedure
|
65 |
|
66 |
### Training hyperparameters
|
67 |
+
The following config and hyperparameters were used during training:
|
68 |
+
model = Wav2Vec2ForCTC.from_pretrained(
|
69 |
+
"facebook/wav2vec2-xls-r-1b",
|
70 |
+
attention_dropout=0.05,
|
71 |
+
hidden_dropout=0.05,
|
72 |
+
feat_proj_dropout=0.05,
|
73 |
+
mask_time_prob=0.55,
|
74 |
+
mask_feature_prob=0.10,
|
75 |
+
layerdrop=0.05,
|
76 |
+
ctc_zero_infinity=True,
|
77 |
+
ctc_loss_reduction="mean",
|
78 |
+
pad_token_id=processor.tokenizer.pad_token_id,
|
79 |
+
vocab_size=len(processor.tokenizer),
|
80 |
+
)
|
81 |
+
from transformers import TrainingArguments
|
82 |
+
|
83 |
+
training_args = TrainingArguments(
|
84 |
+
output_dir=repo_name,
|
85 |
+
group_by_length=True,
|
86 |
+
per_device_train_batch_size=32,
|
87 |
+
gradient_accumulation_steps=2,
|
88 |
+
evaluation_strategy="steps",
|
89 |
+
num_train_epochs=50,
|
90 |
+
gradient_checkpointing=True,
|
91 |
+
fp16=True,
|
92 |
+
save_steps=400,
|
93 |
+
eval_steps=400,
|
94 |
+
logging_steps=400,
|
95 |
+
learning_rate=5.5e-05,
|
96 |
+
warmup_steps=500,
|
97 |
+
save_total_limit=2,
|
98 |
+
push_to_hub=True,
|
99 |
+
report_to="tensorboard")
|
100 |
|
101 |
### Framework versions
|
102 |
- Transformers 4.16.0.dev0
|