wav2vec2-xlsr-1b-mecita-portuguese-all-grade-2-5
This model is a fine-tuned version of jonatasgrosman/wav2vec2-xls-r-1b-portuguese on an unknown dataset. It achieves the following results on the evaluation set:
- Loss: 0.2238
- Wer: 0.1009
- Cer: 0.0334
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 3e-05
- train_batch_size: 16
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 2
- total_train_batch_size: 32
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 100
- mixed_precision_training: Native AMP
Training results
Training Loss | Epoch | Step | Validation Loss | Wer | Cer |
---|---|---|---|---|---|
31.1421 | 1.0 | 58 | 7.4643 | 1.0 | 0.9272 |
8.4474 | 2.0 | 116 | 2.9703 | 0.9992 | 0.9997 |
8.4474 | 3.0 | 174 | 2.8949 | 1.0 | 1.0 |
2.9538 | 4.0 | 232 | 2.2467 | 0.9973 | 0.9806 |
2.9538 | 5.0 | 290 | 0.4163 | 0.2921 | 0.0764 |
1.7017 | 6.0 | 348 | 0.2723 | 0.1421 | 0.0436 |
0.4188 | 7.0 | 406 | 0.2676 | 0.1280 | 0.0421 |
0.4188 | 8.0 | 464 | 0.2381 | 0.1127 | 0.0379 |
0.3001 | 9.0 | 522 | 0.2585 | 0.1127 | 0.0361 |
0.3001 | 10.0 | 580 | 0.2487 | 0.1084 | 0.0359 |
0.2349 | 11.0 | 638 | 0.2271 | 0.1029 | 0.0332 |
0.2349 | 12.0 | 696 | 0.2238 | 0.1009 | 0.0334 |
0.2051 | 13.0 | 754 | 0.2451 | 0.1025 | 0.0333 |
0.1752 | 14.0 | 812 | 0.2595 | 0.1029 | 0.0343 |
0.1752 | 15.0 | 870 | 0.2571 | 0.1048 | 0.0350 |
0.1839 | 16.0 | 928 | 0.2540 | 0.1044 | 0.0344 |
0.1839 | 17.0 | 986 | 0.2417 | 0.1064 | 0.0357 |
0.1664 | 18.0 | 1044 | 0.2350 | 0.1005 | 0.0331 |
0.1796 | 19.0 | 1102 | 0.2415 | 0.0989 | 0.0341 |
0.1796 | 20.0 | 1160 | 0.2348 | 0.0989 | 0.0334 |
0.1446 | 21.0 | 1218 | 0.2538 | 0.0958 | 0.0327 |
0.1446 | 22.0 | 1276 | 0.2357 | 0.0954 | 0.0317 |
0.1518 | 23.0 | 1334 | 0.2329 | 0.0911 | 0.0314 |
0.1518 | 24.0 | 1392 | 0.2412 | 0.0989 | 0.0322 |
0.1185 | 25.0 | 1450 | 0.2365 | 0.0919 | 0.0310 |
0.1273 | 26.0 | 1508 | 0.2390 | 0.0958 | 0.0314 |
0.1273 | 27.0 | 1566 | 0.2380 | 0.0946 | 0.0318 |
0.115 | 28.0 | 1624 | 0.2333 | 0.0970 | 0.0322 |
0.115 | 29.0 | 1682 | 0.2393 | 0.0934 | 0.0313 |
0.1244 | 30.0 | 1740 | 0.2354 | 0.0974 | 0.0319 |
0.1244 | 31.0 | 1798 | 0.2622 | 0.0938 | 0.0323 |
0.1189 | 32.0 | 1856 | 0.2796 | 0.0982 | 0.0322 |
Framework versions
- Transformers 4.28.0
- Pytorch 2.2.1+cu121
- Datasets 2.17.0
- Tokenizers 0.13.3
- Downloads last month
- 0
This model does not have enough activity to be deployed to Inference API (serverless) yet.
Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.