Edit model card

hubert-base-libri-pruning-v1

This model was trained from scratch on the None dataset. It achieves the following results on the evaluation set:

  • Loss: -7634.6143
  • Wer: 0.4250

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.00015
  • train_batch_size: 64
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_steps: 3000
  • num_epochs: 30
  • mixed_precision_training: Native AMP

Training results

Training Loss Epoch Step Validation Loss Wer
-14.3776 1.12 500 -47.0374 0.5235
-106.5986 2.24 1000 -189.7623 0.4825
-291.2003 3.36 1500 -428.0651 0.4726
-567.486 4.48 2000 -761.7246 0.4628
-934.1564 5.61 2500 -1190.6781 0.4580
-1396.4066 6.73 3000 -1714.8335 0.4487
-1930.5184 7.85 3500 -2271.9685 0.4422
-2452.421 8.97 4000 -2802.5127 0.4418
-2953.0952 10.09 4500 -3304.3562 0.4386
-3427.7243 11.21 5000 -3779.5505 0.4357
-3879.0445 12.33 5500 -4227.1289 0.4339
-4302.3395 13.45 6000 -4647.1260 0.4311
-4680.295 14.57 6500 -5039.4692 0.4283
-5054.8855 15.7 7000 -5404.1592 0.4306
-5377.8435 16.82 7500 -5741.4082 0.4286
-5688.8665 17.94 8000 -6051.0688 0.4290
-5990.955 19.06 8500 -6333.1387 0.4284
-6252.404 20.18 9000 -6587.1460 0.4257
-6481.961 21.3 9500 -6814.2788 0.4268
-6695.5835 22.42 10000 -7013.8809 0.4256
-6859.0875 23.54 10500 -7185.9956 0.4255
-7015.5155 24.66 11000 -7330.6577 0.4271
-7170.5215 25.78 11500 -7447.6372 0.4256
-7244.894 26.91 12000 -7537.4111 0.4243
-7320.932 28.03 12500 -7599.7690 0.4248
-7366.4105 29.15 13000 -7634.6143 0.4250

Framework versions

  • Transformers 4.30.0.dev0
  • Pytorch 2.0.1
  • Datasets 2.12.1.dev0
  • Tokenizers 0.13.3
Downloads last month
0
Inference API
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.