mikr commited on
Commit
c404c89
1 Parent(s): 1dba846

update model card README.md

Browse files
Files changed (2) hide show
  1. README.md +80 -0
  2. run.log +74 -0
README.md ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - generated_from_trainer
5
+ datasets:
6
+ - voxpopuli
7
+ metrics:
8
+ - wer
9
+ model-index:
10
+ - name: openai/whisper-small
11
+ results:
12
+ - task:
13
+ name: Automatic Speech Recognition
14
+ type: automatic-speech-recognition
15
+ dataset:
16
+ name: voxpopuli
17
+ type: voxpopuli
18
+ config: hr
19
+ split: test
20
+ args: hr
21
+ metrics:
22
+ - name: Wer
23
+ type: wer
24
+ value: 25.43475821833277
25
+ ---
26
+
27
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
28
+ should probably proofread and complete it, then remove this comment. -->
29
+
30
+ # openai/whisper-small
31
+
32
+ This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the voxpopuli dataset.
33
+ It achieves the following results on the evaluation set:
34
+ - Loss: 0.6405
35
+ - Wer: 25.4348
36
+
37
+ ## Model description
38
+
39
+ More information needed
40
+
41
+ ## Intended uses & limitations
42
+
43
+ More information needed
44
+
45
+ ## Training and evaluation data
46
+
47
+ More information needed
48
+
49
+ ## Training procedure
50
+
51
+ ### Training hyperparameters
52
+
53
+ The following hyperparameters were used during training:
54
+ - learning_rate: 1e-05
55
+ - train_batch_size: 64
56
+ - eval_batch_size: 32
57
+ - seed: 42
58
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
59
+ - lr_scheduler_type: linear
60
+ - lr_scheduler_warmup_steps: 500
61
+ - training_steps: 5000
62
+ - mixed_precision_training: Native AMP
63
+
64
+ ### Training results
65
+
66
+ | Training Loss | Epoch | Step | Validation Loss | Wer |
67
+ |:-------------:|:------:|:----:|:---------------:|:-------:|
68
+ | 0.0028 | 24.01 | 1000 | 0.5303 | 26.1752 |
69
+ | 0.0006 | 49.01 | 2000 | 0.5849 | 25.4123 |
70
+ | 0.0003 | 74.01 | 3000 | 0.6141 | 25.6311 |
71
+ | 0.0002 | 99.01 | 4000 | 0.6325 | 25.3731 |
72
+ | 0.0002 | 124.01 | 5000 | 0.6405 | 25.4348 |
73
+
74
+
75
+ ### Framework versions
76
+
77
+ - Transformers 4.26.0.dev0
78
+ - Pytorch 1.13.0+cu117
79
+ - Datasets 2.7.1.dev0
80
+ - Tokenizers 0.13.2
run.log CHANGED
@@ -363,3 +363,77 @@ xpu_backend=None,
363
  {'loss': 0.0003, 'learning_rate': 3.6133333333333336e-06, 'epoch': 84.0}
364
  {'loss': 0.0003, 'learning_rate': 3.5577777777777785e-06, 'epoch': 84.01}
365
  12/15/2022 06:41:03 - WARNING - datasets.download.streaming_download_manager - Got disconnected from remote data host. Retrying in 5sec [1/20]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
363
  {'loss': 0.0003, 'learning_rate': 3.6133333333333336e-06, 'epoch': 84.0}
364
  {'loss': 0.0003, 'learning_rate': 3.5577777777777785e-06, 'epoch': 84.01}
365
  12/15/2022 06:41:03 - WARNING - datasets.download.streaming_download_manager - Got disconnected from remote data host. Retrying in 5sec [1/20]
366
+ {'loss': 0.0003, 'learning_rate': 3.5022222222222225e-06, 'epoch': 85.0}
367
+ {'loss': 0.0002, 'learning_rate': 3.446666666666667e-06, 'epoch': 86.0}
368
+ {'loss': 0.0003, 'learning_rate': 3.391111111111111e-06, 'epoch': 86.01}
369
+ {'loss': 0.0002, 'learning_rate': 3.335555555555556e-06, 'epoch': 87.0}
370
+ {'loss': 0.0003, 'learning_rate': 3.2800000000000004e-06, 'epoch': 88.0}
371
+ {'loss': 0.0002, 'learning_rate': 3.2244444444444444e-06, 'epoch': 88.01}
372
+ {'loss': 0.0002, 'learning_rate': 3.1688888888888893e-06, 'epoch': 89.0}
373
+ {'loss': 0.0002, 'learning_rate': 3.1133333333333337e-06, 'epoch': 89.01}
374
+ {'loss': 0.0002, 'learning_rate': 3.0577777777777778e-06, 'epoch': 90.0}
375
+ {'loss': 0.0002, 'learning_rate': 3.0022222222222227e-06, 'epoch': 91.0}
376
+ {'loss': 0.0002, 'learning_rate': 2.946666666666667e-06, 'epoch': 91.01}
377
+ {'loss': 0.0002, 'learning_rate': 2.891111111111111e-06, 'epoch': 92.0}
378
+ {'loss': 0.0002, 'learning_rate': 2.835555555555556e-06, 'epoch': 93.0}
379
+ {'loss': 0.0002, 'learning_rate': 2.7800000000000005e-06, 'epoch': 93.01}
380
+ {'loss': 0.0002, 'learning_rate': 2.7244444444444445e-06, 'epoch': 94.0}
381
+ {'loss': 0.0002, 'learning_rate': 2.6688888888888894e-06, 'epoch': 94.01}
382
+ {'loss': 0.0002, 'learning_rate': 2.6133333333333334e-06, 'epoch': 95.0}
383
+ {'loss': 0.0002, 'learning_rate': 2.557777777777778e-06, 'epoch': 96.0}
384
+ {'loss': 0.0002, 'learning_rate': 2.5022222222222224e-06, 'epoch': 96.01}
385
+ {'loss': 0.0002, 'learning_rate': 2.446666666666667e-06, 'epoch': 97.0}
386
+ {'loss': 0.0002, 'learning_rate': 2.3911111111111113e-06, 'epoch': 98.0}
387
+ {'loss': 0.0002, 'learning_rate': 2.3355555555555557e-06, 'epoch': 98.01}
388
+ {'loss': 0.0002, 'learning_rate': 2.28e-06, 'epoch': 99.0}
389
+ {'loss': 0.0002, 'learning_rate': 2.2244444444444447e-06, 'epoch': 99.01}
390
+ {'eval_loss': 0.6324931383132935, 'eval_wer': 25.37305060024683, 'eval_runtime': 650.4597, 'eval_samples_per_second': 1.024, 'eval_steps_per_second': 0.032, 'epoch': 99.01}
391
+ {'loss': 0.0002, 'learning_rate': 2.168888888888889e-06, 'epoch': 100.0}
392
+ {'loss': 0.0002, 'learning_rate': 2.1133333333333336e-06, 'epoch': 101.0}
393
+ {'loss': 0.0002, 'learning_rate': 2.057777777777778e-06, 'epoch': 101.01}
394
+ {'loss': 0.0002, 'learning_rate': 2.0022222222222225e-06, 'epoch': 102.0}
395
+ {'loss': 0.0002, 'learning_rate': 1.9466666666666665e-06, 'epoch': 103.0}
396
+ {'loss': 0.0002, 'learning_rate': 1.8911111111111114e-06, 'epoch': 103.01}
397
+ {'loss': 0.0002, 'learning_rate': 1.8355555555555557e-06, 'epoch': 104.0}
398
+ {'loss': 0.0002, 'learning_rate': 1.7800000000000001e-06, 'epoch': 104.01}
399
+ {'loss': 0.0002, 'learning_rate': 1.7244444444444448e-06, 'epoch': 105.0}
400
+ {'loss': 0.0002, 'learning_rate': 1.668888888888889e-06, 'epoch': 106.0}
401
+ {'loss': 0.0002, 'learning_rate': 1.6133333333333335e-06, 'epoch': 106.01}
402
+ {'loss': 0.0002, 'learning_rate': 1.5577777777777777e-06, 'epoch': 107.0}
403
+ {'loss': 0.0002, 'learning_rate': 1.5022222222222224e-06, 'epoch': 108.0}
404
+ {'loss': 0.0002, 'learning_rate': 1.4466666666666669e-06, 'epoch': 108.01}
405
+ {'loss': 0.0002, 'learning_rate': 1.3911111111111111e-06, 'epoch': 109.0}
406
+ {'loss': 0.0002, 'learning_rate': 1.3355555555555558e-06, 'epoch': 109.01}
407
+ {'loss': 0.0002, 'learning_rate': 1.28e-06, 'epoch': 110.0}
408
+ {'loss': 0.0002, 'learning_rate': 1.2244444444444445e-06, 'epoch': 111.0}
409
+ {'loss': 0.0002, 'learning_rate': 1.168888888888889e-06, 'epoch': 111.01}
410
+ {'loss': 0.0002, 'learning_rate': 1.1133333333333334e-06, 'epoch': 112.0}
411
+ {'loss': 0.0002, 'learning_rate': 1.0577777777777779e-06, 'epoch': 113.0}
412
+ {'loss': 0.0002, 'learning_rate': 1.0022222222222223e-06, 'epoch': 113.01}
413
+ {'loss': 0.0002, 'learning_rate': 9.466666666666667e-07, 'epoch': 114.0}
414
+ {'loss': 0.0002, 'learning_rate': 8.911111111111112e-07, 'epoch': 114.01}
415
+ {'loss': 0.0002, 'learning_rate': 8.355555555555556e-07, 'epoch': 115.0}
416
+ {'loss': 0.0002, 'learning_rate': 7.8e-07, 'epoch': 116.0}
417
+ {'loss': 0.0002, 'learning_rate': 7.244444444444446e-07, 'epoch': 116.01}
418
+ {'loss': 0.0002, 'learning_rate': 6.68888888888889e-07, 'epoch': 117.0}
419
+ {'loss': 0.0002, 'learning_rate': 6.133333333333333e-07, 'epoch': 118.0}
420
+ {'loss': 0.0002, 'learning_rate': 5.577777777777779e-07, 'epoch': 118.01}
421
+ {'loss': 0.0002, 'learning_rate': 5.022222222222222e-07, 'epoch': 119.0}
422
+ {'loss': 0.0002, 'learning_rate': 4.466666666666667e-07, 'epoch': 119.01}
423
+ {'loss': 0.0002, 'learning_rate': 3.9111111111111115e-07, 'epoch': 120.0}
424
+ {'loss': 0.0002, 'learning_rate': 3.3555555555555556e-07, 'epoch': 121.0}
425
+ {'loss': 0.0002, 'learning_rate': 2.8e-07, 'epoch': 121.01}
426
+ {'loss': 0.0002, 'learning_rate': 2.2444444444444445e-07, 'epoch': 122.0}
427
+ {'loss': 0.0002, 'learning_rate': 1.6888888888888888e-07, 'epoch': 123.0}
428
+ {'loss': 0.0002, 'learning_rate': 1.1333333333333336e-07, 'epoch': 123.01}
429
+ {'loss': 0.0002, 'learning_rate': 5.777777777777778e-08, 'epoch': 124.0}
430
+ {'loss': 0.0002, 'learning_rate': 2.2222222222222225e-09, 'epoch': 124.01}
431
+ {'eval_loss': 0.6405040621757507, 'eval_wer': 25.43475821833277, 'eval_runtime': 648.4677, 'eval_samples_per_second': 1.027, 'eval_steps_per_second': 0.032, 'epoch': 124.01}
432
+ {'train_runtime': 39423.3866, 'train_samples_per_second': 8.117, 'train_steps_per_second': 0.127, 'train_loss': 0.04578730283072218, 'epoch': 124.01}
433
+ 12/15/2022 10:11:51 - WARNING - huggingface_hub.repository - Several commits (2) will be pushed upstream.
434
+ 12/15/2022 10:11:51 - WARNING - huggingface_hub.repository - The progress bars may be unreliable.
435
+ 12/15/2022 10:11:57 - WARNING - huggingface_hub.repository - remote: Scanning LFS files for validity, may be slow...
436
+ remote: LFS file scan complete.
437
+ To https://huggingface.co/mikr/whisper-small-hr-vox
438
+ ddcf8d0..1dba846 main -> main
439
+