File size: 2,859 Bytes
e9df99c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
---
tags:
- generated_from_trainer
metrics:
- wer
model-index:
- name: hubert-base-libri-pruning-v1
  results: []
---

<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->

# hubert-base-libri-pruning-v1

This model was trained from scratch on the None dataset.
It achieves the following results on the evaluation set:
- Loss: -7634.6143
- Wer: 0.4250

## Model description

More information needed

## Intended uses & limitations

More information needed

## Training and evaluation data

More information needed

## Training procedure

### Training hyperparameters

The following hyperparameters were used during training:
- learning_rate: 0.00015
- train_batch_size: 64
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 3000
- num_epochs: 30
- mixed_precision_training: Native AMP

### Training results

| Training Loss | Epoch | Step  | Validation Loss | Wer    |
|:-------------:|:-----:|:-----:|:---------------:|:------:|
| -14.3776      | 1.12  | 500   | -47.0374        | 0.5235 |
| -106.5986     | 2.24  | 1000  | -189.7623       | 0.4825 |
| -291.2003     | 3.36  | 1500  | -428.0651       | 0.4726 |
| -567.486      | 4.48  | 2000  | -761.7246       | 0.4628 |
| -934.1564     | 5.61  | 2500  | -1190.6781      | 0.4580 |
| -1396.4066    | 6.73  | 3000  | -1714.8335      | 0.4487 |
| -1930.5184    | 7.85  | 3500  | -2271.9685      | 0.4422 |
| -2452.421     | 8.97  | 4000  | -2802.5127      | 0.4418 |
| -2953.0952    | 10.09 | 4500  | -3304.3562      | 0.4386 |
| -3427.7243    | 11.21 | 5000  | -3779.5505      | 0.4357 |
| -3879.0445    | 12.33 | 5500  | -4227.1289      | 0.4339 |
| -4302.3395    | 13.45 | 6000  | -4647.1260      | 0.4311 |
| -4680.295     | 14.57 | 6500  | -5039.4692      | 0.4283 |
| -5054.8855    | 15.7  | 7000  | -5404.1592      | 0.4306 |
| -5377.8435    | 16.82 | 7500  | -5741.4082      | 0.4286 |
| -5688.8665    | 17.94 | 8000  | -6051.0688      | 0.4290 |
| -5990.955     | 19.06 | 8500  | -6333.1387      | 0.4284 |
| -6252.404     | 20.18 | 9000  | -6587.1460      | 0.4257 |
| -6481.961     | 21.3  | 9500  | -6814.2788      | 0.4268 |
| -6695.5835    | 22.42 | 10000 | -7013.8809      | 0.4256 |
| -6859.0875    | 23.54 | 10500 | -7185.9956      | 0.4255 |
| -7015.5155    | 24.66 | 11000 | -7330.6577      | 0.4271 |
| -7170.5215    | 25.78 | 11500 | -7447.6372      | 0.4256 |
| -7244.894     | 26.91 | 12000 | -7537.4111      | 0.4243 |
| -7320.932     | 28.03 | 12500 | -7599.7690      | 0.4248 |
| -7366.4105    | 29.15 | 13000 | -7634.6143      | 0.4250 |


### Framework versions

- Transformers 4.30.0.dev0
- Pytorch 2.0.1
- Datasets 2.12.1.dev0
- Tokenizers 0.13.3