Marianoleiras commited on
Commit
2d43abc
·
verified ·
1 Parent(s): a8f457c

End of training

Browse files
README.md CHANGED
@@ -2,6 +2,8 @@
2
  library_name: transformers
3
  tags:
4
  - generated_from_trainer
 
 
5
  model-index:
6
  - name: whisper-small-es-ja
7
  results: []
@@ -14,12 +16,8 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  This model was trained from scratch on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
- - eval_loss: 8.4164
18
- - eval_bleu: 21.5256
19
- - eval_runtime: 210.1767
20
- - eval_samples_per_second: 6.499
21
- - eval_steps_per_second: 0.409
22
- - step: 0
23
 
24
  ## Model description
25
 
@@ -39,14 +37,35 @@ More information needed
39
 
40
  The following hyperparameters were used during training:
41
  - learning_rate: 1e-05
42
- - train_batch_size: 32
43
- - eval_batch_size: 16
44
  - seed: 42
 
45
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
46
  - lr_scheduler_type: linear
47
- - training_steps: 5000
48
  - mixed_precision_training: Native AMP
49
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
50
  ### Framework versions
51
 
52
  - Transformers 4.47.1
 
2
  library_name: transformers
3
  tags:
4
  - generated_from_trainer
5
+ metrics:
6
+ - bleu
7
  model-index:
8
  - name: whisper-small-es-ja
9
  results: []
 
16
 
17
  This model was trained from scratch on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
+ - Loss: 1.1810
20
+ - Bleu: 21.4112
 
 
 
 
21
 
22
  ## Model description
23
 
 
37
 
38
  The following hyperparameters were used during training:
39
  - learning_rate: 1e-05
40
+ - train_batch_size: 16
41
+ - eval_batch_size: 8
42
  - seed: 42
43
+ - distributed_type: multi-GPU
44
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
45
  - lr_scheduler_type: linear
46
+ - training_steps: 3500
47
  - mixed_precision_training: Native AMP
48
 
49
+ ### Training results
50
+
51
+ | Training Loss | Epoch | Step | Bleu | Validation Loss |
52
+ |:-------------:|:------:|:----:|:-------:|:---------------:|
53
+ | 1.5787 | 0.3962 | 250 | 11.6756 | 1.5196 |
54
+ | 1.3535 | 0.7924 | 500 | 16.0514 | 1.3470 |
55
+ | 1.0658 | 1.1886 | 750 | 17.7743 | 1.2533 |
56
+ | 1.0303 | 1.5848 | 1000 | 19.1894 | 1.2046 |
57
+ | 0.9893 | 1.9810 | 1250 | 20.1198 | 1.1591 |
58
+ | 0.7569 | 2.3772 | 1500 | 21.0054 | 1.1546 |
59
+ | 0.7571 | 2.7734 | 1750 | 21.6425 | 1.1378 |
60
+ | 0.5557 | 3.1696 | 2000 | 21.7563 | 1.1500 |
61
+ | 0.5612 | 3.5658 | 2250 | 21.1391 | 1.1395 |
62
+ | 0.5581 | 3.9620 | 2500 | 22.0412 | 1.1343 |
63
+ | 0.4144 | 4.3582 | 2750 | 22.2850 | 1.1724 |
64
+ | 0.4114 | 4.7544 | 3000 | 22.1925 | 1.1681 |
65
+ | 0.3005 | 5.1506 | 3250 | 21.4948 | 1.1947 |
66
+ | 0.2945 | 5.5468 | 3500 | 1.1921 | 22.1454 |
67
+
68
+
69
  ### Framework versions
70
 
71
  - Transformers 4.47.1
generation_config.json CHANGED
@@ -160,7 +160,7 @@
160
  "<|yo|>": 50325,
161
  "<|zh|>": 50260
162
  },
163
- "language": "japanese",
164
  "max_initial_timestamp_index": 50,
165
  "max_length": 448,
166
  "no_timestamps_token_id": 50363,
 
160
  "<|yo|>": 50325,
161
  "<|zh|>": 50260
162
  },
163
+ "language": "ja",
164
  "max_initial_timestamp_index": 50,
165
  "max_length": 448,
166
  "no_timestamps_token_id": 50363,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a518d500e70b29ab45c2dbb61c320aadcc5941da8cce399dc560d12d0c17f78f
3
  size 966995080
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f13cb263d1611926ce983abf794049bede8cbc4e40923d0e9a5370e9f9dcf86
3
  size 966995080
runs/Dec19_17-50-54_mad-loc-analytics3/events.out.tfevents.1734627751.mad-loc-analytics3.1007866.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:853bb9b0248d61f2cedca2f184e25e3c5b471078c05fa9da1326dc1188f3f683
3
+ size 407