Edmon02 commited on
Commit
8a102c2
·
verified ·
1 Parent(s): 80b1454

End of training

Browse files
README.md CHANGED
@@ -2,8 +2,6 @@
2
  base_model: Edmon02/speecht5_finetuned_voxpopuli_nl
3
  tags:
4
  - generated_from_trainer
5
- datasets:
6
- - hyvoxpopuli
7
  model-index:
8
  - name: speecht5_finetuned_voxpopuli_hy
9
  results: []
@@ -14,9 +12,9 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  # speecht5_finetuned_voxpopuli_hy
16
 
17
- This model is a fine-tuned version of [Edmon02/speecht5_finetuned_voxpopuli_nl](https://huggingface.co/Edmon02/speecht5_finetuned_voxpopuli_nl) on the hyvoxpopuli dataset.
18
  It achieves the following results on the evaluation set:
19
- - Loss: 0.5655
20
 
21
  ## Model description
22
 
@@ -43,23 +41,22 @@ The following hyperparameters were used during training:
43
  - total_train_batch_size: 32
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
- - lr_scheduler_warmup_steps: 500
47
- - training_steps: 4000
48
- - mixed_precision_training: Native AMP
49
 
50
  ### Training results
51
 
52
- | Training Loss | Epoch | Step | Validation Loss |
53
- |:-------------:|:---------:|:----:|:---------------:|
54
- | 0.5144 | 296.2963 | 1000 | 0.5548 |
55
- | 0.4797 | 592.5926 | 2000 | 0.5446 |
56
- | 0.4674 | 888.8889 | 3000 | 0.5519 |
57
- | 0.4575 | 1185.1852 | 4000 | 0.5655 |
58
 
59
 
60
  ### Framework versions
61
 
62
- - Transformers 4.41.2
63
- - Pytorch 2.3.0+cu121
64
  - Datasets 2.20.0
65
  - Tokenizers 0.19.1
 
2
  base_model: Edmon02/speecht5_finetuned_voxpopuli_nl
3
  tags:
4
  - generated_from_trainer
 
 
5
  model-index:
6
  - name: speecht5_finetuned_voxpopuli_hy
7
  results: []
 
12
 
13
  # speecht5_finetuned_voxpopuli_hy
14
 
15
+ This model is a fine-tuned version of [Edmon02/speecht5_finetuned_voxpopuli_nl](https://huggingface.co/Edmon02/speecht5_finetuned_voxpopuli_nl) on the None dataset.
16
  It achieves the following results on the evaluation set:
17
+ - Loss: 0.6466
18
 
19
  ## Model description
20
 
 
41
  - total_train_batch_size: 32
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: linear
44
+ - lr_scheduler_warmup_steps: 125
45
+ - training_steps: 1000
 
46
 
47
  ### Training results
48
 
49
+ | Training Loss | Epoch | Step | Validation Loss |
50
+ |:-------------:|:------:|:----:|:---------------:|
51
+ | 0.7114 | 2.4600 | 250 | 0.6624 |
52
+ | 0.6999 | 4.9200 | 500 | 0.6542 |
53
+ | 0.7079 | 7.3801 | 750 | 0.6487 |
54
+ | 0.6972 | 9.8401 | 1000 | 0.6466 |
55
 
56
 
57
  ### Framework versions
58
 
59
+ - Transformers 4.42.4
60
+ - Pytorch 2.3.1+cu121
61
  - Datasets 2.20.0
62
  - Tokenizers 0.19.1
added_tokens.json CHANGED
@@ -1,4 +1,3 @@
1
  {
2
- "<ctc_blank>": 80,
3
- "<mask>": 79
4
  }
 
1
  {
2
+ "<pad>": 157
 
3
  }
config.json CHANGED
@@ -65,8 +65,8 @@
65
  "mask_time_min_masks": 2,
66
  "mask_time_prob": 0.05,
67
  "max_length": 1876,
68
- "max_speech_positions": 1876,
69
- "max_text_positions": 600,
70
  "model_type": "speecht5",
71
  "num_conv_pos_embedding_groups": 16,
72
  "num_conv_pos_embeddings": 128,
@@ -85,8 +85,8 @@
85
  "speech_decoder_prenet_layers": 2,
86
  "speech_decoder_prenet_units": 256,
87
  "torch_dtype": "float32",
88
- "transformers_version": "4.41.2",
89
  "use_cache": false,
90
  "use_guided_attention_loss": true,
91
- "vocab_size": 81
92
  }
 
65
  "mask_time_min_masks": 2,
66
  "mask_time_prob": 0.05,
67
  "max_length": 1876,
68
+ "max_speech_positions": 5000,
69
+ "max_text_positions": 512,
70
  "model_type": "speecht5",
71
  "num_conv_pos_embedding_groups": 16,
72
  "num_conv_pos_embeddings": 128,
 
85
  "speech_decoder_prenet_layers": 2,
86
  "speech_decoder_prenet_units": 256,
87
  "torch_dtype": "float32",
88
+ "transformers_version": "4.42.4",
89
  "use_cache": false,
90
  "use_guided_attention_loss": true,
91
+ "vocab_size": 158
92
  }
generation_config.json CHANGED
@@ -5,5 +5,5 @@
5
  "eos_token_id": 2,
6
  "max_length": 1876,
7
  "pad_token_id": 1,
8
- "transformers_version": "4.41.2"
9
  }
 
5
  "eos_token_id": 2,
6
  "max_length": 1876,
7
  "pad_token_id": 1,
8
+ "transformers_version": "4.42.4"
9
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8a9e46e706ed75385aa87712c5521d64a17e1892d4fd63a1d07144e3326cfaa3
3
- size 577789320
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9bcf0a24f4e638b8747787afb342716c35f1d34be8e4ad482af745987e20edfc
3
+ size 578025864
runs/Jul27_10-54-15_caa53b8840ec/events.out.tfevents.1722077676.caa53b8840ec.310.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd5ef7b517d7c76867a8830882d101b412770f06b2b7b024322bac4737861ce0
3
+ size 42741
runs/Jul27_11-53-20_caa53b8840ec/events.out.tfevents.1722081204.caa53b8840ec.16745.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7b574ee8c843d3ab268ba71ce8127660ba0d0e240134c50d902097657ffe147
3
+ size 42909
special_tokens_map.json CHANGED
@@ -1,13 +1,6 @@
1
  {
2
  "bos_token": "<s>",
3
  "eos_token": "</s>",
4
- "mask_token": {
5
- "content": "<mask>",
6
- "lstrip": true,
7
- "normalized": true,
8
- "rstrip": false,
9
- "single_word": false
10
- },
11
  "pad_token": "<pad>",
12
  "unk_token": "<unk>"
13
  }
 
1
  {
2
  "bos_token": "<s>",
3
  "eos_token": "</s>",
 
 
 
 
 
 
 
4
  "pad_token": "<pad>",
5
  "unk_token": "<unk>"
6
  }
spm_char.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7fcc48f3e225f627b1641db410ceb0c8649bd2b0c982e150b03f8be3728ab560
3
- size 238473
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5f1ed236bbb46e584fad7dc1e93e7dcae275eda86e53ed80365bf5531ebc41b
3
+ size 239298
tokenizer_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "added_tokens_decoder": {
3
  "0": {
4
- "content": "<s>",
5
  "lstrip": false,
6
  "normalized": false,
7
  "rstrip": false,
@@ -9,7 +9,7 @@
9
  "special": true
10
  },
11
  "1": {
12
- "content": "<pad>",
13
  "lstrip": false,
14
  "normalized": false,
15
  "rstrip": false,
@@ -24,36 +24,19 @@
24
  "single_word": false,
25
  "special": true
26
  },
27
- "3": {
28
- "content": "<unk>",
29
  "lstrip": false,
30
  "normalized": false,
31
  "rstrip": false,
32
  "single_word": false,
33
  "special": true
34
- },
35
- "79": {
36
- "content": "<mask>",
37
- "lstrip": true,
38
- "normalized": true,
39
- "rstrip": false,
40
- "single_word": false,
41
- "special": true
42
- },
43
- "80": {
44
- "content": "<ctc_blank>",
45
- "lstrip": false,
46
- "normalized": true,
47
- "rstrip": false,
48
- "single_word": false,
49
- "special": false
50
  }
51
  },
52
  "bos_token": "<s>",
53
  "clean_up_tokenization_spaces": true,
54
  "eos_token": "</s>",
55
- "mask_token": "<mask>",
56
- "model_max_length": 600,
57
  "normalize": false,
58
  "pad_token": "<pad>",
59
  "processor_class": "SpeechT5Processor",
 
1
  {
2
  "added_tokens_decoder": {
3
  "0": {
4
+ "content": "<unk>",
5
  "lstrip": false,
6
  "normalized": false,
7
  "rstrip": false,
 
9
  "special": true
10
  },
11
  "1": {
12
+ "content": "<s>",
13
  "lstrip": false,
14
  "normalized": false,
15
  "rstrip": false,
 
24
  "single_word": false,
25
  "special": true
26
  },
27
+ "157": {
28
+ "content": "<pad>",
29
  "lstrip": false,
30
  "normalized": false,
31
  "rstrip": false,
32
  "single_word": false,
33
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
34
  }
35
  },
36
  "bos_token": "<s>",
37
  "clean_up_tokenization_spaces": true,
38
  "eos_token": "</s>",
39
+ "model_max_length": 1000000000000000019884624838656,
 
40
  "normalize": false,
41
  "pad_token": "<pad>",
42
  "processor_class": "SpeechT5Processor",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:923a08646dc770d432227f22698377c2adcfd47e3021a435b4577c2c85fda902
3
  size 5304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a12c0040b3a2d22c02d61dbedce1862573e49552d1b47c121c84008c2c0b45d
3
  size 5304