jimregan commited on
Commit
52c67c1
1 Parent(s): fa8b155
checkpoints/checkpoint_epoch=099.ckpt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:3ba3661ef19717291a05d8a473ee3c4e1a6ff19c6335dfa4131f3009b6d135d7
3
- size 218839178
 
 
 
 
checkpoints/checkpoint_epoch=199.ckpt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:e6f7593be80cb623c222e630ad6375dfecd37f8fb757eae8f06e187de96ddbdd
3
- size 218839561
 
 
 
 
checkpoints/checkpoint_epoch=299.ckpt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:2c5dc3f51e058c3a21137aab9ace1846644d6e87f5076cc03dfc6a1ea4030f4e
3
- size 218839944
 
 
 
 
checkpoints/checkpoint_epoch=399.ckpt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:1151cd7c9ce36258a5e8f125471dd33f32ffddcee137bf11f14b052cd1089fb6
3
- size 218840327
 
 
 
 
checkpoints/checkpoint_epoch=499.ckpt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:802947c447968f893cfafb4c3361c8641c671f665225ef7fc867cf18b1f2c16f
3
- size 218840710
 
 
 
 
checkpoints/checkpoint_epoch=599.ckpt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:fed2174b0a971098edb110d40815a0013e9c29ab119f7d3cde9cdfb8853d1a61
3
- size 218841029
 
 
 
 
checkpoints/checkpoint_epoch=699.ckpt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:17a076f2cc33d47a6707c19aba5587c693b878ebb0e0e3d17988f083f9e6d8e2
3
- size 218841412
 
 
 
 
checkpoints/checkpoint_epoch=799.ckpt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:69d6ac3da37a1e63557ad4dd50e85fe9132d15a2dbe3df705c10f26b0403f488
3
- size 218841795
 
 
 
 
checkpoints/checkpoint_epoch=899.ckpt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:b68cd1ce3255316a71119ba608db65cce3a3dba1d401ec000ad685025c962720
3
- size 218842178
 
 
 
 
tensorboard/events.out.tfevents.1729366325.0959f84f9bd7.22.0 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:d3d60c60b7a379fd0664bf83cf906a2798468a6f2d274507a84e327333be1301
3
- size 192900735
 
 
 
 
tensorboard/hparams.yaml DELETED
@@ -1,115 +0,0 @@
1
- model:
2
- _target_: matcha.models.matcha_tts.MatchaTTS
3
- n_vocab: 178
4
- n_spks: ${data.n_spks}
5
- spk_emb_dim: 64
6
- n_feats: 80
7
- data_statistics: ${data.data_statistics}
8
- out_size: null
9
- prior_loss: true
10
- use_precomputed_durations: ${data.load_durations}
11
- encoder:
12
- encoder_type: RoPE Encoder
13
- encoder_params:
14
- n_feats: ${model.n_feats}
15
- n_channels: 192
16
- filter_channels: 768
17
- filter_channels_dp: 256
18
- n_heads: 2
19
- n_layers: 6
20
- kernel_size: 3
21
- p_dropout: 0.1
22
- spk_emb_dim: 64
23
- n_spks: 1
24
- prenet: true
25
- duration_predictor_params:
26
- filter_channels_dp: ${model.encoder.encoder_params.filter_channels_dp}
27
- kernel_size: 3
28
- p_dropout: ${model.encoder.encoder_params.p_dropout}
29
- decoder:
30
- channels:
31
- - 256
32
- - 256
33
- dropout: 0.05
34
- attention_head_dim: 64
35
- n_blocks: 1
36
- num_mid_blocks: 2
37
- num_heads: 2
38
- act_fn: snakebeta
39
- cfm:
40
- name: CFM
41
- solver: euler
42
- sigma_min: 0.0001
43
- optimizer:
44
- _target_: torch.optim.Adam
45
- _partial_: true
46
- lr: 0.0001
47
- weight_decay: 0.0
48
- model/params/total: 18204193
49
- model/params/trainable: 18204193
50
- model/params/non_trainable: 0
51
- data:
52
- _target_: matcha.data.text_mel_datamodule.TextMelDataModule
53
- name: ljspeech
54
- train_filelist_path: data/LJSpeech-1.1/train.txt
55
- valid_filelist_path: data/LJSpeech-1.1/val.txt
56
- batch_size: 32
57
- num_workers: 20
58
- pin_memory: true
59
- cleaners:
60
- - english_cleaners2
61
- add_blank: true
62
- n_spks: 1
63
- n_fft: 1024
64
- n_feats: 80
65
- sample_rate: 22050
66
- hop_length: 256
67
- win_length: 1024
68
- f_min: 0
69
- f_max: 8000
70
- data_statistics:
71
- mel_mean: -5.536622
72
- mel_std: 2.116101
73
- seed: ${seed}
74
- load_durations: false
75
- trainer:
76
- _target_: lightning.pytorch.trainer.Trainer
77
- default_root_dir: ${paths.output_dir}
78
- max_epochs: -1
79
- accelerator: gpu
80
- devices:
81
- - 0
82
- precision: 16-mixed
83
- check_val_every_n_epoch: 1
84
- deterministic: false
85
- gradient_clip_val: 5.0
86
- callbacks:
87
- model_checkpoint:
88
- _target_: lightning.pytorch.callbacks.ModelCheckpoint
89
- dirpath: ${paths.output_dir}/checkpoints
90
- filename: checkpoint_{epoch:03d}
91
- monitor: epoch
92
- verbose: false
93
- save_last: true
94
- save_top_k: 10
95
- mode: max
96
- auto_insert_metric_name: true
97
- save_weights_only: false
98
- every_n_train_steps: null
99
- train_time_interval: null
100
- every_n_epochs: 100
101
- save_on_train_epoch_end: null
102
- model_summary:
103
- _target_: lightning.pytorch.callbacks.RichModelSummary
104
- max_depth: 3
105
- rich_progress_bar:
106
- _target_: lightning.pytorch.callbacks.RichProgressBar
107
- extras:
108
- ignore_warnings: false
109
- enforce_tags: true
110
- print_config: true
111
- task_name: train
112
- tags:
113
- - ljspeech
114
- ckpt_path: null
115
- seed: 1234