binhnx8 commited on
Commit
42b3e14
1 Parent(s): 32318f6

Upload folder using huggingface_hub

Browse files
melody_v1/.hydra/config.yaml ADDED
@@ -0,0 +1,114 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ paths:
2
+ run_dir: results/${project}
3
+ ckpt_dir: ${paths.run_dir}/checkpoints
4
+ trainer:
5
+ _target_: lightning.pytorch.trainer.Trainer
6
+ default_root_dir: ${paths.run_dir}
7
+ accelerator: gpu
8
+ num_nodes: 1
9
+ devices: auto
10
+ strategy:
11
+ _target_: lightning.pytorch.strategies.DDPStrategy
12
+ process_group_backend: nccl
13
+ precision: bf16-true
14
+ check_val_every_n_epoch: null
15
+ val_check_interval: 1000
16
+ max_steps: 100000
17
+ benchmark: true
18
+ accumulate_grad_batches: 1
19
+ gradient_clip_val: 1.0
20
+ gradient_clip_algorithm: norm
21
+ limit_val_batches: 10
22
+ callbacks:
23
+ model_checkpoint:
24
+ _target_: lightning.pytorch.callbacks.ModelCheckpoint
25
+ dirpath: ${paths.ckpt_dir}
26
+ filename: step_{step:09d}
27
+ save_last: false
28
+ save_top_k: 3
29
+ monitor: step
30
+ mode: max
31
+ every_n_epochs: null
32
+ every_n_train_steps: ${trainer.val_check_interval}
33
+ auto_insert_metric_name: false
34
+ model_summary:
35
+ _target_: lightning.pytorch.callbacks.ModelSummary
36
+ max_depth: 2
37
+ learning_rate_monitor:
38
+ _target_: lightning.pytorch.callbacks.LearningRateMonitor
39
+ logging_interval: step
40
+ log_momentum: false
41
+ grad_norm_monitor:
42
+ _target_: fish_speech.callbacks.GradNormMonitor
43
+ norm_type: 2
44
+ logging_interval: step
45
+ logger:
46
+ tensorboard:
47
+ _target_: lightning.pytorch.loggers.tensorboard.TensorBoardLogger
48
+ save_dir: ${paths.run_dir}/tensorboard/
49
+ name: null
50
+ log_graph: false
51
+ default_hp_metric: true
52
+ prefix: ''
53
+ train: true
54
+ test: false
55
+ project: melody_v1
56
+ max_length: 1024
57
+ pretrained_ckpt_path: checkpoints/fish-speech-1.2
58
+ tokenizer:
59
+ _target_: transformers.AutoTokenizer.from_pretrained
60
+ pretrained_model_name_or_path: ${pretrained_ckpt_path}
61
+ train_dataset:
62
+ _target_: fish_speech.datasets.semantic.AutoTextSemanticInstructionDataset
63
+ proto_files:
64
+ - data/protos
65
+ tokenizer: ${tokenizer}
66
+ causal: true
67
+ max_length: ${max_length}
68
+ use_speaker: false
69
+ interactive_prob: 0.7
70
+ val_dataset:
71
+ _target_: fish_speech.datasets.semantic.AutoTextSemanticInstructionDataset
72
+ proto_files:
73
+ - data/protos
74
+ tokenizer: ${tokenizer}
75
+ causal: true
76
+ max_length: ${max_length}
77
+ use_speaker: false
78
+ interactive_prob: 0.7
79
+ data:
80
+ _target_: fish_speech.datasets.semantic.SemanticDataModule
81
+ train_dataset: ${train_dataset}
82
+ val_dataset: ${val_dataset}
83
+ num_workers: 4
84
+ batch_size: 4
85
+ tokenizer: ${tokenizer}
86
+ max_length: ${max_length}
87
+ model:
88
+ _target_: fish_speech.models.text2semantic.lit_module.TextToSemantic
89
+ model:
90
+ _target_: fish_speech.models.text2semantic.llama.BaseTransformer.from_pretrained
91
+ path: ${pretrained_ckpt_path}
92
+ load_weights: true
93
+ max_length: ${max_length}
94
+ lora_config:
95
+ _target_: fish_speech.models.text2semantic.lora.LoraConfig
96
+ r: 8
97
+ lora_alpha: 16
98
+ lora_dropout: 0.01
99
+ optimizer:
100
+ _target_: torch.optim.AdamW
101
+ _partial_: true
102
+ lr: 0.0001
103
+ weight_decay: 0.01
104
+ betas:
105
+ - 0.9
106
+ - 0.95
107
+ eps: 1.0e-05
108
+ lr_scheduler:
109
+ _target_: torch.optim.lr_scheduler.LambdaLR
110
+ _partial_: true
111
+ lr_lambda:
112
+ _target_: fish_speech.scheduler.get_constant_schedule_with_warmup_lr_lambda
113
+ _partial_: true
114
+ num_warmup_steps: 50
melody_v1/.hydra/hydra.yaml ADDED
@@ -0,0 +1,157 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: ${paths.run_dir}
4
+ sweep:
5
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - project=melody_v1
116
+ - +lora@model.model.lora_config=r_8_alpha_16
117
+ job:
118
+ name: train
119
+ chdir: null
120
+ override_dirname: +lora@model.model.lora_config=r_8_alpha_16,project=melody_v1
121
+ id: ???
122
+ num: ???
123
+ config_name: text2semantic_finetune
124
+ env_set: {}
125
+ env_copy: []
126
+ config:
127
+ override_dirname:
128
+ kv_sep: '='
129
+ item_sep: ','
130
+ exclude_keys: []
131
+ runtime:
132
+ version: 1.3.2
133
+ version_base: '1.3'
134
+ cwd: /mnt/f/repo_fish-speech
135
+ config_sources:
136
+ - path: hydra.conf
137
+ schema: pkg
138
+ provider: hydra
139
+ - path: /mnt/f/repo_fish-speech/fish_speech/configs
140
+ schema: file
141
+ provider: main
142
+ - path: ''
143
+ schema: structured
144
+ provider: schema
145
+ output_dir: /mnt/f/repo_fish-speech/results/melody_v1
146
+ choices:
147
+ lora@model.model.lora_config: r_8_alpha_16
148
+ hydra/env: default
149
+ hydra/callbacks: null
150
+ hydra/job_logging: default
151
+ hydra/hydra_logging: default
152
+ hydra/hydra_help: default
153
+ hydra/help: default
154
+ hydra/sweeper: basic
155
+ hydra/launcher: basic
156
+ hydra/output: default
157
+ verbose: false
melody_v1/.hydra/overrides.yaml ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ - project=melody_v1
2
+ - +lora@model.model.lora_config=r_8_alpha_16
melody_v1/checkpoints/step_000091000.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a96c5e9e8e17fc07a51f9922240e0f4e86aebd7efde7d6e59837ddf8fae5b0e
3
+ size 30476671
melody_v1/checkpoints/step_000092000.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d219a4cef3a3d62a4d771b0bf212b150aad1d1995d5ef71dae5da31af0745e5
3
+ size 30476671
melody_v1/checkpoints/step_000093000.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58a3c281927aa33a97ad5fffadd802eba83649aceb18fad12c50dad4e3ffd0fd
3
+ size 30476671
melody_v1/tensorboard/version_0/events.out.tfevents.1720798951.DESKTOP-E8VFO84.111147.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24bd05df1f1566a96147eef5b563a272b6dae3830e7fdbd43e6775ca08a16f23
3
+ size 831279
melody_v1/tensorboard/version_0/hparams.yaml ADDED
@@ -0,0 +1,107 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ _target_: fish_speech.models.text2semantic.lit_module.TextToSemantic
3
+ model:
4
+ _target_: fish_speech.models.text2semantic.llama.BaseTransformer.from_pretrained
5
+ path: checkpoints/fish-speech-1.2
6
+ load_weights: true
7
+ max_length: 1024
8
+ lora_config:
9
+ _target_: fish_speech.models.text2semantic.lora.LoraConfig
10
+ r: 8
11
+ lora_alpha: 16
12
+ lora_dropout: 0.01
13
+ optimizer:
14
+ _target_: torch.optim.AdamW
15
+ _partial_: true
16
+ lr: 0.0001
17
+ weight_decay: 0.01
18
+ betas:
19
+ - 0.9
20
+ - 0.95
21
+ eps: 1.0e-05
22
+ lr_scheduler:
23
+ _target_: torch.optim.lr_scheduler.LambdaLR
24
+ _partial_: true
25
+ lr_lambda:
26
+ _target_: fish_speech.scheduler.get_constant_schedule_with_warmup_lr_lambda
27
+ _partial_: true
28
+ num_warmup_steps: 50
29
+ model/params/total: 495286272
30
+ model/params/trainable: 5017600
31
+ model/params/non_trainable: 490268672
32
+ data:
33
+ _target_: fish_speech.datasets.semantic.SemanticDataModule
34
+ train_dataset:
35
+ _target_: fish_speech.datasets.semantic.AutoTextSemanticInstructionDataset
36
+ proto_files:
37
+ - data/protos
38
+ tokenizer:
39
+ _target_: transformers.AutoTokenizer.from_pretrained
40
+ pretrained_model_name_or_path: checkpoints/fish-speech-1.2
41
+ causal: true
42
+ max_length: 1024
43
+ use_speaker: false
44
+ interactive_prob: 0.7
45
+ val_dataset:
46
+ _target_: fish_speech.datasets.semantic.AutoTextSemanticInstructionDataset
47
+ proto_files:
48
+ - data/protos
49
+ tokenizer:
50
+ _target_: transformers.AutoTokenizer.from_pretrained
51
+ pretrained_model_name_or_path: checkpoints/fish-speech-1.2
52
+ causal: true
53
+ max_length: 1024
54
+ use_speaker: false
55
+ interactive_prob: 0.7
56
+ num_workers: 4
57
+ batch_size: 4
58
+ tokenizer:
59
+ _target_: transformers.AutoTokenizer.from_pretrained
60
+ pretrained_model_name_or_path: checkpoints/fish-speech-1.2
61
+ max_length: 1024
62
+ trainer:
63
+ _target_: lightning.pytorch.trainer.Trainer
64
+ default_root_dir: results/melody_v1
65
+ accelerator: gpu
66
+ num_nodes: 1
67
+ devices: auto
68
+ strategy:
69
+ _target_: lightning.pytorch.strategies.DDPStrategy
70
+ process_group_backend: nccl
71
+ precision: bf16-true
72
+ check_val_every_n_epoch: null
73
+ val_check_interval: 1000
74
+ max_steps: 100000
75
+ benchmark: true
76
+ accumulate_grad_batches: 1
77
+ gradient_clip_val: 1.0
78
+ gradient_clip_algorithm: norm
79
+ limit_val_batches: 10
80
+ callbacks:
81
+ model_checkpoint:
82
+ _target_: lightning.pytorch.callbacks.ModelCheckpoint
83
+ dirpath: results/melody_v1/checkpoints
84
+ filename: step_{step:09d}
85
+ save_last: false
86
+ save_top_k: 3
87
+ monitor: step
88
+ mode: max
89
+ every_n_epochs: null
90
+ every_n_train_steps: 1000
91
+ auto_insert_metric_name: false
92
+ model_summary:
93
+ _target_: lightning.pytorch.callbacks.ModelSummary
94
+ max_depth: 2
95
+ learning_rate_monitor:
96
+ _target_: lightning.pytorch.callbacks.LearningRateMonitor
97
+ logging_interval: step
98
+ log_momentum: false
99
+ grad_norm_monitor:
100
+ _target_: fish_speech.callbacks.GradNormMonitor
101
+ norm_type: 2
102
+ logging_interval: step
103
+ extras: null
104
+ task_name: null
105
+ tags: null
106
+ ckpt_path: null
107
+ seed: null
melody_v1/train.log ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2024-07-12 22:42:24,744][__main__][INFO] - [rank: 0] Instantiating datamodule <fish_speech.datasets.semantic.SemanticDataModule>
2
+ [2024-07-12 22:42:25,360][datasets][INFO] - PyTorch version 2.3.1 available.
3
+ [2024-07-12 22:42:25,663][__main__][INFO] - [rank: 0] Instantiating model <fish_speech.models.text2semantic.lit_module.TextToSemantic>
4
+ [2024-07-12 22:42:25,719][fish_speech.models.text2semantic.llama][INFO] - [rank: 0] Override max_seq_len to 1024
5
+ [2024-07-12 22:42:25,755][fish_speech.models.text2semantic.llama][INFO] - [rank: 0] Loading model from checkpoints/fish-speech-1.2, config: DualARModelArgs(model_type='dual_ar', vocab_size=32000, n_layer=24, n_head=16, dim=1024, intermediate_size=4096, n_local_heads=2, head_dim=64, rope_base=1000000.0, norm_eps=1e-06, max_seq_len=1024, dropout=0.1, tie_word_embeddings=False, attention_qkv_bias=False, codebook_size=1024, num_codebooks=4, use_gradient_checkpointing=True, initializer_range=0.02, n_fast_layer=4)
6
+ [2024-07-12 22:42:31,758][fish_speech.models.text2semantic.llama][INFO] - [rank: 0] LoRA setup: LoraConfig(r=8, lora_alpha=16, lora_dropout=0.01)
7
+ [2024-07-12 22:42:31,832][fish_speech.models.text2semantic.llama][INFO] - [rank: 0] Loaded weights with error: _IncompatibleKeys(missing_keys=['embeddings.lora_A', 'embeddings.lora_B', 'codebook_embeddings.lora_A', 'codebook_embeddings.lora_B', 'layers.0.attention.wqkv.lora_A', 'layers.0.attention.wqkv.lora_B', 'layers.0.attention.wo.lora_A', 'layers.0.attention.wo.lora_B', 'layers.0.feed_forward.w1.lora_A', 'layers.0.feed_forward.w1.lora_B', 'layers.0.feed_forward.w3.lora_A', 'layers.0.feed_forward.w3.lora_B', 'layers.0.feed_forward.w2.lora_A', 'layers.0.feed_forward.w2.lora_B', 'layers.1.attention.wqkv.lora_A', 'layers.1.attention.wqkv.lora_B', 'layers.1.attention.wo.lora_A', 'layers.1.attention.wo.lora_B', 'layers.1.feed_forward.w1.lora_A', 'layers.1.feed_forward.w1.lora_B', 'layers.1.feed_forward.w3.lora_A', 'layers.1.feed_forward.w3.lora_B', 'layers.1.feed_forward.w2.lora_A', 'layers.1.feed_forward.w2.lora_B', 'layers.2.attention.wqkv.lora_A', 'layers.2.attention.wqkv.lora_B', 'layers.2.attention.wo.lora_A', 'layers.2.attention.wo.lora_B', 'layers.2.feed_forward.w1.lora_A', 'layers.2.feed_forward.w1.lora_B', 'layers.2.feed_forward.w3.lora_A', 'layers.2.feed_forward.w3.lora_B', 'layers.2.feed_forward.w2.lora_A', 'layers.2.feed_forward.w2.lora_B', 'layers.3.attention.wqkv.lora_A', 'layers.3.attention.wqkv.lora_B', 'layers.3.attention.wo.lora_A', 'layers.3.attention.wo.lora_B', 'layers.3.feed_forward.w1.lora_A', 'layers.3.feed_forward.w1.lora_B', 'layers.3.feed_forward.w3.lora_A', 'layers.3.feed_forward.w3.lora_B', 'layers.3.feed_forward.w2.lora_A', 'layers.3.feed_forward.w2.lora_B', 'layers.4.attention.wqkv.lora_A', 'layers.4.attention.wqkv.lora_B', 'layers.4.attention.wo.lora_A', 'layers.4.attention.wo.lora_B', 'layers.4.feed_forward.w1.lora_A', 'layers.4.feed_forward.w1.lora_B', 'layers.4.feed_forward.w3.lora_A', 'layers.4.feed_forward.w3.lora_B', 'layers.4.feed_forward.w2.lora_A', 'layers.4.feed_forward.w2.lora_B', 'layers.5.attention.wqkv.lora_A', 'layers.5.attention.wqkv.lora_B', 'layers.5.attention.wo.lora_A', 'layers.5.attention.wo.lora_B', 'layers.5.feed_forward.w1.lora_A', 'layers.5.feed_forward.w1.lora_B', 'layers.5.feed_forward.w3.lora_A', 'layers.5.feed_forward.w3.lora_B', 'layers.5.feed_forward.w2.lora_A', 'layers.5.feed_forward.w2.lora_B', 'layers.6.attention.wqkv.lora_A', 'layers.6.attention.wqkv.lora_B', 'layers.6.attention.wo.lora_A', 'layers.6.attention.wo.lora_B', 'layers.6.feed_forward.w1.lora_A', 'layers.6.feed_forward.w1.lora_B', 'layers.6.feed_forward.w3.lora_A', 'layers.6.feed_forward.w3.lora_B', 'layers.6.feed_forward.w2.lora_A', 'layers.6.feed_forward.w2.lora_B', 'layers.7.attention.wqkv.lora_A', 'layers.7.attention.wqkv.lora_B', 'layers.7.attention.wo.lora_A', 'layers.7.attention.wo.lora_B', 'layers.7.feed_forward.w1.lora_A', 'layers.7.feed_forward.w1.lora_B', 'layers.7.feed_forward.w3.lora_A', 'layers.7.feed_forward.w3.lora_B', 'layers.7.feed_forward.w2.lora_A', 'layers.7.feed_forward.w2.lora_B', 'layers.8.attention.wqkv.lora_A', 'layers.8.attention.wqkv.lora_B', 'layers.8.attention.wo.lora_A', 'layers.8.attention.wo.lora_B', 'layers.8.feed_forward.w1.lora_A', 'layers.8.feed_forward.w1.lora_B', 'layers.8.feed_forward.w3.lora_A', 'layers.8.feed_forward.w3.lora_B', 'layers.8.feed_forward.w2.lora_A', 'layers.8.feed_forward.w2.lora_B', 'layers.9.attention.wqkv.lora_A', 'layers.9.attention.wqkv.lora_B', 'layers.9.attention.wo.lora_A', 'layers.9.attention.wo.lora_B', 'layers.9.feed_forward.w1.lora_A', 'layers.9.feed_forward.w1.lora_B', 'layers.9.feed_forward.w3.lora_A', 'layers.9.feed_forward.w3.lora_B', 'layers.9.feed_forward.w2.lora_A', 'layers.9.feed_forward.w2.lora_B', 'layers.10.attention.wqkv.lora_A', 'layers.10.attention.wqkv.lora_B', 'layers.10.attention.wo.lora_A', 'layers.10.attention.wo.lora_B', 'layers.10.feed_forward.w1.lora_A', 'layers.10.feed_forward.w1.lora_B', 'layers.10.feed_forward.w3.lora_A', 'layers.10.feed_forward.w3.lora_B', 'layers.10.feed_forward.w2.lora_A', 'layers.10.feed_forward.w2.lora_B', 'layers.11.attention.wqkv.lora_A', 'layers.11.attention.wqkv.lora_B', 'layers.11.attention.wo.lora_A', 'layers.11.attention.wo.lora_B', 'layers.11.feed_forward.w1.lora_A', 'layers.11.feed_forward.w1.lora_B', 'layers.11.feed_forward.w3.lora_A', 'layers.11.feed_forward.w3.lora_B', 'layers.11.feed_forward.w2.lora_A', 'layers.11.feed_forward.w2.lora_B', 'layers.12.attention.wqkv.lora_A', 'layers.12.attention.wqkv.lora_B', 'layers.12.attention.wo.lora_A', 'layers.12.attention.wo.lora_B', 'layers.12.feed_forward.w1.lora_A', 'layers.12.feed_forward.w1.lora_B', 'layers.12.feed_forward.w3.lora_A', 'layers.12.feed_forward.w3.lora_B', 'layers.12.feed_forward.w2.lora_A', 'layers.12.feed_forward.w2.lora_B', 'layers.13.attention.wqkv.lora_A', 'layers.13.attention.wqkv.lora_B', 'layers.13.attention.wo.lora_A', 'layers.13.attention.wo.lora_B', 'layers.13.feed_forward.w1.lora_A', 'layers.13.feed_forward.w1.lora_B', 'layers.13.feed_forward.w3.lora_A', 'layers.13.feed_forward.w3.lora_B', 'layers.13.feed_forward.w2.lora_A', 'layers.13.feed_forward.w2.lora_B', 'layers.14.attention.wqkv.lora_A', 'layers.14.attention.wqkv.lora_B', 'layers.14.attention.wo.lora_A', 'layers.14.attention.wo.lora_B', 'layers.14.feed_forward.w1.lora_A', 'layers.14.feed_forward.w1.lora_B', 'layers.14.feed_forward.w3.lora_A', 'layers.14.feed_forward.w3.lora_B', 'layers.14.feed_forward.w2.lora_A', 'layers.14.feed_forward.w2.lora_B', 'layers.15.attention.wqkv.lora_A', 'layers.15.attention.wqkv.lora_B', 'layers.15.attention.wo.lora_A', 'layers.15.attention.wo.lora_B', 'layers.15.feed_forward.w1.lora_A', 'layers.15.feed_forward.w1.lora_B', 'layers.15.feed_forward.w3.lora_A', 'layers.15.feed_forward.w3.lora_B', 'layers.15.feed_forward.w2.lora_A', 'layers.15.feed_forward.w2.lora_B', 'layers.16.attention.wqkv.lora_A', 'layers.16.attention.wqkv.lora_B', 'layers.16.attention.wo.lora_A', 'layers.16.attention.wo.lora_B', 'layers.16.feed_forward.w1.lora_A', 'layers.16.feed_forward.w1.lora_B', 'layers.16.feed_forward.w3.lora_A', 'layers.16.feed_forward.w3.lora_B', 'layers.16.feed_forward.w2.lora_A', 'layers.16.feed_forward.w2.lora_B', 'layers.17.attention.wqkv.lora_A', 'layers.17.attention.wqkv.lora_B', 'layers.17.attention.wo.lora_A', 'layers.17.attention.wo.lora_B', 'layers.17.feed_forward.w1.lora_A', 'layers.17.feed_forward.w1.lora_B', 'layers.17.feed_forward.w3.lora_A', 'layers.17.feed_forward.w3.lora_B', 'layers.17.feed_forward.w2.lora_A', 'layers.17.feed_forward.w2.lora_B', 'layers.18.attention.wqkv.lora_A', 'layers.18.attention.wqkv.lora_B', 'layers.18.attention.wo.lora_A', 'layers.18.attention.wo.lora_B', 'layers.18.feed_forward.w1.lora_A', 'layers.18.feed_forward.w1.lora_B', 'layers.18.feed_forward.w3.lora_A', 'layers.18.feed_forward.w3.lora_B', 'layers.18.feed_forward.w2.lora_A', 'layers.18.feed_forward.w2.lora_B', 'layers.19.attention.wqkv.lora_A', 'layers.19.attention.wqkv.lora_B', 'layers.19.attention.wo.lora_A', 'layers.19.attention.wo.lora_B', 'layers.19.feed_forward.w1.lora_A', 'layers.19.feed_forward.w1.lora_B', 'layers.19.feed_forward.w3.lora_A', 'layers.19.feed_forward.w3.lora_B', 'layers.19.feed_forward.w2.lora_A', 'layers.19.feed_forward.w2.lora_B', 'layers.20.attention.wqkv.lora_A', 'layers.20.attention.wqkv.lora_B', 'layers.20.attention.wo.lora_A', 'layers.20.attention.wo.lora_B', 'layers.20.feed_forward.w1.lora_A', 'layers.20.feed_forward.w1.lora_B', 'layers.20.feed_forward.w3.lora_A', 'layers.20.feed_forward.w3.lora_B', 'layers.20.feed_forward.w2.lora_A', 'layers.20.feed_forward.w2.lora_B', 'layers.21.attention.wqkv.lora_A', 'layers.21.attention.wqkv.lora_B', 'layers.21.attention.wo.lora_A', 'layers.21.attention.wo.lora_B', 'layers.21.feed_forward.w1.lora_A', 'layers.21.feed_forward.w1.lora_B', 'layers.21.feed_forward.w3.lora_A', 'layers.21.feed_forward.w3.lora_B', 'layers.21.feed_forward.w2.lora_A', 'layers.21.feed_forward.w2.lora_B', 'layers.22.attention.wqkv.lora_A', 'layers.22.attention.wqkv.lora_B', 'layers.22.attention.wo.lora_A', 'layers.22.attention.wo.lora_B', 'layers.22.feed_forward.w1.lora_A', 'layers.22.feed_forward.w1.lora_B', 'layers.22.feed_forward.w3.lora_A', 'layers.22.feed_forward.w3.lora_B', 'layers.22.feed_forward.w2.lora_A', 'layers.22.feed_forward.w2.lora_B', 'layers.23.attention.wqkv.lora_A', 'layers.23.attention.wqkv.lora_B', 'layers.23.attention.wo.lora_A', 'layers.23.attention.wo.lora_B', 'layers.23.feed_forward.w1.lora_A', 'layers.23.feed_forward.w1.lora_B', 'layers.23.feed_forward.w3.lora_A', 'layers.23.feed_forward.w3.lora_B', 'layers.23.feed_forward.w2.lora_A', 'layers.23.feed_forward.w2.lora_B', 'output.lora_A', 'output.lora_B', 'fast_embeddings.lora_A', 'fast_embeddings.lora_B', 'fast_layers.0.attention.wqkv.lora_A', 'fast_layers.0.attention.wqkv.lora_B', 'fast_layers.0.attention.wo.lora_A', 'fast_layers.0.attention.wo.lora_B', 'fast_layers.0.feed_forward.w1.lora_A', 'fast_layers.0.feed_forward.w1.lora_B', 'fast_layers.0.feed_forward.w3.lora_A', 'fast_layers.0.feed_forward.w3.lora_B', 'fast_layers.0.feed_forward.w2.lora_A', 'fast_layers.0.feed_forward.w2.lora_B', 'fast_layers.1.attention.wqkv.lora_A', 'fast_layers.1.attention.wqkv.lora_B', 'fast_layers.1.attention.wo.lora_A', 'fast_layers.1.attention.wo.lora_B', 'fast_layers.1.feed_forward.w1.lora_A', 'fast_layers.1.feed_forward.w1.lora_B', 'fast_layers.1.feed_forward.w3.lora_A', 'fast_layers.1.feed_forward.w3.lora_B', 'fast_layers.1.feed_forward.w2.lora_A', 'fast_layers.1.feed_forward.w2.lora_B', 'fast_layers.2.attention.wqkv.lora_A', 'fast_layers.2.attention.wqkv.lora_B', 'fast_layers.2.attention.wo.lora_A', 'fast_layers.2.attention.wo.lora_B', 'fast_layers.2.feed_forward.w1.lora_A', 'fast_layers.2.feed_forward.w1.lora_B', 'fast_layers.2.feed_forward.w3.lora_A', 'fast_layers.2.feed_forward.w3.lora_B', 'fast_layers.2.feed_forward.w2.lora_A', 'fast_layers.2.feed_forward.w2.lora_B', 'fast_layers.3.attention.wqkv.lora_A', 'fast_layers.3.attention.wqkv.lora_B', 'fast_layers.3.attention.wo.lora_A', 'fast_layers.3.attention.wo.lora_B', 'fast_layers.3.feed_forward.w1.lora_A', 'fast_layers.3.feed_forward.w1.lora_B', 'fast_layers.3.feed_forward.w3.lora_A', 'fast_layers.3.feed_forward.w3.lora_B', 'fast_layers.3.feed_forward.w2.lora_A', 'fast_layers.3.feed_forward.w2.lora_B', 'fast_output.lora_A', 'fast_output.lora_B'], unexpected_keys=[])
8
+ [2024-07-12 22:42:31,844][__main__][INFO] - [rank: 0] Instantiating callbacks...
9
+ [2024-07-12 22:42:31,844][fish_speech.utils.instantiators][INFO] - [rank: 0] Instantiating callback <lightning.pytorch.callbacks.ModelCheckpoint>
10
+ [2024-07-12 22:42:31,847][fish_speech.utils.instantiators][INFO] - [rank: 0] Instantiating callback <lightning.pytorch.callbacks.ModelSummary>
11
+ [2024-07-12 22:42:31,847][fish_speech.utils.instantiators][INFO] - [rank: 0] Instantiating callback <lightning.pytorch.callbacks.LearningRateMonitor>
12
+ [2024-07-12 22:42:31,848][fish_speech.utils.instantiators][INFO] - [rank: 0] Instantiating callback <fish_speech.callbacks.GradNormMonitor>
13
+ [2024-07-12 22:42:31,865][__main__][INFO] - [rank: 0] Instantiating loggers...
14
+ [2024-07-12 22:42:31,865][fish_speech.utils.instantiators][INFO] - [rank: 0] Instantiating logger <lightning.pytorch.loggers.tensorboard.TensorBoardLogger>
15
+ [2024-07-12 22:42:31,873][__main__][INFO] - [rank: 0] Instantiating trainer <lightning.pytorch.trainer.Trainer>
16
+ [2024-07-12 22:42:31,917][__main__][INFO] - [rank: 0] Logging hyperparameters!
17
+ [2024-07-12 22:42:31,998][__main__][INFO] - [rank: 0] Starting training!
18
+ [2024-07-12 22:42:59,315][fish_speech.models.text2semantic.lit_module][INFO] - [rank: 0] Set weight decay: 0.01 for 432 parameters
19
+ [2024-07-12 22:42:59,315][fish_speech.models.text2semantic.lit_module][INFO] - [rank: 0] Set weight decay: 0.0 for 61 parameters
20
+ [2024-07-12 22:42:59,586][fish_speech.datasets.semantic][INFO] - [rank: 0] Reading 2 / 2 files
21
+ [2024-07-12 22:42:59,586][fish_speech.datasets.semantic][INFO] - [rank: 0] Reading 1 / 2 files
22
+ [2024-07-12 22:42:59,586][fish_speech.datasets.semantic][INFO] - [rank: 0] Reading 2 / 2 files
23
+ [2024-07-12 22:42:59,586][fish_speech.datasets.semantic][INFO] - [rank: 0] Reading 1 / 2 files
24
+ [2024-07-12 22:42:59,687][fish_speech.datasets.semantic][INFO] - [rank: 0] Read total 3 groups of data
25
+ [2024-07-12 22:42:59,714][fish_speech.datasets.semantic][INFO] - [rank: 0] Read total 2 groups of data
26
+ [2024-07-12 22:42:59,784][fish_speech.datasets.semantic][INFO] - [rank: 0] Read total 6 groups of data
27
+ [2024-07-12 22:42:59,804][fish_speech.datasets.semantic][INFO] - [rank: 0] Read total 4 groups of data
28
+ [2024-07-12 22:43:00,779][fish_speech.datasets.semantic][INFO] - [rank: 0] Reading 2 / 2 files
29
+ [2024-07-12 22:43:00,779][fish_speech.datasets.semantic][INFO] - [rank: 0] Reading 2 / 2 files
30
+ [2024-07-12 22:43:00,780][fish_speech.datasets.semantic][INFO] - [rank: 0] Reading 1 / 2 files
31
+ [2024-07-12 22:43:00,780][fish_speech.datasets.semantic][INFO] - [rank: 0] Reading 1 / 2 files
32
+ [2024-07-12 22:43:00,891][fish_speech.datasets.semantic][INFO] - [rank: 0] Read total 3 groups of data
33
+ [2024-07-12 22:43:00,903][fish_speech.datasets.semantic][INFO] - [rank: 0] Read total 2 groups of data
34
+ [2024-07-12 22:43:00,967][fish_speech.datasets.semantic][INFO] - [rank: 0] Read total 6 groups of data
35
+ [2024-07-12 22:43:00,999][fish_speech.datasets.semantic][INFO] - [rank: 0] Read total 4 groups of data
36
+ [2024-07-13 14:47:53,112][fish_speech.utils.utils][INFO] - [rank: 0] Output dir: results/melody_v1