model: _target_: fish_speech.models.text2semantic.lit_module.TextToSemantic model: _target_: fish_speech.models.text2semantic.llama.BaseTransformer.from_pretrained path: checkpoints/fish-speech-1.2 load_weights: true max_length: 1024 lora_config: _target_: fish_speech.models.text2semantic.lora.LoraConfig r: 8 lora_alpha: 16 lora_dropout: 0.01 optimizer: _target_: torch.optim.AdamW _partial_: true lr: 0.0001 weight_decay: 0.01 betas: - 0.9 - 0.95 eps: 1.0e-05 lr_scheduler: _target_: torch.optim.lr_scheduler.LambdaLR _partial_: true lr_lambda: _target_: fish_speech.scheduler.get_constant_schedule_with_warmup_lr_lambda _partial_: true num_warmup_steps: 50 model/params/total: 495286272 model/params/trainable: 5017600 model/params/non_trainable: 490268672 data: _target_: fish_speech.datasets.semantic.SemanticDataModule train_dataset: _target_: fish_speech.datasets.semantic.AutoTextSemanticInstructionDataset proto_files: - data/protos tokenizer: _target_: transformers.AutoTokenizer.from_pretrained pretrained_model_name_or_path: checkpoints/fish-speech-1.2 causal: true max_length: 1024 use_speaker: false interactive_prob: 0.7 val_dataset: _target_: fish_speech.datasets.semantic.AutoTextSemanticInstructionDataset proto_files: - data/protos tokenizer: _target_: transformers.AutoTokenizer.from_pretrained pretrained_model_name_or_path: checkpoints/fish-speech-1.2 causal: true max_length: 1024 use_speaker: false interactive_prob: 0.7 num_workers: 4 batch_size: 4 tokenizer: _target_: transformers.AutoTokenizer.from_pretrained pretrained_model_name_or_path: checkpoints/fish-speech-1.2 max_length: 1024 trainer: _target_: lightning.pytorch.trainer.Trainer default_root_dir: results/mix_v2 accelerator: gpu num_nodes: 1 devices: auto strategy: _target_: lightning.pytorch.strategies.DDPStrategy process_group_backend: nccl precision: bf16-true check_val_every_n_epoch: null val_check_interval: 1000 max_steps: 100000 benchmark: true accumulate_grad_batches: 1 gradient_clip_val: 1.0 gradient_clip_algorithm: norm limit_val_batches: 10 callbacks: model_checkpoint: _target_: lightning.pytorch.callbacks.ModelCheckpoint dirpath: results/mix_v2/checkpoints filename: step_{step:09d} save_last: false save_top_k: 5 monitor: step mode: max every_n_epochs: null every_n_train_steps: 1000 auto_insert_metric_name: false model_summary: _target_: lightning.pytorch.callbacks.ModelSummary max_depth: 2 learning_rate_monitor: _target_: lightning.pytorch.callbacks.LearningRateMonitor logging_interval: step log_momentum: false grad_norm_monitor: _target_: fish_speech.callbacks.GradNormMonitor norm_type: 2 logging_interval: step extras: null task_name: null tags: null ckpt_path: null seed: null