File size: 1,260 Bytes
9ba5132
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
output_dir: "logs/s1"
train:
  seed: 1234
  epochs: 15
  batch_size: 8
  save_every_n_epoch: 5
  precision: 32
  if_save_latest: true
  if_save_every_weights: true
  exp_name: "gpt_training"
  half_weights_save_dir: "weights/s1"
  wandb:
    project: "gpt-sovits-hindi"
    name: "stage1_training"
    entity: null
    log_interval: 100

optimizer:
  lr_init: 0.0001
  lr: 0.0004
  lr_end: 0.00001
  warmup_steps: 500
  decay_steps: 1000

data:
  training_files: "data8"
  max_sec: 60
  max_frames: 60
  filter_length: 2048
  hop_length: 640
  win_length: 2048
  mel_channels: 128
  mel_fmin: 0.0
  mel_fmax: null
  cleaned_text: true
  num_workers: 4
  batch_size: 8
  pad_val: 1024

# Data paths
train_semantic_path: "data8/semantic.tsv"
train_phoneme_path: "data8/phoneme.txt"

model:
  hidden_dim: 768
  embedding_dim: 768
  n_layer: 12
  head: 12
  n_embd: 768
  vocab_size: 2048
  block_size: 1000
  embd_pdrop: 0.1
  resid_pdrop: 0.1
  attn_pdrop: 0.1
  semantic_dim: 1024
  num_layers: 6
  ffn_hidden: 3072
  dropout: 0.1
  attention_dropout: 0.1
  hidden_dropout: 0.1
  max_text_positions: 2048
  max_mel_positions: 8000
  prenet_dim: 384
  postnet_dim: 384
  prenet_layers: 3
  postnet_layers: 3
  phoneme_vocab_size: 2048
  EOS: 2047
  pad_val: 1024