Safetensors
tts
vc
svs
svc
music
RMSnow commited on
Commit
795b27d
·
verified ·
1 Parent(s): be32f72

Upload folder using huggingface_hub

Browse files
Files changed (46) hide show
  1. .gitattributes +2 -0
  2. acoustic_modeling/fm_emilia101k_singnet7k_repa/config.json +62 -0
  3. acoustic_modeling/fm_emilia101k_singnet7k_repa/model.safetensors +3 -0
  4. acoustic_modeling/fm_emilia101k_singnet7k_repa/whisper_stats.pt +3 -0
  5. acoustic_modeling/fm_emilia101k_singnet7k_repa_text/config.json +63 -0
  6. acoustic_modeling/fm_emilia101k_singnet7k_repa_text/model.safetensors +3 -0
  7. acoustic_modeling/fm_emilia101k_singnet7k_repa_text/whisper_stats.pt +3 -0
  8. contentstyle_modeling/posttrained/added_tokens.json +0 -0
  9. contentstyle_modeling/posttrained/amphion_config.json +73 -0
  10. contentstyle_modeling/posttrained/config.json +29 -0
  11. contentstyle_modeling/posttrained/generation_config.json +14 -0
  12. contentstyle_modeling/posttrained/merges.txt +0 -0
  13. contentstyle_modeling/posttrained/model.safetensors +3 -0
  14. contentstyle_modeling/posttrained/optimizer.pt +3 -0
  15. contentstyle_modeling/posttrained/rng_state_0.pth +3 -0
  16. contentstyle_modeling/posttrained/rng_state_1.pth +3 -0
  17. contentstyle_modeling/posttrained/rng_state_2.pth +3 -0
  18. contentstyle_modeling/posttrained/rng_state_3.pth +3 -0
  19. contentstyle_modeling/posttrained/rng_state_4.pth +3 -0
  20. contentstyle_modeling/posttrained/rng_state_5.pth +3 -0
  21. contentstyle_modeling/posttrained/rng_state_6.pth +3 -0
  22. contentstyle_modeling/posttrained/scheduler.pt +3 -0
  23. contentstyle_modeling/posttrained/special_tokens_map.json +0 -0
  24. contentstyle_modeling/posttrained/tokenizer.json +3 -0
  25. contentstyle_modeling/posttrained/tokenizer_config.json +0 -0
  26. contentstyle_modeling/posttrained/trainer_state.json +0 -0
  27. contentstyle_modeling/posttrained/training_args.bin +3 -0
  28. contentstyle_modeling/posttrained/vocab.json +0 -0
  29. contentstyle_modeling/pretrained/added_tokens.json +0 -0
  30. contentstyle_modeling/pretrained/config.json +28 -0
  31. contentstyle_modeling/pretrained/generation_config.json +14 -0
  32. contentstyle_modeling/pretrained/merges.txt +0 -0
  33. contentstyle_modeling/pretrained/model.safetensors +3 -0
  34. contentstyle_modeling/pretrained/optimizer.bin +3 -0
  35. contentstyle_modeling/pretrained/random_states_0.pkl +3 -0
  36. contentstyle_modeling/pretrained/scheduler.bin +3 -0
  37. contentstyle_modeling/pretrained/special_tokens_map.json +0 -0
  38. contentstyle_modeling/pretrained/tokenizer.json +3 -0
  39. contentstyle_modeling/pretrained/tokenizer_config.json +0 -0
  40. contentstyle_modeling/pretrained/vocab.json +0 -0
  41. tokenizer/contentstyle_fvq16384_12.5hz/model.safetensors +3 -0
  42. tokenizer/prosody_fvq512_6.25hz/model.safetensors +3 -0
  43. vocoder/config.json +84 -0
  44. vocoder/model.safetensors +3 -0
  45. vocoder/model_1.safetensors +3 -0
  46. vocoder/model_2.safetensors +3 -0
.gitattributes CHANGED
@@ -33,3 +33,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ contentstyle_modeling/posttrained/tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
+ contentstyle_modeling/pretrained/tokenizer.json filter=lfs diff=lfs merge=lfs -text
acoustic_modeling/fm_emilia101k_singnet7k_repa/config.json ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_type": "FlowMatchingTransformer",
3
+ "preprocess": {
4
+ "hop_size": 480,
5
+ "sample_rate": 24000,
6
+ "n_fft": 1920,
7
+ "num_mels": 128,
8
+ "win_size": 1920,
9
+ "fmin": 0,
10
+ "fmax": 12000,
11
+ "mel_var": 8.14,
12
+ "mel_mean": -4.92,
13
+ "f0_fmin": 50.0,
14
+ "f0_fmax": 1100.0,
15
+ "load_phone": false,
16
+ "load_chromagram": true,
17
+ "load_semantic_features": true,
18
+ },
19
+ "model": {
20
+ "flow_matching_transformer": {
21
+ "mel_dim": 128,
22
+ "hidden_size": 1024,
23
+ "num_layers": 16,
24
+ "num_heads": 16,
25
+ "cfg_scale": 0.2,
26
+ "use_cond_code": true, // false means Hidden features
27
+ // "cond_dim": 1024, // HuBERT features dimension
28
+ "cond_codebook_size": 16384, // VQ Codebook Size
29
+ "cond_scale_factor": 4, // 1 means not use ReTrans. 4 means 12.5Hz * 4 = 50Hz. This should be aligned with the frame rate with Mels
30
+ "sigma": 1e-5,
31
+ "time_scheduler": "cos",
32
+ "whisper_perturb": false,
33
+ "repa": {
34
+ "layer_index": 5, // Use the Wav2Vec2Bert features to align. 5 means the 6th layer.
35
+ "output_dim": 1024, // The dimension of the Wav2Vec2Bert features.
36
+ "loss_type": "cos", // "cos" or "l1". By default is "l1"
37
+ },
38
+ },
39
+ "cond_sample_rate": 16000, // whisper: 16000
40
+ "coco": {
41
+ "coco_type": "content_style", // content, style, or content_style
42
+ "downsample_rate": 4, // The original frame rate is 50 Hz, downsample to 12.5 Hz
43
+ "codebook_size": 16384,
44
+ "hidden_size": 1024, // Representations Dim
45
+ "codebook_dim": 8,
46
+ "encoder": {
47
+ "vocos_dim": 384,
48
+ "vocos_intermediate_dim": 2048,
49
+ "vocos_num_layers": 12,
50
+ },
51
+ "decoder": {
52
+ "vocos_dim": 384,
53
+ "vocos_intermediate_dim": 2048,
54
+ "vocos_num_layers": 12,
55
+ },
56
+ "use_normed_whisper": true,
57
+ "whisper_stats_path": "models/svc/vevosing/config/whisper_stats.pt",
58
+ "whisper_dim": 1024,
59
+ "chromagram_dim": 24,
60
+ },
61
+ },
62
+ }
acoustic_modeling/fm_emilia101k_singnet7k_repa/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef3733b3f92cf8f38e32f6d161f58751247b199cffc15cc1562e76c1289a7186
3
+ size 1451496488
acoustic_modeling/fm_emilia101k_singnet7k_repa/whisper_stats.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6117052b3e23e9075a79cc208ecc24328ae2f71a7dd4c9793db7c90a88b4a519
3
+ size 9215
acoustic_modeling/fm_emilia101k_singnet7k_repa_text/config.json ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_type": "FlowMatchingTransformer",
3
+ "preprocess": {
4
+ "hop_size": 480,
5
+ "sample_rate": 24000,
6
+ "n_fft": 1920,
7
+ "num_mels": 128,
8
+ "win_size": 1920,
9
+ "fmin": 0,
10
+ "fmax": 12000,
11
+ "mel_var": 8.14,
12
+ "mel_mean": -4.92,
13
+ "f0_fmin": 50.0,
14
+ "f0_fmax": 1100.0,
15
+ "load_phone": false,
16
+ "load_singnet_phone": false,
17
+ "load_chromagram": true,
18
+ "load_semantic_features": true,
19
+ "load_text": true,
20
+ "pad_token_id": 32000,
21
+ "tokenizer_path": "microsoft/Phi-3.5-mini-instruct"
22
+ },
23
+ "model": {
24
+ "flow_matching_transformer": {
25
+ "mel_dim": 128,
26
+ "hidden_size": 1024,
27
+ "num_layers": 16,
28
+ "num_heads": 16,
29
+ "cfg_scale": 0.2,
30
+ "use_cond_code": true, // false means Hidden features
31
+ // "cond_dim": 1024, // HuBERT features dimension
32
+ "cond_codebook_size": 16384, // VQ Codebook Size
33
+ "cond_scale_factor": 4, // 1 means not use ReTrans. 4 means 12.5Hz * 4 = 50Hz. This should be aligned with the frame rate with Mels
34
+ "sigma": 1e-5,
35
+ "time_scheduler": "cos",
36
+ "whisper_perturb": false,
37
+ "repa": {
38
+ "layer_index": 5, // Use the Wav2Vec2Bert features to align. 5 means the 6th layer.
39
+ "output_dim": 1024, // The dimension of the Wav2Vec2Bert features.
40
+ "loss_type": "cos", // "cos" or "l1". By default is "l1"
41
+ },
42
+ "use_text_as_condition": true,
43
+ "text_cond_codebook_size": 32100
44
+ },
45
+ "cond_sample_rate": 16000, // whisper: 16000
46
+ "coco": {
47
+ "coco_type": "content_style", // content, style, or content_style
48
+ "downsample_rate": 4, // The original frame rate is 50 Hz, downsample to 12.5 Hz
49
+ "codebook_size": 16384,
50
+ "hidden_size": 1024, // Representations Dim
51
+ "codebook_dim": 8,
52
+ "encoder": {
53
+ "vocos_dim": 384,
54
+ "vocos_intermediate_dim": 2048,
55
+ "vocos_num_layers": 12,
56
+ },
57
+ "use_normed_whisper": true,
58
+ "whisper_stats_path": "whisper_stats.pt",
59
+ "whisper_dim": 1024,
60
+ "chromagram_dim": 24,
61
+ },
62
+ },
63
+ }
acoustic_modeling/fm_emilia101k_singnet7k_repa_text/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c972f24183e1347d6b767c9d2804959de0506bfcb4836992aba9cd33ce4e9ed1
3
+ size 1616553544
acoustic_modeling/fm_emilia101k_singnet7k_repa_text/whisper_stats.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6117052b3e23e9075a79cc208ecc24328ae2f71a7dd4c9793db7c90a88b4a519
3
+ size 9215
contentstyle_modeling/posttrained/added_tokens.json ADDED
The diff for this file is too large to render. See raw diff
 
contentstyle_modeling/posttrained/amphion_config.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "preprocess": {
3
+ "hop_size": 480,
4
+ "sample_rate": 24000,
5
+ "n_fft": 1920,
6
+ "num_mels": 128,
7
+ "win_size": 1920,
8
+ "fmin": 0,
9
+ "fmax": 12000,
10
+ "mel_var": 8.14,
11
+ "mel_mean": -4.92,
12
+ "f0_fmin": 50.0,
13
+ "f0_fmax": 1100.0,
14
+ "wav_code_frame_rate": 18.75, // Vevo2: 12.5 (Content-Style Code) + 6.25 (Prosody Code) = 18.75
15
+ "min_dur": 1,
16
+ "max_dur": 30,
17
+ "drop_prosody_id_prob": -1, // Dropping prosody ids means the Text-to-CS, while not dropping means the Text+Note-to-CS,
18
+ "pad_token_id": 151643, // <|endoftext|> for Qwen2.5-0.5B-Instruct,
19
+ "eos_token": "<|im_end|>",
20
+ "eos_token_id": 151645, // <|im_end|> for Qwen2.5-0.5B-Instruct,
21
+ // "tokenizer_path": "/mnt/data4/zhangxueyao/SpeechGenerationYC_ckpts/ckpts/vevo2/pretrained/Qwen2.5-0.5B-Instruct-add_prosody_contentstyle"
22
+ },
23
+ "model": {
24
+ // "pretrained_model_path": "/mnt/data4/zhangxueyao/SpeechGenerationYC_ckpts/ckpts/vevo2/pretrained/Qwen2.5-0.5B-Instruct-add_prosody_contentstyle", // Qwen2.5 Model
25
+ // "rl_init_model_path": "/mnt/data4/zhangxueyao/SpeechGenerationYC_ckpts/ckpts/vevo2/llm_dpo/dpo_qwen0.5B_intp2_highsim_3e-5/checkpoint_backup/epoch-0023_step-0027000_loss-0.000961", // DPO Model
26
+ "use_intelligibility_reward": true,
27
+ "use_chromagram_reward": true,
28
+ "use_target_length_reward": true,
29
+ "reward_combination_strategy": "advantage_first", // "reward_first" or "advantage_first"
30
+ "coco_style": {
31
+ "coco_type": "style", // content, style, or content_style
32
+ "downsample_rate": 8, // The original frame rate is 50 Hz, downsample to 6.25 Hz
33
+ "codebook_size": 512,
34
+ "hidden_size": 1024, // Representations Dim
35
+ "codebook_dim": 8,
36
+ "encoder": {
37
+ "vocos_dim": 384,
38
+ "vocos_intermediate_dim": 2048,
39
+ "vocos_num_layers": 12,
40
+ },
41
+ "decoder": {
42
+ "vocos_dim": 384,
43
+ "vocos_intermediate_dim": 2048,
44
+ "vocos_num_layers": 12,
45
+ },
46
+ "use_normed_whisper": true,
47
+ "whisper_stats_path": "models/svc/vevosing/config/whisper_stats.pt",
48
+ "whisper_dim": 1024,
49
+ "chromagram_dim": 24,
50
+ },
51
+ "coco_content_style": {
52
+ "coco_type": "content_style", // content, style, or content_style
53
+ "downsample_rate": 4, // The original frame rate is 50 Hz, downsample to 12.5 Hz
54
+ "codebook_size": 16384,
55
+ "hidden_size": 1024, // Representations Dim
56
+ "codebook_dim": 8,
57
+ "encoder": {
58
+ "vocos_dim": 384,
59
+ "vocos_intermediate_dim": 2048,
60
+ "vocos_num_layers": 12,
61
+ },
62
+ "decoder": {
63
+ "vocos_dim": 384,
64
+ "vocos_intermediate_dim": 2048,
65
+ "vocos_num_layers": 12,
66
+ },
67
+ "use_normed_whisper": true,
68
+ "whisper_stats_path": "models/svc/vevosing/config/whisper_stats.pt",
69
+ "whisper_dim": 1024,
70
+ "chromagram_dim": 24,
71
+ },
72
+ },
73
+ }
contentstyle_modeling/posttrained/config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "epoch-0023_step-0027000_loss-0.000961",
3
+ "architectures": [
4
+ "Qwen2ForCausalLM"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 151643,
8
+ "eos_token_id": 151645,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 896,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 4864,
13
+ "max_position_embeddings": 32768,
14
+ "max_window_layers": 21,
15
+ "model_type": "qwen2",
16
+ "num_attention_heads": 14,
17
+ "num_hidden_layers": 24,
18
+ "num_key_value_heads": 2,
19
+ "rms_norm_eps": 1e-06,
20
+ "rope_scaling": null,
21
+ "rope_theta": 1000000.0,
22
+ "sliding_window": null,
23
+ "tie_word_embeddings": true,
24
+ "torch_dtype": "bfloat16",
25
+ "transformers_version": "4.47.1",
26
+ "use_cache": true,
27
+ "use_sliding_window": false,
28
+ "vocab_size": 168565
29
+ }
contentstyle_modeling/posttrained/generation_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 151643,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 151645,
6
+ 151643
7
+ ],
8
+ "pad_token_id": 151643,
9
+ "repetition_penalty": 1.1,
10
+ "temperature": 0.7,
11
+ "top_k": 20,
12
+ "top_p": 0.8,
13
+ "transformers_version": "4.47.1"
14
+ }
contentstyle_modeling/posttrained/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
contentstyle_modeling/posttrained/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b056f8b4f0616dd5eb8598a0cc3fe5a4bd9dcdffc2dbe344156a16691496cf16
3
+ size 1017897016
contentstyle_modeling/posttrained/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:00d6a8b09b3c9b33a89928078d4419801da83503ee168ad218b54bcb4b2de80c
3
+ size 2035976634
contentstyle_modeling/posttrained/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a0d21fa6169554d53edc945a579970dc3448968f10ab1416295d1014dbbb9f49
3
+ size 15728
contentstyle_modeling/posttrained/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:50cace7ce599554e1b01251eb0e29892e1a9894d33e4aa3ac9b2cc0a33b31d45
3
+ size 15792
contentstyle_modeling/posttrained/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f7dbd212f619e6e1aa021322d3a3eb2379e92d7530cc581bf9d5da6d99dac64
3
+ size 15728
contentstyle_modeling/posttrained/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d595703e986b376bcbbbe983db3f7f9d54467320ea6892831e36d4ddcb28c7d
3
+ size 15792
contentstyle_modeling/posttrained/rng_state_4.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54858b6df6db301adc27c00cf899bf8783437a01f590241a20b7b3b0c576436f
3
+ size 15728
contentstyle_modeling/posttrained/rng_state_5.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8209d49950616ee1e3115d8ba85e44f92746f2cf274d7e65daa66a5ddbffb516
3
+ size 15792
contentstyle_modeling/posttrained/rng_state_6.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4c9c457ef48482488944e26e066bff5b43a88ecd73457a8cf068e2568980401d
3
+ size 15664
contentstyle_modeling/posttrained/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:42f755b43c342264783f3a2fa30b921b7bf7353b3e3649a4cf4c8ac31ede21ad
3
+ size 1064
contentstyle_modeling/posttrained/special_tokens_map.json ADDED
The diff for this file is too large to render. See raw diff
 
contentstyle_modeling/posttrained/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b4afa3f042a3dd5ddf573c035c68f88810d01d714fe3341054e9688c0eebef7
3
+ size 13891825
contentstyle_modeling/posttrained/tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff
 
contentstyle_modeling/posttrained/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
contentstyle_modeling/posttrained/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d8b222db8e20e9927c6193a02f9e01d452e03a5f706ae69e0bc6b500a0711f7
3
+ size 6264
contentstyle_modeling/posttrained/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
contentstyle_modeling/pretrained/added_tokens.json ADDED
The diff for this file is too large to render. See raw diff
 
contentstyle_modeling/pretrained/config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/storage/zhangxueyao/workspace/SpeechGenerationYC/pretrained/Qwen2.5-0.5B-Instruct",
3
+ "architectures": [
4
+ "Qwen2ForCausalLM"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 151643,
8
+ "eos_token_id": 151645,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 896,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 4864,
13
+ "max_position_embeddings": 32768,
14
+ "max_window_layers": 21,
15
+ "model_type": "qwen2",
16
+ "num_attention_heads": 14,
17
+ "num_hidden_layers": 24,
18
+ "num_key_value_heads": 2,
19
+ "rms_norm_eps": 1e-06,
20
+ "rope_theta": 1000000.0,
21
+ "sliding_window": null,
22
+ "tie_word_embeddings": true,
23
+ "torch_dtype": "bfloat16",
24
+ "transformers_version": "4.44.1",
25
+ "use_cache": true,
26
+ "use_sliding_window": false,
27
+ "vocab_size": 168565
28
+ }
contentstyle_modeling/pretrained/generation_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 151643,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 151645,
6
+ 151643
7
+ ],
8
+ "pad_token_id": 151643,
9
+ "repetition_penalty": 1.1,
10
+ "temperature": 0.7,
11
+ "top_k": 20,
12
+ "top_p": 0.8,
13
+ "transformers_version": "4.44.1"
14
+ }
contentstyle_modeling/pretrained/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
contentstyle_modeling/pretrained/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:531be07b9eb7337988cda77842ced780f0cea3d6d30395ca19224f338698551f
3
+ size 1017897016
contentstyle_modeling/pretrained/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5773426880bb14310f7f2091ecdb25c40ca5fb0ff13a1673462d8f9c4d4bfd77
3
+ size 2035976442
contentstyle_modeling/pretrained/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd94f420a504ac49f408ed27b8ed54e94ccb1843fdae48a98404069db675e292
3
+ size 16036
contentstyle_modeling/pretrained/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e945d4b96759596852af2d7406b1573fb5c0ef023de89ac9a7a1a7ab37cdbdf
3
+ size 1000
contentstyle_modeling/pretrained/special_tokens_map.json ADDED
The diff for this file is too large to render. See raw diff
 
contentstyle_modeling/pretrained/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc6a67d19953836a0b0f3ab9385f2a29194fe98feadf07ef8628a9eb6d58509e
3
+ size 12544343
contentstyle_modeling/pretrained/tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff
 
contentstyle_modeling/pretrained/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer/contentstyle_fvq16384_12.5hz/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bebdcea39e2d0134dbcce193aed0e7b0d393a866346317cd196389e40e9151b0
3
+ size 244781696
tokenizer/prosody_fvq512_6.25hz/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:754ba5928dfe4dd3c4b6e91ce396e2a50cff52937053c0e3e4ac3a854a4b8ae4
3
+ size 273642504
vocoder/config.json ADDED
@@ -0,0 +1,84 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_type": "Vocoder",
3
+ "preprocess": {
4
+ "hop_size": 480,
5
+ "sample_rate": 24000,
6
+ "max_length": 36000,
7
+ "n_fft": 1920,
8
+ "num_mels": 128,
9
+ "win_size": 1920,
10
+ "fmin": 0,
11
+ "fmax": 12000,
12
+ "mel_var": 8.14,
13
+ "mel_mean": -4.92,
14
+ "load_phone": false,
15
+ "load_chromagram": false
16
+ },
17
+ "model": {
18
+ "vocos": {
19
+ "input_channels": 128,
20
+ "dim": 1024,
21
+ "intermediate_dim": 4096,
22
+ "num_layers": 30,
23
+ "n_fft": 1920,
24
+ "hop_size": 480,
25
+ "padding": "same"
26
+ },
27
+ "period_gan": {
28
+ "max_downsample_channels": 1024,
29
+ "channels": 64,
30
+ "channel_increasing_factor": 2
31
+ },
32
+ "spec_gan": {
33
+ "stft_params": {
34
+ "fft_sizes": [
35
+ 128,
36
+ 256,
37
+ 512,
38
+ 1024,
39
+ 2048
40
+ ],
41
+ "hop_sizes": [
42
+ 32,
43
+ 64,
44
+ 128,
45
+ 256,
46
+ 512
47
+ ],
48
+ "win_lengths": [
49
+ 128,
50
+ 256,
51
+ 512,
52
+ 1024,
53
+ 2048
54
+ ],
55
+ "window": "hann_window"
56
+ },
57
+ "in_channels": 1,
58
+ "out_channels": 1,
59
+ "channels": 64,
60
+ "kernel_sizes": [
61
+ 5,
62
+ 3
63
+ ],
64
+ "max_downsample_channels": 1024,
65
+ "down_scales": [
66
+ 2,
67
+ 2,
68
+ 2
69
+ ],
70
+ "use_weight_norm": true,
71
+ "use_complex": false
72
+ }
73
+ },
74
+ "loss": {
75
+ "mel_loss": {
76
+ "sample_rate": 24000
77
+ },
78
+ "disc_loss_weight": 1.0,
79
+ "mel_loss_weight": 10.0,
80
+ "adv_loss_weight": 2.0,
81
+ "fm_loss_weight": 2.0,
82
+ "spec_fm_loss_weight": 1.0
83
+ },
84
+ }
vocoder/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b5d1a46b19351c9a71bd8a5a59dd16be0be2ddefe70d3a0b4915d9a425e56d3
3
+ size 1020206416
vocoder/model_1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:850799d78699134b969056183fc9d490c51f8d8154d0ed00fae3e738b6b30af6
3
+ size 69768280
vocoder/model_2.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:56130fd13d5fbe828e56d61edb0049d35700db0472a866b8167d1d217d2687f8
3
+ size 180693296