Munaza10 commited on
Commit
9f8de7a
·
verified ·
1 Parent(s): 9dbc586

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer/tokenizer.json filter=lfs diff=lfs merge=lfs -text
audio_token_detokenizer/config.json ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AceStepAudioTokenDetokenizer",
3
+ "_diffusers_version": "0.39.0.dev0",
4
+ "attention_bias": false,
5
+ "attention_dropout": 0.0,
6
+ "audio_acoustic_hidden_dim": 64,
7
+ "head_dim": 128,
8
+ "hidden_size": 2048,
9
+ "intermediate_size": 6144,
10
+ "layer_types": [
11
+ "sliding_attention",
12
+ "full_attention"
13
+ ],
14
+ "num_attention_heads": 16,
15
+ "num_attention_pooler_hidden_layers": 2,
16
+ "num_key_value_heads": 8,
17
+ "pool_window_size": 5,
18
+ "rms_norm_eps": 1e-06,
19
+ "rope_theta": 1000000,
20
+ "sliding_window": 128
21
+ }
audio_token_detokenizer/diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:57f4bb761e6305efe4c28e9e7e7e0e353fffc5dec28af28a6a7d5429396c8edf
3
+ size 210026424
audio_tokenizer/config.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AceStepAudioTokenizer",
3
+ "_diffusers_version": "0.39.0.dev0",
4
+ "attention_bias": false,
5
+ "attention_dropout": 0.0,
6
+ "audio_acoustic_hidden_dim": 64,
7
+ "fsq_dim": 2048,
8
+ "fsq_input_levels": [
9
+ 8,
10
+ 8,
11
+ 8,
12
+ 5,
13
+ 5,
14
+ 5
15
+ ],
16
+ "fsq_input_num_quantizers": 1,
17
+ "head_dim": 128,
18
+ "hidden_size": 2048,
19
+ "intermediate_size": 6144,
20
+ "layer_types": [
21
+ "sliding_attention",
22
+ "full_attention"
23
+ ],
24
+ "num_attention_heads": 16,
25
+ "num_attention_pooler_hidden_layers": 2,
26
+ "num_key_value_heads": 8,
27
+ "pool_window_size": 5,
28
+ "rms_norm_eps": 1e-06,
29
+ "rope_theta": 1000000,
30
+ "sliding_window": 128
31
+ }
audio_tokenizer/diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b6e89e32223dc1d91ad522550d6a2744abc9f8659879f154005703b2140d260
3
+ size 210068124
condition_encoder/config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AceStepConditionEncoder",
3
+ "_diffusers_version": "0.39.0.dev0",
4
+ "attention_bias": false,
5
+ "attention_dropout": 0.0,
6
+ "head_dim": 128,
7
+ "hidden_size": 2048,
8
+ "intermediate_size": 6144,
9
+ "layer_types": null,
10
+ "num_attention_heads": 16,
11
+ "num_key_value_heads": 8,
12
+ "num_lyric_encoder_hidden_layers": 8,
13
+ "num_timbre_encoder_hidden_layers": 4,
14
+ "rms_norm_eps": 1e-06,
15
+ "rope_theta": 1000000,
16
+ "sliding_window": 128,
17
+ "text_hidden_dim": 1024,
18
+ "timbre_hidden_dim": 64
19
+ }
condition_encoder/diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac9cf5e23bfeb2de0d0876ab5b836f5d05b4c8bd22455cc548bd2a4f56b49fab
3
+ size 1218676312
model_index.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AceStepPipeline",
3
+ "_diffusers_version": "0.39.0.dev0",
4
+ "audio_token_detokenizer": [
5
+ "ace_step",
6
+ "AceStepAudioTokenDetokenizer"
7
+ ],
8
+ "audio_tokenizer": [
9
+ "ace_step",
10
+ "AceStepAudioTokenizer"
11
+ ],
12
+ "condition_encoder": [
13
+ "ace_step",
14
+ "AceStepConditionEncoder"
15
+ ],
16
+ "scheduler": [
17
+ "diffusers",
18
+ "FlowMatchEulerDiscreteScheduler"
19
+ ],
20
+ "text_encoder": [
21
+ "transformers",
22
+ "Qwen3Model"
23
+ ],
24
+ "tokenizer": [
25
+ "transformers",
26
+ "Qwen2Tokenizer"
27
+ ],
28
+ "transformer": [
29
+ "diffusers",
30
+ "AceStepTransformer1DModel"
31
+ ],
32
+ "vae": [
33
+ "diffusers",
34
+ "AutoencoderOobleck"
35
+ ]
36
+ }
scheduler/scheduler_config.json ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "FlowMatchEulerDiscreteScheduler",
3
+ "_diffusers_version": "0.39.0.dev0",
4
+ "base_image_seq_len": 256,
5
+ "base_shift": 0.5,
6
+ "invert_sigmas": false,
7
+ "max_image_seq_len": 4096,
8
+ "max_shift": 1.15,
9
+ "num_train_timesteps": 1,
10
+ "shift": 1.0,
11
+ "shift_terminal": null,
12
+ "stochastic_sampling": false,
13
+ "time_shift_type": "exponential",
14
+ "use_beta_sigmas": false,
15
+ "use_dynamic_shifting": false,
16
+ "use_exponential_sigmas": false,
17
+ "use_karras_sigmas": false
18
+ }
silence_latent.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a778e9dd942f5e8b2c09c55370782d318834432b03dabbcdf70e6ed49ad6358b
3
+ size 3841215
text_encoder/config.json ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen3Model"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 151643,
8
+ "dtype": "bfloat16",
9
+ "eos_token_id": 151643,
10
+ "head_dim": 128,
11
+ "hidden_act": "silu",
12
+ "hidden_size": 1024,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 3072,
15
+ "layer_types": [
16
+ "full_attention",
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention"
44
+ ],
45
+ "max_position_embeddings": 32768,
46
+ "max_window_layers": 28,
47
+ "model_type": "qwen3",
48
+ "num_attention_heads": 16,
49
+ "num_hidden_layers": 28,
50
+ "num_key_value_heads": 8,
51
+ "pad_token_id": null,
52
+ "rms_norm_eps": 1e-06,
53
+ "rope_parameters": {
54
+ "rope_theta": 1000000,
55
+ "rope_type": "default"
56
+ },
57
+ "sliding_window": null,
58
+ "tie_word_embeddings": true,
59
+ "transformers_version": "5.5.4",
60
+ "use_cache": true,
61
+ "use_sliding_window": false,
62
+ "vocab_size": 151669
63
+ }
text_encoder/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0437e45c94563b09e13cb7a64478fc406947a93cb34a7e05870fc8dcd48e23fd
3
+ size 1191586416
tokenizer/chat_template.jinja ADDED
@@ -0,0 +1,85 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools %}
2
+ {{- '<|im_start|>system\n' }}
3
+ {%- if messages[0].role == 'system' %}
4
+ {{- messages[0].content + '\n\n' }}
5
+ {%- endif %}
6
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
7
+ {%- for tool in tools %}
8
+ {{- "\n" }}
9
+ {{- tool | tojson }}
10
+ {%- endfor %}
11
+ {{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
12
+ {%- else %}
13
+ {%- if messages[0].role == 'system' %}
14
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
15
+ {%- endif %}
16
+ {%- endif %}
17
+ {%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
18
+ {%- for message in messages[::-1] %}
19
+ {%- set index = (messages|length - 1) - loop.index0 %}
20
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('<tool_response>') and message.content.endswith('</tool_response>')) %}
21
+ {%- set ns.multi_step_tool = false %}
22
+ {%- set ns.last_query_index = index %}
23
+ {%- endif %}
24
+ {%- endfor %}
25
+ {%- for message in messages %}
26
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
27
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
28
+ {%- elif message.role == "assistant" %}
29
+ {%- set content = message.content %}
30
+ {%- set reasoning_content = '' %}
31
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
32
+ {%- set reasoning_content = message.reasoning_content %}
33
+ {%- else %}
34
+ {%- if '</think>' in message.content %}
35
+ {%- set content = message.content.split('</think>')[-1].lstrip('\n') %}
36
+ {%- set reasoning_content = message.content.split('</think>')[0].rstrip('\n').split('<think>')[-1].lstrip('\n') %}
37
+ {%- endif %}
38
+ {%- endif %}
39
+ {%- if loop.index0 > ns.last_query_index %}
40
+ {%- if loop.last or (not loop.last and reasoning_content) %}
41
+ {{- '<|im_start|>' + message.role + '\n<think>\n' + reasoning_content.strip('\n') + '\n</think>\n\n' + content.lstrip('\n') }}
42
+ {%- else %}
43
+ {{- '<|im_start|>' + message.role + '\n' + content }}
44
+ {%- endif %}
45
+ {%- else %}
46
+ {{- '<|im_start|>' + message.role + '\n' + content }}
47
+ {%- endif %}
48
+ {%- if message.tool_calls %}
49
+ {%- for tool_call in message.tool_calls %}
50
+ {%- if (loop.first and content) or (not loop.first) %}
51
+ {{- '\n' }}
52
+ {%- endif %}
53
+ {%- if tool_call.function %}
54
+ {%- set tool_call = tool_call.function %}
55
+ {%- endif %}
56
+ {{- '<tool_call>\n{"name": "' }}
57
+ {{- tool_call.name }}
58
+ {{- '", "arguments": ' }}
59
+ {%- if tool_call.arguments is string %}
60
+ {{- tool_call.arguments }}
61
+ {%- else %}
62
+ {{- tool_call.arguments | tojson }}
63
+ {%- endif %}
64
+ {{- '}\n</tool_call>' }}
65
+ {%- endfor %}
66
+ {%- endif %}
67
+ {{- '<|im_end|>\n' }}
68
+ {%- elif message.role == "tool" %}
69
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
70
+ {{- '<|im_start|>user' }}
71
+ {%- endif %}
72
+ {{- '\n<tool_response>\n' }}
73
+ {{- message.content }}
74
+ {{- '\n</tool_response>' }}
75
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
76
+ {{- '<|im_end|>\n' }}
77
+ {%- endif %}
78
+ {%- endif %}
79
+ {%- endfor %}
80
+ {%- if add_generation_prompt %}
81
+ {{- '<|im_start|>assistant\n' }}
82
+ {%- if enable_thinking is defined and enable_thinking is false %}
83
+ {{- '<think>\n\n</think>\n\n' }}
84
+ {%- endif %}
85
+ {%- endif %}
tokenizer/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93623af029cdc69b87f2864d3b2cc2424fdf16684f15e139b5b9d08ec34ced91
3
+ size 11423701
tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "backend": "tokenizers",
4
+ "bos_token": null,
5
+ "clean_up_tokenization_spaces": false,
6
+ "eos_token": "<|im_end|>",
7
+ "errors": "replace",
8
+ "is_local": true,
9
+ "model_max_length": 131072,
10
+ "pad_token": "<|endoftext|>",
11
+ "split_special_tokens": false,
12
+ "tokenizer_class": "Qwen2Tokenizer",
13
+ "unk_token": null
14
+ }
transformer/config.json ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AceStepTransformer1DModel",
3
+ "_diffusers_version": "0.39.0.dev0",
4
+ "attention_bias": false,
5
+ "attention_dropout": 0.0,
6
+ "audio_acoustic_hidden_dim": 64,
7
+ "encoder_hidden_size": 2048,
8
+ "head_dim": 128,
9
+ "hidden_size": 2048,
10
+ "in_channels": 192,
11
+ "intermediate_size": 6144,
12
+ "is_turbo": false,
13
+ "layer_types": [
14
+ "sliding_attention",
15
+ "full_attention",
16
+ "sliding_attention",
17
+ "full_attention",
18
+ "sliding_attention",
19
+ "full_attention",
20
+ "sliding_attention",
21
+ "full_attention",
22
+ "sliding_attention",
23
+ "full_attention",
24
+ "sliding_attention",
25
+ "full_attention",
26
+ "sliding_attention",
27
+ "full_attention",
28
+ "sliding_attention",
29
+ "full_attention",
30
+ "sliding_attention",
31
+ "full_attention",
32
+ "sliding_attention",
33
+ "full_attention",
34
+ "sliding_attention",
35
+ "full_attention",
36
+ "sliding_attention",
37
+ "full_attention"
38
+ ],
39
+ "model_version": null,
40
+ "num_attention_heads": 16,
41
+ "num_hidden_layers": 24,
42
+ "num_key_value_heads": 8,
43
+ "patch_size": 2,
44
+ "rms_norm_eps": 1e-06,
45
+ "rope_theta": 1000000,
46
+ "sliding_window": 128
47
+ }
transformer/diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d19d04152d3e261d52d64c1b02cffe6eb0f0e585646ec8982bffd1bf3a57ba8
3
+ size 3150968664
vae/config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderOobleck",
3
+ "_diffusers_version": "0.39.0.dev0",
4
+ "_name_or_path": "train_model/huggingface/models--ACE-Step--Ace-Step1.5/snapshots/19671f406d603126926c1b7e2adc169acbcade22/vae",
5
+ "audio_channels": 2,
6
+ "channel_multiples": [
7
+ 1,
8
+ 2,
9
+ 4,
10
+ 8,
11
+ 16
12
+ ],
13
+ "decoder_channels": 128,
14
+ "decoder_input_channels": 64,
15
+ "downsampling_ratios": [
16
+ 2,
17
+ 4,
18
+ 4,
19
+ 6,
20
+ 10
21
+ ],
22
+ "encoder_hidden_size": 128,
23
+ "sampling_rate": 48000
24
+ }
vae/diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da17edb604c40deaf09e9b24974e590d1ca83a374070e5d0884cfa4bed9a99b0
3
+ size 337431388