ylacombe HF staff commited on
Commit
fda967b
1 Parent(s): b192d44

ed7f5fa954a9c6588d225487406f90ca22511e1ef02f7399ee2a44117dd709d0

Browse files
Files changed (1) hide show
  1. config.json +99 -15
config.json CHANGED
@@ -81,11 +81,105 @@
81
  "use_bfloat16": false,
82
  "use_cache": true
83
  },
84
- "coarse_infer_token": 12050,
85
- "coarse_rate_hz": 75,
86
- "coarse_semantic_pad_token": 12048,
87
- "codebook_size": 1024,
88
- "context_window_size": 1024,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
89
  "fine_acoustics_config": {
90
  "_name_or_path": "",
91
  "add_cross_attention": false,
@@ -165,10 +259,6 @@
165
  "use_cache": true
166
  },
167
  "model_type": "bark",
168
- "n_coarse_codebooks": 2,
169
- "n_fine_codebooks": 8,
170
- "pretrained_encodec_name_or_path": "facebook/encodec_24khz",
171
- "sample_rate": 24000,
172
  "semantic_config": {
173
  "_name_or_path": "",
174
  "add_cross_attention": false,
@@ -247,12 +337,6 @@
247
  "use_bfloat16": false,
248
  "use_cache": true
249
  },
250
- "semantic_infer_token": 129599,
251
- "semantic_pad_token": 10000,
252
- "semantic_rate_hz": 49.9,
253
- "semantic_vocab_size": 10000,
254
- "text_encoding_offset": 10048,
255
- "text_pad_token": 129595,
256
  "torch_dtype": "float32",
257
  "transformers_version": null
258
  }
 
81
  "use_bfloat16": false,
82
  "use_cache": true
83
  },
84
+ "codec_config": {
85
+ "_name_or_path": "ArthurZ/encodec_24khz",
86
+ "add_cross_attention": false,
87
+ "architectures": [
88
+ "EncodecModel"
89
+ ],
90
+ "audio_channels": 1,
91
+ "bad_words_ids": null,
92
+ "begin_suppress_tokens": null,
93
+ "bos_token_id": null,
94
+ "chunk_length_s": null,
95
+ "chunk_size_feed_forward": 0,
96
+ "codebook_dim": 128,
97
+ "codebook_size": 1024,
98
+ "compress": 2,
99
+ "cross_attention_hidden_size": null,
100
+ "decoder_start_token_id": null,
101
+ "dilation_growth_rate": 2,
102
+ "diversity_penalty": 0.0,
103
+ "do_sample": false,
104
+ "early_stopping": false,
105
+ "encoder_no_repeat_ngram_size": 0,
106
+ "eos_token_id": null,
107
+ "exponential_decay_length_penalty": null,
108
+ "finetuning_task": null,
109
+ "forced_bos_token_id": null,
110
+ "forced_eos_token_id": null,
111
+ "hidden_size": 128,
112
+ "id2label": {
113
+ "0": "LABEL_0",
114
+ "1": "LABEL_1"
115
+ },
116
+ "is_decoder": false,
117
+ "is_encoder_decoder": false,
118
+ "kernel_size": 7,
119
+ "label2id": {
120
+ "LABEL_0": 0,
121
+ "LABEL_1": 1
122
+ },
123
+ "last_kernel_size": 7,
124
+ "length_penalty": 1.0,
125
+ "max_length": 20,
126
+ "min_length": 0,
127
+ "model_type": "encodec",
128
+ "no_repeat_ngram_size": 0,
129
+ "norm_type": "weight_norm",
130
+ "normalize": false,
131
+ "num_beam_groups": 1,
132
+ "num_beams": 1,
133
+ "num_filters": 32,
134
+ "num_lstm_layers": 2,
135
+ "num_residual_layers": 1,
136
+ "num_return_sequences": 1,
137
+ "output_attentions": false,
138
+ "output_hidden_states": false,
139
+ "output_scores": false,
140
+ "overlap": null,
141
+ "pad_mode": "reflect",
142
+ "pad_token_id": null,
143
+ "prefix": null,
144
+ "problem_type": null,
145
+ "pruned_heads": {},
146
+ "remove_invalid_values": false,
147
+ "repetition_penalty": 1.0,
148
+ "residual_kernel_size": 3,
149
+ "return_dict": true,
150
+ "return_dict_in_generate": false,
151
+ "sampling_rate": 24000,
152
+ "sep_token_id": null,
153
+ "suppress_tokens": null,
154
+ "target_bandwidths": [
155
+ 1.5,
156
+ 3.0,
157
+ 6.0,
158
+ 12.0,
159
+ 24.0
160
+ ],
161
+ "task_specific_params": null,
162
+ "temperature": 1.0,
163
+ "tf_legacy_loss": false,
164
+ "tie_encoder_decoder": false,
165
+ "tie_word_embeddings": true,
166
+ "tokenizer_class": null,
167
+ "top_k": 50,
168
+ "top_p": 1.0,
169
+ "torch_dtype": "float32",
170
+ "torchscript": false,
171
+ "transformers_version": "4.31.0.dev0",
172
+ "trim_right_ratio": 1.0,
173
+ "typical_p": 1.0,
174
+ "upsampling_ratios": [
175
+ 8,
176
+ 5,
177
+ 4,
178
+ 2
179
+ ],
180
+ "use_bfloat16": false,
181
+ "use_causal_conv": true
182
+ },
183
  "fine_acoustics_config": {
184
  "_name_or_path": "",
185
  "add_cross_attention": false,
 
259
  "use_cache": true
260
  },
261
  "model_type": "bark",
 
 
 
 
262
  "semantic_config": {
263
  "_name_or_path": "",
264
  "add_cross_attention": false,
 
337
  "use_bfloat16": false,
338
  "use_cache": true
339
  },
 
 
 
 
 
 
340
  "torch_dtype": "float32",
341
  "transformers_version": null
342
  }