ylacombe HF staff commited on
Commit
6eb35c9
1 Parent(s): 568882c

Upload MusicgenMelodyForConditionalGeneration

Browse files
config.json CHANGED
@@ -105,7 +105,7 @@
105
  "add_cross_attention": false,
106
  "architectures": null,
107
  "attention_dropout": 0.0,
108
- "audio_channels": 1,
109
  "bad_words_ids": null,
110
  "begin_suppress_tokens": null,
111
  "bos_token_id": 2048,
@@ -145,7 +145,7 @@
145
  "num_attention_heads": 24,
146
  "num_beam_groups": 1,
147
  "num_beams": 1,
148
- "num_codebooks": 4,
149
  "num_hidden_layers": 48,
150
  "num_return_sequences": 1,
151
  "output_attentions": false,
 
105
  "add_cross_attention": false,
106
  "architectures": null,
107
  "attention_dropout": 0.0,
108
+ "audio_channels": 2,
109
  "bad_words_ids": null,
110
  "begin_suppress_tokens": null,
111
  "bos_token_id": 2048,
 
145
  "num_attention_heads": 24,
146
  "num_beam_groups": 1,
147
  "num_beams": 1,
148
+ "num_codebooks": 8,
149
  "num_hidden_layers": 48,
150
  "num_return_sequences": 1,
151
  "output_attentions": false,
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4d60f540d8a15ae3c50f63f5d4f646c0dd4bfac6442c22e481117554d0465062
3
- size 4966269528
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e20ca49644e26560e90ccf54557852c59c13e8863337bd1562fd906ef0400515
3
+ size 4997739056
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:01268b3a5f9819026a9ee6b76df35574e7dd521de3e6431c7fad16b44ce5a9c0
3
- size 1263380400
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95932e81bf5b2b02777a245a42dc65b9ab13362460ae77f7a1c3c0b50737e3c6
3
+ size 1332599632
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 6229559576
4
  },
5
  "weight_map": {
6
  "audio_enc_to_dec_proj.bias": "model-00002-of-00002.safetensors",
@@ -125,11 +125,19 @@
125
  "decoder.lm_heads.1.weight": "model-00002-of-00002.safetensors",
126
  "decoder.lm_heads.2.weight": "model-00002-of-00002.safetensors",
127
  "decoder.lm_heads.3.weight": "model-00002-of-00002.safetensors",
 
 
 
 
128
  "decoder.model.decoder.embed_positions.weights": "model-00001-of-00002.safetensors",
129
  "decoder.model.decoder.embed_tokens.0.weight": "model-00001-of-00002.safetensors",
130
  "decoder.model.decoder.embed_tokens.1.weight": "model-00001-of-00002.safetensors",
131
  "decoder.model.decoder.embed_tokens.2.weight": "model-00001-of-00002.safetensors",
132
  "decoder.model.decoder.embed_tokens.3.weight": "model-00001-of-00002.safetensors",
 
 
 
 
133
  "decoder.model.decoder.layer_norm.bias": "model-00002-of-00002.safetensors",
134
  "decoder.model.decoder.layer_norm.weight": "model-00002-of-00002.safetensors",
135
  "decoder.model.decoder.layers.0.fc1.weight": "model-00001-of-00002.safetensors",
@@ -447,11 +455,11 @@
447
  "decoder.model.decoder.layers.37.final_layer_norm.bias": "model-00002-of-00002.safetensors",
448
  "decoder.model.decoder.layers.37.final_layer_norm.weight": "model-00002-of-00002.safetensors",
449
  "decoder.model.decoder.layers.37.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
450
- "decoder.model.decoder.layers.37.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
451
- "decoder.model.decoder.layers.37.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
452
  "decoder.model.decoder.layers.37.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
453
- "decoder.model.decoder.layers.37.self_attn_layer_norm.bias": "model-00001-of-00002.safetensors",
454
- "decoder.model.decoder.layers.37.self_attn_layer_norm.weight": "model-00001-of-00002.safetensors",
455
  "decoder.model.decoder.layers.38.fc1.weight": "model-00002-of-00002.safetensors",
456
  "decoder.model.decoder.layers.38.fc2.weight": "model-00002-of-00002.safetensors",
457
  "decoder.model.decoder.layers.38.final_layer_norm.bias": "model-00002-of-00002.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 6330247448
4
  },
5
  "weight_map": {
6
  "audio_enc_to_dec_proj.bias": "model-00002-of-00002.safetensors",
 
125
  "decoder.lm_heads.1.weight": "model-00002-of-00002.safetensors",
126
  "decoder.lm_heads.2.weight": "model-00002-of-00002.safetensors",
127
  "decoder.lm_heads.3.weight": "model-00002-of-00002.safetensors",
128
+ "decoder.lm_heads.4.weight": "model-00002-of-00002.safetensors",
129
+ "decoder.lm_heads.5.weight": "model-00002-of-00002.safetensors",
130
+ "decoder.lm_heads.6.weight": "model-00002-of-00002.safetensors",
131
+ "decoder.lm_heads.7.weight": "model-00002-of-00002.safetensors",
132
  "decoder.model.decoder.embed_positions.weights": "model-00001-of-00002.safetensors",
133
  "decoder.model.decoder.embed_tokens.0.weight": "model-00001-of-00002.safetensors",
134
  "decoder.model.decoder.embed_tokens.1.weight": "model-00001-of-00002.safetensors",
135
  "decoder.model.decoder.embed_tokens.2.weight": "model-00001-of-00002.safetensors",
136
  "decoder.model.decoder.embed_tokens.3.weight": "model-00001-of-00002.safetensors",
137
+ "decoder.model.decoder.embed_tokens.4.weight": "model-00001-of-00002.safetensors",
138
+ "decoder.model.decoder.embed_tokens.5.weight": "model-00001-of-00002.safetensors",
139
+ "decoder.model.decoder.embed_tokens.6.weight": "model-00001-of-00002.safetensors",
140
+ "decoder.model.decoder.embed_tokens.7.weight": "model-00001-of-00002.safetensors",
141
  "decoder.model.decoder.layer_norm.bias": "model-00002-of-00002.safetensors",
142
  "decoder.model.decoder.layer_norm.weight": "model-00002-of-00002.safetensors",
143
  "decoder.model.decoder.layers.0.fc1.weight": "model-00001-of-00002.safetensors",
 
455
  "decoder.model.decoder.layers.37.final_layer_norm.bias": "model-00002-of-00002.safetensors",
456
  "decoder.model.decoder.layers.37.final_layer_norm.weight": "model-00002-of-00002.safetensors",
457
  "decoder.model.decoder.layers.37.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
458
+ "decoder.model.decoder.layers.37.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
459
+ "decoder.model.decoder.layers.37.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
460
  "decoder.model.decoder.layers.37.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
461
+ "decoder.model.decoder.layers.37.self_attn_layer_norm.bias": "model-00002-of-00002.safetensors",
462
+ "decoder.model.decoder.layers.37.self_attn_layer_norm.weight": "model-00002-of-00002.safetensors",
463
  "decoder.model.decoder.layers.38.fc1.weight": "model-00002-of-00002.safetensors",
464
  "decoder.model.decoder.layers.38.fc2.weight": "model-00002-of-00002.safetensors",
465
  "decoder.model.decoder.layers.38.final_layer_norm.bias": "model-00002-of-00002.safetensors",