ylacombe HF staff commited on
Commit
bbffab3
1 Parent(s): 998bfeb

Model save

Browse files
README.md CHANGED
@@ -2,8 +2,6 @@
2
  license: cc-by-nc-4.0
3
  library_name: peft
4
  tags:
5
- - text-to-audio
6
- - ylacombe/tiny-punk
7
  - generated_from_trainer
8
  base_model: facebook/musicgen-melody
9
  model-index:
@@ -14,12 +12,11 @@ model-index:
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
  should probably proofread and complete it, then remove this comment. -->
16
 
 
 
17
  # musicgen-melody-lora-punk
18
 
19
- This model is a fine-tuned version of [facebook/musicgen-melody](https://huggingface.co/facebook/musicgen-melody) on the YLACOMBE/TINY-PUNK - DEFAULT dataset.
20
- It achieves the following results on the evaluation set:
21
- - Loss: 4.7288
22
- - Clap: 0.2581
23
 
24
  ## Model description
25
 
@@ -56,7 +53,7 @@ The following hyperparameters were used during training:
56
  ### Framework versions
57
 
58
  - PEFT 0.10.0
59
- - Transformers 4.40.0.dev0
60
  - Pytorch 2.1.2+cu121
61
- - Datasets 2.18.0
62
  - Tokenizers 0.19.1
 
2
  license: cc-by-nc-4.0
3
  library_name: peft
4
  tags:
 
 
5
  - generated_from_trainer
6
  base_model: facebook/musicgen-melody
7
  model-index:
 
12
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
  should probably proofread and complete it, then remove this comment. -->
14
 
15
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/ylacombe/huggingface/runs/fibz5z40)
16
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/ylacombe/huggingface/runs/fibz5z40)
17
  # musicgen-melody-lora-punk
18
 
19
+ This model is a fine-tuned version of [facebook/musicgen-melody](https://huggingface.co/facebook/musicgen-melody) on an unknown dataset.
 
 
 
20
 
21
  ## Model description
22
 
 
53
  ### Framework versions
54
 
55
  - PEFT 0.10.0
56
+ - Transformers 4.41.0.dev0
57
  - Pytorch 2.1.2+cu121
58
+ - Datasets 2.19.0
59
  - Tokenizers 0.19.1
adapter_config.json CHANGED
@@ -23,22 +23,22 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
 
26
  "out_proj",
27
- "lm_heads.2",
28
- "embed_tokens.0",
29
- "embed_tokens.2",
30
- "fc1",
31
- "embed_tokens.1",
32
  "audio_enc_to_dec_proj",
33
- "v_proj",
34
- "lm_heads.0",
35
- "embed_tokens.3",
36
  "enc_to_dec_proj",
 
 
37
  "lm_heads.3",
38
- "q_proj",
39
  "k_proj",
40
- "lm_heads.1",
41
- "fc2"
 
 
 
 
42
  ],
43
  "task_type": null,
44
  "use_dora": false,
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
+ "fc2",
27
  "out_proj",
 
 
 
 
 
28
  "audio_enc_to_dec_proj",
29
+ "lm_heads.1",
 
 
30
  "enc_to_dec_proj",
31
+ "embed_tokens.1",
32
+ "embed_tokens.0",
33
  "lm_heads.3",
34
+ "lm_heads.2",
35
  "k_proj",
36
+ "embed_tokens.3",
37
+ "embed_tokens.2",
38
+ "v_proj",
39
+ "q_proj",
40
+ "fc1",
41
+ "lm_heads.0"
42
  ],
43
  "task_type": null,
44
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6f8c8b2a7fab93911c35795ccf207946125da8511d14f46a5356b2fdcf9cf93b
3
  size 87103456
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:27de65e63aa3e9805c34c5625f400c03e527dcd5d7702695b8dbe0cdb6695191
3
  size 87103456
all_results.json CHANGED
@@ -1,15 +1,15 @@
1
  {
2
  "epoch": 3.764705882352941,
3
- "eval_clap": 0.2581212818622589,
4
- "eval_loss": 4.728785514831543,
5
- "eval_runtime": 140.1969,
6
  "eval_samples": 5,
7
- "eval_samples_per_second": 0.036,
8
- "eval_steps_per_second": 0.036,
9
- "total_flos": 35328595626816.0,
10
- "train_loss": 8.98837435245514,
11
- "train_runtime": 239.7476,
12
  "train_samples": 33,
13
- "train_samples_per_second": 0.551,
14
- "train_steps_per_second": 0.033
15
  }
 
1
  {
2
  "epoch": 3.764705882352941,
3
+ "eval_clap": -0.034163739532232285,
4
+ "eval_loss": 4.7287702560424805,
5
+ "eval_runtime": 95.1156,
6
  "eval_samples": 5,
7
+ "eval_samples_per_second": 0.053,
8
+ "eval_steps_per_second": 0.053,
9
+ "total_flos": 3464493556500.0,
10
+ "train_loss": 8.955864429473877,
11
+ "train_runtime": 45.9597,
12
  "train_samples": 33,
13
+ "train_samples_per_second": 2.872,
14
+ "train_steps_per_second": 0.174
15
  }
config.json CHANGED
@@ -295,5 +295,5 @@
295
  "vocab_size": 32128
296
  },
297
  "torch_dtype": "float32",
298
- "transformers_version": "4.40.0.dev0"
299
  }
 
295
  "vocab_size": 32128
296
  },
297
  "torch_dtype": "float32",
298
+ "transformers_version": "4.41.0.dev0"
299
  }
eval_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "epoch": 3.764705882352941,
3
- "eval_clap": 0.2581212818622589,
4
- "eval_loss": 4.728785514831543,
5
- "eval_runtime": 140.1969,
6
  "eval_samples": 5,
7
- "eval_samples_per_second": 0.036,
8
- "eval_steps_per_second": 0.036
9
  }
 
1
  {
2
  "epoch": 3.764705882352941,
3
+ "eval_clap": -0.034163739532232285,
4
+ "eval_loss": 4.7287702560424805,
5
+ "eval_runtime": 95.1156,
6
  "eval_samples": 5,
7
+ "eval_samples_per_second": 0.053,
8
+ "eval_steps_per_second": 0.053
9
  }
train_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "epoch": 3.764705882352941,
3
- "total_flos": 35328595626816.0,
4
- "train_loss": 8.98837435245514,
5
- "train_runtime": 239.7476,
6
  "train_samples": 33,
7
- "train_samples_per_second": 0.551,
8
- "train_steps_per_second": 0.033
9
  }
 
1
  {
2
  "epoch": 3.764705882352941,
3
+ "total_flos": 3464493556500.0,
4
+ "train_loss": 8.955864429473877,
5
+ "train_runtime": 45.9597,
6
  "train_samples": 33,
7
+ "train_samples_per_second": 2.872,
8
+ "train_steps_per_second": 0.174
9
  }
trainer_state.json CHANGED
@@ -10,76 +10,76 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.47058823529411764,
13
- "grad_norm": 1.3573589324951172,
14
  "learning_rate": 0.000175,
15
- "loss": 9.573,
16
  "step": 1
17
  },
18
  {
19
  "epoch": 0.9411764705882353,
20
- "grad_norm": 1.2206748723983765,
21
  "learning_rate": 0.00015000000000000001,
22
- "loss": 9.4647,
23
  "step": 2
24
  },
25
  {
26
  "epoch": 1.4117647058823528,
27
- "grad_norm": 1.1659976243972778,
28
  "learning_rate": 0.000125,
29
- "loss": 9.2541,
30
  "step": 3
31
  },
32
  {
33
  "epoch": 1.8823529411764706,
34
- "grad_norm": 1.3362257480621338,
35
  "learning_rate": 0.0001,
36
- "loss": 9.0185,
37
  "step": 4
38
  },
39
  {
40
  "epoch": 2.3529411764705883,
41
- "grad_norm": 1.568061351776123,
42
  "learning_rate": 7.500000000000001e-05,
43
- "loss": 8.8428,
44
  "step": 5
45
  },
46
  {
47
  "epoch": 2.8235294117647056,
48
- "grad_norm": 1.6825159788131714,
49
  "learning_rate": 5e-05,
50
- "loss": 8.7436,
51
  "step": 6
52
  },
53
  {
54
  "epoch": 3.2941176470588234,
55
- "grad_norm": 2.018841028213501,
56
  "learning_rate": 2.5e-05,
57
- "loss": 8.5273,
58
  "step": 7
59
  },
60
  {
61
  "epoch": 3.764705882352941,
62
- "grad_norm": 2.155460834503174,
63
  "learning_rate": 0.0,
64
- "loss": 8.483,
65
  "step": 8
66
  },
67
  {
68
  "epoch": 3.764705882352941,
69
  "step": 8,
70
- "total_flos": 35328595626816.0,
71
- "train_loss": 8.98837435245514,
72
- "train_runtime": 239.7476,
73
- "train_samples_per_second": 0.551,
74
- "train_steps_per_second": 0.033
75
  }
76
  ],
77
- "logging_steps": 1,
78
  "max_steps": 8,
79
  "num_input_tokens_seen": 0,
80
  "num_train_epochs": 4,
81
  "save_steps": 500,
82
- "total_flos": 35328595626816.0,
83
  "train_batch_size": 2,
84
  "trial_name": null,
85
  "trial_params": null
 
10
  "log_history": [
11
  {
12
  "epoch": 0.47058823529411764,
13
+ "grad_norm": 1.1788710355758667,
14
  "learning_rate": 0.000175,
15
+ "loss": 9.5576,
16
  "step": 1
17
  },
18
  {
19
  "epoch": 0.9411764705882353,
20
+ "grad_norm": 1.1743119955062866,
21
  "learning_rate": 0.00015000000000000001,
22
+ "loss": 9.422,
23
  "step": 2
24
  },
25
  {
26
  "epoch": 1.4117647058823528,
27
+ "grad_norm": 1.1545751094818115,
28
  "learning_rate": 0.000125,
29
+ "loss": 9.2273,
30
  "step": 3
31
  },
32
  {
33
  "epoch": 1.8823529411764706,
34
+ "grad_norm": 1.3188327550888062,
35
  "learning_rate": 0.0001,
36
+ "loss": 8.9962,
37
  "step": 4
38
  },
39
  {
40
  "epoch": 2.3529411764705883,
41
+ "grad_norm": 1.6398828029632568,
42
  "learning_rate": 7.500000000000001e-05,
43
+ "loss": 8.8142,
44
  "step": 5
45
  },
46
  {
47
  "epoch": 2.8235294117647056,
48
+ "grad_norm": 1.806210994720459,
49
  "learning_rate": 5e-05,
50
+ "loss": 8.7001,
51
  "step": 6
52
  },
53
  {
54
  "epoch": 3.2941176470588234,
55
+ "grad_norm": 2.153493881225586,
56
  "learning_rate": 2.5e-05,
57
+ "loss": 8.4994,
58
  "step": 7
59
  },
60
  {
61
  "epoch": 3.764705882352941,
62
+ "grad_norm": 2.2270724773406982,
63
  "learning_rate": 0.0,
64
+ "loss": 8.4301,
65
  "step": 8
66
  },
67
  {
68
  "epoch": 3.764705882352941,
69
  "step": 8,
70
+ "total_flos": 3464493556500.0,
71
+ "train_loss": 8.955864429473877,
72
+ "train_runtime": 45.9597,
73
+ "train_samples_per_second": 2.872,
74
+ "train_steps_per_second": 0.174
75
  }
76
  ],
77
+ "logging_steps": 1.0,
78
  "max_steps": 8,
79
  "num_input_tokens_seen": 0,
80
  "num_train_epochs": 4,
81
  "save_steps": 500,
82
+ "total_flos": 3464493556500.0,
83
  "train_batch_size": 2,
84
  "trial_name": null,
85
  "trial_params": null
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6009f07836a646f9d53a044c08722604b2902c4c2f06c0193166fd870efdc120
3
- size 5112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:494294fe4c26ed33c87c68afd6fe7edca29e0c4e94f126655fc837e7ed49caf4
3
+ size 5176