DevPanda004 commited on
Commit
13d4b8e
1 Parent(s): 869413b

Training in progress, step 30

Browse files
adapter_config.json CHANGED
@@ -1,11 +1,13 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": {
4
- "base_model_class": "MusicgenMelodyForConditionalGeneration",
5
- "parent_library": "transformers.models.musicgen_melody.modeling_musicgen_melody"
6
  },
7
- "base_model_name_or_path": "facebook/musicgen-melody",
8
  "bias": "none",
 
 
9
  "fan_in_fan_out": false,
10
  "inference_mode": true,
11
  "init_lora_weights": true,
@@ -14,6 +16,7 @@
14
  "layers_to_transform": null,
15
  "loftq_config": {},
16
  "lora_alpha": 16,
 
17
  "lora_dropout": 0.05,
18
  "megatron_config": null,
19
  "megatron_core": "megatron.core",
@@ -23,22 +26,22 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
- "out_proj",
 
 
 
27
  "q_proj",
28
- "lm_heads.0",
 
29
  "audio_enc_to_dec_proj",
 
30
  "embed_tokens.3",
31
- "fc2",
 
32
  "embed_tokens.2",
 
33
  "fc1",
34
- "lm_heads.2",
35
- "k_proj",
36
- "lm_heads.1",
37
- "embed_tokens.1",
38
- "v_proj",
39
- "lm_heads.3",
40
- "embed_tokens.0",
41
- "enc_to_dec_proj"
42
  ],
43
  "task_type": null,
44
  "use_dora": false,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": {
4
+ "base_model_class": "PeftModel",
5
+ "parent_library": "peft.peft_model"
6
  },
7
+ "base_model_name_or_path": null,
8
  "bias": "none",
9
+ "eva_config": null,
10
+ "exclude_modules": null,
11
  "fan_in_fan_out": false,
12
  "inference_mode": true,
13
  "init_lora_weights": true,
 
16
  "layers_to_transform": null,
17
  "loftq_config": {},
18
  "lora_alpha": 16,
19
+ "lora_bias": false,
20
  "lora_dropout": 0.05,
21
  "megatron_config": null,
22
  "megatron_core": "megatron.core",
 
26
  "rank_pattern": {},
27
  "revision": null,
28
  "target_modules": [
29
+ "enc_to_dec_proj",
30
+ "lm_heads.3",
31
+ "embed_tokens.0",
32
+ "v_proj",
33
  "q_proj",
34
+ "lm_heads.2",
35
+ "out_proj",
36
  "audio_enc_to_dec_proj",
37
+ "lm_heads.1",
38
  "embed_tokens.3",
39
+ "lm_heads.0",
40
+ "embed_tokens.1",
41
  "embed_tokens.2",
42
+ "fc2",
43
  "fc1",
44
+ "k_proj"
 
 
 
 
 
 
 
45
  ],
46
  "task_type": null,
47
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bd9be4656a77bb5db3858cf9a56e37f919319f53a4e22a8519c21a119984bb3e
3
- size 87103456
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c99a72f7b079b01fc7223efef55eb5e61323c38ae014fc1d35fd467a811fbd8b
3
+ size 87113584
tokenizer.json CHANGED
@@ -1,7 +1,14 @@
1
  {
2
  "version": "1.0",
3
  "truncation": null,
4
- "padding": null,
 
 
 
 
 
 
 
5
  "added_tokens": [
6
  {
7
  "id": 0,
 
1
  {
2
  "version": "1.0",
3
  "truncation": null,
4
+ "padding": {
5
+ "strategy": "BatchLongest",
6
+ "direction": "Right",
7
+ "pad_to_multiple_of": null,
8
+ "pad_id": 0,
9
+ "pad_type_id": 0,
10
+ "pad_token": "<pad>"
11
+ },
12
  "added_tokens": [
13
  {
14
  "id": 0,
tokenizer_config.json CHANGED
@@ -930,6 +930,7 @@
930
  "clean_up_tokenization_spaces": true,
931
  "eos_token": "</s>",
932
  "extra_ids": 100,
 
933
  "max_length": null,
934
  "model_max_length": 512,
935
  "pad_to_multiple_of": null,
 
930
  "clean_up_tokenization_spaces": true,
931
  "eos_token": "</s>",
932
  "extra_ids": 100,
933
+ "extra_special_tokens": {},
934
  "max_length": null,
935
  "model_max_length": 512,
936
  "pad_to_multiple_of": null,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:698a3f013ad287a54ad827f8d0123dedfeae90e5f16059c944789346e604dcac
3
- size 5432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:474032c744046ce77da45153cf18c34e9bc0cf658b971afc86b7f5c3ad3f80d7
3
+ size 5496