adamjweintraut commited on
Commit
16b6c56
1 Parent(s): ff952ee

Training in progress, step 500, checkpoint

Browse files
last-checkpoint/config.json CHANGED
@@ -70,5 +70,5 @@
70
  "torch_dtype": "float32",
71
  "transformers_version": "4.39.3",
72
  "use_cache": true,
73
- "vocab_size": 50265
74
  }
 
70
  "torch_dtype": "float32",
71
  "transformers_version": "4.39.3",
72
  "use_cache": true,
73
+ "vocab_size": 50266
74
  }
last-checkpoint/generation_config.json CHANGED
@@ -1,13 +1,25 @@
1
  {
2
- "_from_model_config": true,
3
  "bos_token_id": 0,
 
4
  "decoder_start_token_id": 2,
 
5
  "early_stopping": true,
6
  "eos_token_id": 2,
7
  "forced_bos_token_id": 0,
8
  "forced_eos_token_id": 2,
 
 
 
9
  "no_repeat_ngram_size": 3,
10
  "num_beams": 4,
11
  "pad_token_id": 1,
12
- "transformers_version": "4.39.3"
 
 
 
 
 
 
 
 
13
  }
 
1
  {
 
2
  "bos_token_id": 0,
3
+ "clean_up_tokenization_spaces": true,
4
  "decoder_start_token_id": 2,
5
+ "do_sample": true,
6
  "early_stopping": true,
7
  "eos_token_id": 2,
8
  "forced_bos_token_id": 0,
9
  "forced_eos_token_id": 2,
10
+ "max_new_tokens": 512,
11
+ "min_new_tokens": 128,
12
+ "n_examples": null,
13
  "no_repeat_ngram_size": 3,
14
  "num_beams": 4,
15
  "pad_token_id": 1,
16
+ "padding": "max_length",
17
+ "renormalize_logits": true,
18
+ "repetition_penalty": 0.6,
19
+ "skip_special_tokens": true,
20
+ "temperature": 0.85,
21
+ "top_k": 0,
22
+ "top_p": 0.9,
23
+ "transformers_version": "4.39.3",
24
+ "truncation": true
25
  }
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a524d8ba515353da172f3ea29869d4f5f00248b6d2240eed45c56334429fd438
3
- size 1625426996
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd0c5d1a06334178c514a152dd882e679196e9868a59af7410748d27894ac32c
3
+ size 1625431096
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f0027089ec21b8d311f17fffcc479548564f54962fcf65bad2bca1f5a26f0401
3
- size 3250759951
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ac5ac3c62d050e51d00061c2322dbd23123261ca56b37d80731f50d946fc74a
3
+ size 3250768143
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3a1cd061e18f9b50a57b16facc0f8e912690855b9236bc6da220d4623bf5e43d
3
  size 14308
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e0a4e9ea4482d0bb626a007aaa253cff387f0e1939224e8a014c99a65084df8
3
  size 14308
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:eed6fc6c0be5639a175fb541ebe63899052d1e643a0aac07601ff1059d2c27f3
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8efdbb74ab678951ea69a6311703aa1542d2a82a0472c9e6d420dde455af8e63
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,140 +1,35 @@
1
  {
2
- "best_metric": 0.8160156011581421,
3
- "best_model_checkpoint": "/content/drive/MyDrive/W210 Capstone - Lyric Generation with Melody/loaf/models/loaf/bart/bart-finetuned-loaf-512-lyrictoplan/checkpoint-4000",
4
- "epoch": 3.5618878005342833,
5
  "eval_steps": 500,
6
- "global_step": 4000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.45,
13
- "grad_norm": 5.676259994506836,
14
- "learning_rate": 4.443455031166518e-05,
15
- "loss": 1.4176,
16
  "step": 500
17
  },
18
  {
19
  "epoch": 0.45,
20
- "eval_loss": 1.1072946786880493,
21
- "eval_runtime": 98.8901,
22
- "eval_samples_per_second": 6.29,
23
- "eval_steps_per_second": 1.578,
24
  "step": 500
25
- },
26
- {
27
- "epoch": 0.89,
28
- "grad_norm": 7.331023216247559,
29
- "learning_rate": 3.886910062333037e-05,
30
- "loss": 1.1228,
31
- "step": 1000
32
- },
33
- {
34
- "epoch": 0.89,
35
- "eval_loss": 0.9914930462837219,
36
- "eval_runtime": 99.0385,
37
- "eval_samples_per_second": 6.28,
38
- "eval_steps_per_second": 1.575,
39
- "step": 1000
40
- },
41
- {
42
- "epoch": 1.34,
43
- "grad_norm": 2.772461414337158,
44
- "learning_rate": 3.330365093499555e-05,
45
- "loss": 0.9568,
46
- "step": 1500
47
- },
48
- {
49
- "epoch": 1.34,
50
- "eval_loss": 0.9346542358398438,
51
- "eval_runtime": 99.0311,
52
- "eval_samples_per_second": 6.281,
53
- "eval_steps_per_second": 1.575,
54
- "step": 1500
55
- },
56
- {
57
- "epoch": 1.78,
58
- "grad_norm": 4.1792778968811035,
59
- "learning_rate": 2.7738201246660732e-05,
60
- "loss": 0.888,
61
- "step": 2000
62
- },
63
- {
64
- "epoch": 1.78,
65
- "eval_loss": 0.8794981241226196,
66
- "eval_runtime": 98.9567,
67
- "eval_samples_per_second": 6.286,
68
- "eval_steps_per_second": 1.576,
69
- "step": 2000
70
- },
71
- {
72
- "epoch": 2.23,
73
- "grad_norm": 2.9945931434631348,
74
- "learning_rate": 2.2172751558325912e-05,
75
- "loss": 0.8179,
76
- "step": 2500
77
- },
78
- {
79
- "epoch": 2.23,
80
- "eval_loss": 0.8773565888404846,
81
- "eval_runtime": 99.5216,
82
- "eval_samples_per_second": 6.25,
83
- "eval_steps_per_second": 1.567,
84
- "step": 2500
85
- },
86
- {
87
- "epoch": 2.67,
88
- "grad_norm": 1.6972075700759888,
89
- "learning_rate": 1.6607301869991096e-05,
90
- "loss": 0.7929,
91
- "step": 3000
92
- },
93
- {
94
- "epoch": 2.67,
95
- "eval_loss": 0.8550799489021301,
96
- "eval_runtime": 99.4863,
97
- "eval_samples_per_second": 6.252,
98
- "eval_steps_per_second": 1.568,
99
- "step": 3000
100
- },
101
- {
102
- "epoch": 3.12,
103
- "grad_norm": 1.3822712898254395,
104
- "learning_rate": 1.1041852181656279e-05,
105
- "loss": 0.7256,
106
- "step": 3500
107
- },
108
- {
109
- "epoch": 3.12,
110
- "eval_loss": 0.8212681412696838,
111
- "eval_runtime": 99.4565,
112
- "eval_samples_per_second": 6.254,
113
- "eval_steps_per_second": 1.569,
114
- "step": 3500
115
- },
116
- {
117
- "epoch": 3.56,
118
- "grad_norm": 2.2662038803100586,
119
- "learning_rate": 5.476402493321461e-06,
120
- "loss": 0.6979,
121
- "step": 4000
122
- },
123
- {
124
- "epoch": 3.56,
125
- "eval_loss": 0.8160156011581421,
126
- "eval_runtime": 99.3184,
127
- "eval_samples_per_second": 6.263,
128
- "eval_steps_per_second": 1.571,
129
- "step": 4000
130
  }
131
  ],
132
  "logging_steps": 500,
133
- "max_steps": 4492,
134
  "num_input_tokens_seen": 0,
135
- "num_train_epochs": 4,
136
  "save_steps": 500,
137
- "total_flos": 1.7327084846186496e+16,
138
  "train_batch_size": 4,
139
  "trial_name": null,
140
  "trial_params": null
 
1
  {
2
+ "best_metric": 3.263693332672119,
3
+ "best_model_checkpoint": "/content/drive/MyDrive/W210 Capstone - Lyric Generation with Melody/loaf/models/loaf/bart/bart-finetuned-loaf-512-lyrictoplan/checkpoint-500",
4
+ "epoch": 0.4452359750667854,
5
  "eval_steps": 500,
6
+ "global_step": 500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.45,
13
+ "grad_norm": 29.74579620361328,
14
+ "learning_rate": 4.271297120807361e-05,
15
+ "loss": 2.451,
16
  "step": 500
17
  },
18
  {
19
  "epoch": 0.45,
20
+ "eval_loss": 3.263693332672119,
21
+ "eval_runtime": 39.3865,
22
+ "eval_samples_per_second": 15.792,
23
+ "eval_steps_per_second": 3.961,
24
  "step": 500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
25
  }
26
  ],
27
  "logging_steps": 500,
28
+ "max_steps": 3369,
29
  "num_input_tokens_seen": 0,
30
+ "num_train_epochs": 3,
31
  "save_steps": 500,
32
+ "total_flos": 2167104602112000.0,
33
  "train_batch_size": 4,
34
  "trial_name": null,
35
  "trial_params": null
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f998e93eefaec729748687a686b83a7b841a02229dd7c4807b29862479a3dc23
3
  size 5432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14ba679ed25fed17e1d592e85271f6f57742fd304dbe529e46555f0566aaa169
3
  size 5432