rootxhacker commited on
Commit
d3953b6
·
verified ·
1 Parent(s): f13746c

Training in progress, step 18500, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f56b521d1a393817f17047a20ecf2bb1e83ee9f43692fbe30db28020b9d79b36
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a40b3ec9fac131b936a25a3717ad20559ac04c1beac8d89dd6e2dfe4e69599f5
3
  size 36730224
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:af75d0581fbdc90030a2d5f4ec75948b6c9a777f25105f2635a11a66e19a8991
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0c5233fd48a8ddd485f6a2ec03ff9389775661db4a9e073cc9e6578c35f6acc
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aec74914111c893531d77390ed26a8e62f47c5fd368563e3e2ea395d9971bdd0
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a637703e63dd26bf3fc5d45021980dfb8c6c15b36b34b2b812e241dcfa5b3276
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fb36538bca0e8f82a230d205fcda169c895fd4416ff12f18badea8938ea79e34
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac82d08ac6c84227c6b48fa89b5e8a7f058764ced7fff8850851393f056e0943
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:04fb24a5e33627ec28e9d8236874cf729cbb5859a1714c36c632683bea019abf
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1ac76f538d93ba96883051bed95b0b8b1845abfe936b318341f400f4768f02a
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": 16750,
3
  "best_metric": 1.4708431959152222,
4
  "best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-15500",
5
- "epoch": 1.3845088839320052,
6
  "eval_steps": 250,
7
- "global_step": 18000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -3104,6 +3104,92 @@
3104
  "eval_samples_per_second": 55.559,
3105
  "eval_steps_per_second": 13.89,
3106
  "step": 18000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3107
  }
3108
  ],
3109
  "logging_steps": 50,
 
2
  "best_global_step": 16750,
3
  "best_metric": 1.4708431959152222,
4
  "best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-15500",
5
+ "epoch": 1.4229674640412275,
6
  "eval_steps": 250,
7
+ "global_step": 18500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
3104
  "eval_samples_per_second": 55.559,
3105
  "eval_steps_per_second": 13.89,
3106
  "step": 18000
3107
+ },
3108
+ {
3109
+ "epoch": 1.3883547419429274,
3110
+ "grad_norm": 1.6790105104446411,
3111
+ "learning_rate": 4.7670306238741216e-05,
3112
+ "loss": 1.4791,
3113
+ "step": 18050
3114
+ },
3115
+ {
3116
+ "epoch": 1.3922005999538496,
3117
+ "grad_norm": 1.1840436458587646,
3118
+ "learning_rate": 4.74129627587221e-05,
3119
+ "loss": 1.4021,
3120
+ "step": 18100
3121
+ },
3122
+ {
3123
+ "epoch": 1.396046457964772,
3124
+ "grad_norm": 1.7883968353271484,
3125
+ "learning_rate": 4.715561927870298e-05,
3126
+ "loss": 1.4637,
3127
+ "step": 18150
3128
+ },
3129
+ {
3130
+ "epoch": 1.3998923159756942,
3131
+ "grad_norm": 1.2177505493164062,
3132
+ "learning_rate": 4.689827579868387e-05,
3133
+ "loss": 1.5123,
3134
+ "step": 18200
3135
+ },
3136
+ {
3137
+ "epoch": 1.4037381739866164,
3138
+ "grad_norm": 1.439232349395752,
3139
+ "learning_rate": 4.6640932318664756e-05,
3140
+ "loss": 1.4579,
3141
+ "step": 18250
3142
+ },
3143
+ {
3144
+ "epoch": 1.4037381739866164,
3145
+ "eval_loss": 1.4953014850616455,
3146
+ "eval_runtime": 17.9127,
3147
+ "eval_samples_per_second": 55.826,
3148
+ "eval_steps_per_second": 13.957,
3149
+ "step": 18250
3150
+ },
3151
+ {
3152
+ "epoch": 1.4075840319975388,
3153
+ "grad_norm": 2.0796408653259277,
3154
+ "learning_rate": 4.638358883864563e-05,
3155
+ "loss": 1.4295,
3156
+ "step": 18300
3157
+ },
3158
+ {
3159
+ "epoch": 1.4114298900084608,
3160
+ "grad_norm": 1.3032926321029663,
3161
+ "learning_rate": 4.6126245358626515e-05,
3162
+ "loss": 1.4733,
3163
+ "step": 18350
3164
+ },
3165
+ {
3166
+ "epoch": 1.4152757480193832,
3167
+ "grad_norm": 0.9058660864830017,
3168
+ "learning_rate": 4.58689018786074e-05,
3169
+ "loss": 1.4446,
3170
+ "step": 18400
3171
+ },
3172
+ {
3173
+ "epoch": 1.4191216060303053,
3174
+ "grad_norm": 2.05460786819458,
3175
+ "learning_rate": 4.561155839858828e-05,
3176
+ "loss": 1.4133,
3177
+ "step": 18450
3178
+ },
3179
+ {
3180
+ "epoch": 1.4229674640412275,
3181
+ "grad_norm": 0.8309249877929688,
3182
+ "learning_rate": 4.535421491856917e-05,
3183
+ "loss": 1.456,
3184
+ "step": 18500
3185
+ },
3186
+ {
3187
+ "epoch": 1.4229674640412275,
3188
+ "eval_loss": 1.480312466621399,
3189
+ "eval_runtime": 18.2137,
3190
+ "eval_samples_per_second": 54.904,
3191
+ "eval_steps_per_second": 13.726,
3192
+ "step": 18500
3193
  }
3194
  ],
3195
  "logging_steps": 50,