xyfJASON commited on
Commit
3c427a4
1 Parent(s): e332f9b

Upload ddpm_pred_x0_cifar10 checkpoints and training logs

Browse files
Files changed (47) hide show
  1. ddpm_pred_x0_cifar10/ckpt/step079999/ema.pt +3 -0
  2. ddpm_pred_x0_cifar10/ckpt/step079999/meta.pt +3 -0
  3. ddpm_pred_x0_cifar10/ckpt/step079999/model.pt +3 -0
  4. ddpm_pred_x0_cifar10/ckpt/step079999/optimizer.pt +3 -0
  5. ddpm_pred_x0_cifar10/ckpt/step159999/ema.pt +3 -0
  6. ddpm_pred_x0_cifar10/ckpt/step159999/meta.pt +3 -0
  7. ddpm_pred_x0_cifar10/ckpt/step159999/model.pt +3 -0
  8. ddpm_pred_x0_cifar10/ckpt/step159999/optimizer.pt +3 -0
  9. ddpm_pred_x0_cifar10/ckpt/step239999/ema.pt +3 -0
  10. ddpm_pred_x0_cifar10/ckpt/step239999/meta.pt +3 -0
  11. ddpm_pred_x0_cifar10/ckpt/step239999/model.pt +3 -0
  12. ddpm_pred_x0_cifar10/ckpt/step239999/optimizer.pt +3 -0
  13. ddpm_pred_x0_cifar10/ckpt/step319999/ema.pt +3 -0
  14. ddpm_pred_x0_cifar10/ckpt/step319999/meta.pt +3 -0
  15. ddpm_pred_x0_cifar10/ckpt/step319999/model.pt +3 -0
  16. ddpm_pred_x0_cifar10/ckpt/step319999/optimizer.pt +3 -0
  17. ddpm_pred_x0_cifar10/ckpt/step399999/ema.pt +3 -0
  18. ddpm_pred_x0_cifar10/ckpt/step399999/meta.pt +3 -0
  19. ddpm_pred_x0_cifar10/ckpt/step399999/model.pt +3 -0
  20. ddpm_pred_x0_cifar10/ckpt/step399999/optimizer.pt +3 -0
  21. ddpm_pred_x0_cifar10/ckpt/step479999/ema.pt +3 -0
  22. ddpm_pred_x0_cifar10/ckpt/step479999/meta.pt +3 -0
  23. ddpm_pred_x0_cifar10/ckpt/step479999/model.pt +3 -0
  24. ddpm_pred_x0_cifar10/ckpt/step479999/optimizer.pt +3 -0
  25. ddpm_pred_x0_cifar10/ckpt/step559999/ema.pt +3 -0
  26. ddpm_pred_x0_cifar10/ckpt/step559999/meta.pt +3 -0
  27. ddpm_pred_x0_cifar10/ckpt/step559999/model.pt +3 -0
  28. ddpm_pred_x0_cifar10/ckpt/step559999/optimizer.pt +3 -0
  29. ddpm_pred_x0_cifar10/ckpt/step639999/ema.pt +3 -0
  30. ddpm_pred_x0_cifar10/ckpt/step639999/meta.pt +3 -0
  31. ddpm_pred_x0_cifar10/ckpt/step639999/model.pt +3 -0
  32. ddpm_pred_x0_cifar10/ckpt/step639999/optimizer.pt +3 -0
  33. ddpm_pred_x0_cifar10/ckpt/step719999/ema.pt +3 -0
  34. ddpm_pred_x0_cifar10/ckpt/step719999/meta.pt +3 -0
  35. ddpm_pred_x0_cifar10/ckpt/step719999/model.pt +3 -0
  36. ddpm_pred_x0_cifar10/ckpt/step719999/optimizer.pt +3 -0
  37. ddpm_pred_x0_cifar10/ckpt/step799999/ema.pt +3 -0
  38. ddpm_pred_x0_cifar10/ckpt/step799999/meta.pt +3 -0
  39. ddpm_pred_x0_cifar10/ckpt/step799999/model.pt +3 -0
  40. ddpm_pred_x0_cifar10/ckpt/step799999/optimizer.pt +3 -0
  41. ddpm_pred_x0_cifar10/config-2023-08-07-07-49-01.yaml +54 -0
  42. ddpm_pred_x0_cifar10/config-2023-08-09-23-34-33.yaml +54 -0
  43. ddpm_pred_x0_cifar10/output-2023-08-07-07-49-01.log +847 -0
  44. ddpm_pred_x0_cifar10/output-2023-08-09-23-34-33.log +0 -0
  45. ddpm_pred_x0_cifar10/samples.zip +3 -0
  46. ddpm_pred_x0_cifar10/tensorboard/events.out.tfevents.1691408944.aa-ESC4000-G4.24711.0 +3 -0
  47. ddpm_pred_x0_cifar10/tensorboard/events.out.tfevents.1691638477.aa-ESC4000-G4.12367.0 +3 -0
ddpm_pred_x0_cifar10/ckpt/step079999/ema.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e666f49c500ce6fe4861467bee440518c18d87877644156d1929b3cd910ca071
3
+ size 143067273
ddpm_pred_x0_cifar10/ckpt/step079999/meta.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:460ecfe9af7c3c2749db8e2d57285af468303df85ec0957a7e666d8bf7bc5387
3
+ size 425
ddpm_pred_x0_cifar10/ckpt/step079999/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e097b35bfae47bcba0ac9c90c9214779d6e22b00bdc3384066ce87297accb27
3
+ size 143084317
ddpm_pred_x0_cifar10/ckpt/step079999/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1eabd446b8ce194d9c702f4b3ee89024aef5c41b3643a91d50aad6db7521b304
3
+ size 286245381
ddpm_pred_x0_cifar10/ckpt/step159999/ema.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bfdef0e713407ba8ee5658c3ade1035d7591287357c1c63d63996d1838dde02a
3
+ size 143067273
ddpm_pred_x0_cifar10/ckpt/step159999/meta.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c084ac1d346ecee3fb1e4d361399e0c2850bdba7e412be1f08db0ae69cd059c
3
+ size 425
ddpm_pred_x0_cifar10/ckpt/step159999/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d925d5eef0fde7149524c922f6c5b1e525a73512672f96d552a5729dab1c3f83
3
+ size 143084317
ddpm_pred_x0_cifar10/ckpt/step159999/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8462a49bfd582335fd535ff259208a4acb62a3e579cfc5c5a342f2ab57de082c
3
+ size 286245381
ddpm_pred_x0_cifar10/ckpt/step239999/ema.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d31c3e953df4e8b1a6b959754e3c55b4d0d80f4ac4131cedac599ee05732fd2
3
+ size 143067273
ddpm_pred_x0_cifar10/ckpt/step239999/meta.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4df47f45c956c46e67d34dd50aa6b4413b30d5ca1c3db1a0c6534d2659d8afbf
3
+ size 425
ddpm_pred_x0_cifar10/ckpt/step239999/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c11cb487cdb3d63f70da5a94cba34858fb717c644ec28f190d5403ab9e74ad36
3
+ size 143084317
ddpm_pred_x0_cifar10/ckpt/step239999/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f336dff86b18025f20ef321646b06a5b3b1f74f8f2630f92702ca164b7705ce
3
+ size 286245381
ddpm_pred_x0_cifar10/ckpt/step319999/ema.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3786b19277c131b381805135684027edc2c2958bd2de43a40aa114d5f70986e2
3
+ size 143067273
ddpm_pred_x0_cifar10/ckpt/step319999/meta.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f5499fc4a910fdf81fe9bd42925dde65531a71584d5cba241e210753ed32a71
3
+ size 425
ddpm_pred_x0_cifar10/ckpt/step319999/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3fa82ef1e1960e0b6ed5942d9a31bd9b312ede685b1b1838b0e487be64990f98
3
+ size 143084317
ddpm_pred_x0_cifar10/ckpt/step319999/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2f0bac98db2393c14d39e118ffba87cad1a4bb3b1f879404e95266865ebeb3a
3
+ size 286245381
ddpm_pred_x0_cifar10/ckpt/step399999/ema.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3c4b7c0d40c02cc673509ec66ea15b5c03530c5f793a78d202fefd88eb10dab
3
+ size 143067273
ddpm_pred_x0_cifar10/ckpt/step399999/meta.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9f6dc55e9ad20a9d9fb49c912a1c2df80668233313c10de9d33705ec250ea9c
3
+ size 425
ddpm_pred_x0_cifar10/ckpt/step399999/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:727309417f4cd17eeb975340b4da653adb831697a80b93b5a1b64623c8857f04
3
+ size 143084317
ddpm_pred_x0_cifar10/ckpt/step399999/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dab9f540504edc3d3c6e508368fe8bdae815c489c4d19e7c88db90a9e9800354
3
+ size 286249989
ddpm_pred_x0_cifar10/ckpt/step479999/ema.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e28c7f24ea88ff0b7694ae73812fc3b79519c69396d3266dcce19bf1c9f9c3c
3
+ size 143067273
ddpm_pred_x0_cifar10/ckpt/step479999/meta.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:985bdda8a39e9fb74da57b328f1891a570fad0d423cdd81c18e3de102b8ebd4f
3
+ size 425
ddpm_pred_x0_cifar10/ckpt/step479999/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c8f86ed4de0ea7286f39b8a443f1f4671d9a3c5305048a9209c0e76debee0eb
3
+ size 143084317
ddpm_pred_x0_cifar10/ckpt/step479999/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d05595006895fe164173ba4d5b1025408495854bdeac076135623bd24b5ee28
3
+ size 286249989
ddpm_pred_x0_cifar10/ckpt/step559999/ema.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d60723e582b6b5fbcbb3eb8b8d40a637cf4fcf83b907a9e23e406405c440cf11
3
+ size 143067273
ddpm_pred_x0_cifar10/ckpt/step559999/meta.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cdc2d49c8d25a3b6ff963b03a7b56b5f32d8c3f976155b6a480b476766196f8f
3
+ size 425
ddpm_pred_x0_cifar10/ckpt/step559999/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6635c847fd00c3329feca454b18ef3b4890f6cc3224134f67820bf0a7219544
3
+ size 143084317
ddpm_pred_x0_cifar10/ckpt/step559999/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e600dd82f0b7a03a657c0ca0c27dcaa8f3011c6acc10dc99032b10f7ef22a89
3
+ size 286249989
ddpm_pred_x0_cifar10/ckpt/step639999/ema.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a877b7043569d4e306033c88449bbd9a55c459be0a88e7a00611b659e43cad6
3
+ size 143067273
ddpm_pred_x0_cifar10/ckpt/step639999/meta.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f9174d79f6b5349c61129e6a460309b4fbb79ef89b46c7ebf3a3ef369e264a8e
3
+ size 425
ddpm_pred_x0_cifar10/ckpt/step639999/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8193ea6179292427b4b053e21b852e718a48ef729ef5e1888c97be3e72b4cb9
3
+ size 143084317
ddpm_pred_x0_cifar10/ckpt/step639999/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41611e3721463acc18ab85f35c4a1927e9246d3a9303c7f4402040e4cef9af18
3
+ size 286249989
ddpm_pred_x0_cifar10/ckpt/step719999/ema.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0381c4709e6109e4a5844a80c2a201890900c53c36fa7ac57cfe317f0e4679be
3
+ size 143067273
ddpm_pred_x0_cifar10/ckpt/step719999/meta.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:baefc936b268cd126cebe024329539b0016de73a64b053eed6cdd7374d85e0ea
3
+ size 425
ddpm_pred_x0_cifar10/ckpt/step719999/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e65dc301140cb5d09e95874a1a8c7ecd02c798c3058880bcf6d7a11c76010495
3
+ size 143084317
ddpm_pred_x0_cifar10/ckpt/step719999/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ea18303173867a6b43b6dcd15dd81246180c4bf00465967c9b3f45386e8c13e
3
+ size 286249989
ddpm_pred_x0_cifar10/ckpt/step799999/ema.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71474fe843c08630f00d1eb6e9bd54b0ed716514751669d3038ba2bbaabc2976
3
+ size 143067273
ddpm_pred_x0_cifar10/ckpt/step799999/meta.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7fc389681c8eb2922fbeabf0c843cedbad10aaaa30455d0d2b53d1e86ad206b
3
+ size 425
ddpm_pred_x0_cifar10/ckpt/step799999/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a9c1d6adb3a6589581d9fe08c285dbb8ac79e0d267721e7731e092128120a37
3
+ size 143084317
ddpm_pred_x0_cifar10/ckpt/step799999/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:79fb294e85b13efbc16852bb3196b7388455cc396661c3a19bbc446e6bed0733
3
+ size 286249989
ddpm_pred_x0_cifar10/config-2023-08-07-07-49-01.yaml ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ seed: 2022
2
+ data:
3
+ name: CIFAR-10
4
+ dataroot: ~/data/CIFAR-10/
5
+ img_size: 32
6
+ img_channels: 3
7
+ num_classes: 10
8
+ dataloader:
9
+ num_workers: 4
10
+ pin_memory: true
11
+ prefetch_factor: 2
12
+ micro_batch: 0
13
+ model:
14
+ target: models.unet.UNet
15
+ params:
16
+ in_channels: 3
17
+ out_channels: 3
18
+ dim: 128
19
+ dim_mults:
20
+ - 1
21
+ - 2
22
+ - 2
23
+ - 2
24
+ use_attn:
25
+ - false
26
+ - true
27
+ - false
28
+ - false
29
+ num_res_blocks: 2
30
+ n_heads: 1
31
+ dropout: 0.1
32
+ ema_decay: 0.9999
33
+ ema_gradual: true
34
+ diffusion:
35
+ target: diffusions.ddpm.DDPM
36
+ params:
37
+ total_steps: 1000
38
+ beta_schedule: linear
39
+ beta_start: 0.0001
40
+ beta_end: 0.02
41
+ objective: pred_x0
42
+ var_type: fixed_large
43
+ train:
44
+ n_steps: 800000
45
+ batch_size: 128
46
+ clip_grad_norm: 1.0
47
+ print_freq: 400
48
+ save_freq: 10000
49
+ sample_freq: 5000
50
+ n_samples: 64
51
+ optim:
52
+ target: torch.optim.Adam
53
+ params:
54
+ lr: 0.0002
ddpm_pred_x0_cifar10/config-2023-08-09-23-34-33.yaml ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ seed: 2022
2
+ data:
3
+ name: CIFAR-10
4
+ dataroot: ~/data/CIFAR-10/
5
+ img_size: 32
6
+ img_channels: 3
7
+ num_classes: 10
8
+ dataloader:
9
+ num_workers: 4
10
+ pin_memory: true
11
+ prefetch_factor: 2
12
+ micro_batch: 0
13
+ model:
14
+ target: models.unet.UNet
15
+ params:
16
+ in_channels: 3
17
+ out_channels: 3
18
+ dim: 128
19
+ dim_mults:
20
+ - 1
21
+ - 2
22
+ - 2
23
+ - 2
24
+ use_attn:
25
+ - false
26
+ - true
27
+ - false
28
+ - false
29
+ num_res_blocks: 2
30
+ n_heads: 1
31
+ dropout: 0.1
32
+ ema_decay: 0.9999
33
+ ema_gradual: true
34
+ diffusion:
35
+ target: diffusions.ddpm.DDPM
36
+ params:
37
+ total_steps: 1000
38
+ beta_schedule: linear
39
+ beta_start: 0.0001
40
+ beta_end: 0.02
41
+ objective: pred_x0
42
+ var_type: fixed_large
43
+ train:
44
+ n_steps: 800000
45
+ batch_size: 128
46
+ clip_grad_norm: 1.0
47
+ print_freq: 400
48
+ save_freq: 10000
49
+ sample_freq: 5000
50
+ n_samples: 64
51
+ optim:
52
+ target: torch.optim.Adam
53
+ params:
54
+ lr: 0.0002
ddpm_pred_x0_cifar10/output-2023-08-07-07-49-01.log ADDED
@@ -0,0 +1,847 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2023-08-07 07:49:04,646 - INFO - Experiment directory: ./runs/ddpm_cifar10_pred_x0
2
+ 2023-08-07 07:49:04,647 - INFO - Number of processes: 4
3
+ 2023-08-07 07:49:04,647 - INFO - Distributed type: DistributedType.MULTI_GPU
4
+ 2023-08-07 07:49:04,647 - INFO - Mixed precision: no
5
+ 2023-08-07 07:49:04,647 - INFO - ==============================
6
+ 2023-08-07 07:49:06,157 - INFO - Size of training set: 50000
7
+ 2023-08-07 07:49:06,157 - INFO - Batch size per process: 32
8
+ 2023-08-07 07:49:06,157 - INFO - Total batch size: 128
9
+ 2023-08-07 07:49:06,157 - INFO - ==============================
10
+ 2023-08-07 07:49:06,976 - INFO - Start training...
11
+ 2023-08-07 07:50:31,019 - INFO - [Train] step: 399, loss: 0.119830, lr: 0.000200
12
+ 2023-08-07 07:51:53,317 - INFO - [Train] step: 799, loss: 0.150414, lr: 0.000200
13
+ 2023-08-07 07:53:15,251 - INFO - [Train] step: 1199, loss: 0.125284, lr: 0.000200
14
+ 2023-08-07 07:54:37,181 - INFO - [Train] step: 1599, loss: 0.163405, lr: 0.000200
15
+ 2023-08-07 07:55:59,305 - INFO - [Train] step: 1999, loss: 0.124573, lr: 0.000200
16
+ 2023-08-07 07:57:21,488 - INFO - [Train] step: 2399, loss: 0.121558, lr: 0.000200
17
+ 2023-08-07 07:58:43,758 - INFO - [Train] step: 2799, loss: 0.140495, lr: 0.000200
18
+ 2023-08-07 08:00:06,082 - INFO - [Train] step: 3199, loss: 0.120343, lr: 0.000200
19
+ 2023-08-07 08:01:28,422 - INFO - [Train] step: 3599, loss: 0.126704, lr: 0.000200
20
+ 2023-08-07 08:02:50,514 - INFO - [Train] step: 3999, loss: 0.129013, lr: 0.000200
21
+ 2023-08-07 08:04:12,450 - INFO - [Train] step: 4399, loss: 0.131795, lr: 0.000200
22
+ 2023-08-07 08:05:34,597 - INFO - [Train] step: 4799, loss: 0.125565, lr: 0.000200
23
+ 2023-08-07 08:07:17,806 - INFO - [Train] step: 5199, loss: 0.090839, lr: 0.000200
24
+ 2023-08-07 08:08:40,145 - INFO - [Train] step: 5599, loss: 0.109404, lr: 0.000200
25
+ 2023-08-07 08:10:02,054 - INFO - [Train] step: 5999, loss: 0.123738, lr: 0.000200
26
+ 2023-08-07 08:11:23,822 - INFO - [Train] step: 6399, loss: 0.119509, lr: 0.000200
27
+ 2023-08-07 08:12:45,922 - INFO - [Train] step: 6799, loss: 0.099912, lr: 0.000200
28
+ 2023-08-07 08:14:07,996 - INFO - [Train] step: 7199, loss: 0.087818, lr: 0.000200
29
+ 2023-08-07 08:15:30,343 - INFO - [Train] step: 7599, loss: 0.119198, lr: 0.000200
30
+ 2023-08-07 08:16:52,457 - INFO - [Train] step: 7999, loss: 0.115591, lr: 0.000200
31
+ 2023-08-07 08:18:14,790 - INFO - [Train] step: 8399, loss: 0.099107, lr: 0.000200
32
+ 2023-08-07 08:19:36,938 - INFO - [Train] step: 8799, loss: 0.148510, lr: 0.000200
33
+ 2023-08-07 08:20:59,065 - INFO - [Train] step: 9199, loss: 0.136262, lr: 0.000200
34
+ 2023-08-07 08:22:21,136 - INFO - [Train] step: 9599, loss: 0.087655, lr: 0.000200
35
+ 2023-08-07 08:23:42,670 - INFO - [Train] step: 9999, loss: 0.128666, lr: 0.000200
36
+ 2023-08-07 08:25:28,794 - INFO - [Train] step: 10399, loss: 0.117899, lr: 0.000200
37
+ 2023-08-07 08:26:50,704 - INFO - [Train] step: 10799, loss: 0.098106, lr: 0.000200
38
+ 2023-08-07 08:28:12,833 - INFO - [Train] step: 11199, loss: 0.104961, lr: 0.000200
39
+ 2023-08-07 08:29:35,175 - INFO - [Train] step: 11599, loss: 0.106024, lr: 0.000200
40
+ 2023-08-07 08:30:57,087 - INFO - [Train] step: 11999, loss: 0.095780, lr: 0.000200
41
+ 2023-08-07 08:32:19,193 - INFO - [Train] step: 12399, loss: 0.110405, lr: 0.000200
42
+ 2023-08-07 08:33:41,542 - INFO - [Train] step: 12799, loss: 0.116221, lr: 0.000200
43
+ 2023-08-07 08:35:03,633 - INFO - [Train] step: 13199, loss: 0.097295, lr: 0.000200
44
+ 2023-08-07 08:36:25,794 - INFO - [Train] step: 13599, loss: 0.146900, lr: 0.000200
45
+ 2023-08-07 08:37:48,075 - INFO - [Train] step: 13999, loss: 0.122996, lr: 0.000200
46
+ 2023-08-07 08:39:10,261 - INFO - [Train] step: 14399, loss: 0.133118, lr: 0.000200
47
+ 2023-08-07 08:40:32,561 - INFO - [Train] step: 14799, loss: 0.077836, lr: 0.000200
48
+ 2023-08-07 08:42:17,045 - INFO - [Train] step: 15199, loss: 0.129952, lr: 0.000200
49
+ 2023-08-07 08:43:39,221 - INFO - [Train] step: 15599, loss: 0.110417, lr: 0.000200
50
+ 2023-08-07 08:45:01,943 - INFO - [Train] step: 15999, loss: 0.099892, lr: 0.000200
51
+ 2023-08-07 08:46:24,017 - INFO - [Train] step: 16399, loss: 0.156371, lr: 0.000200
52
+ 2023-08-07 08:47:46,151 - INFO - [Train] step: 16799, loss: 0.099175, lr: 0.000200
53
+ 2023-08-07 08:49:07,453 - INFO - [Train] step: 17199, loss: 0.133834, lr: 0.000200
54
+ 2023-08-07 08:50:28,968 - INFO - [Train] step: 17599, loss: 0.135953, lr: 0.000200
55
+ 2023-08-07 08:51:51,301 - INFO - [Train] step: 17999, loss: 0.120979, lr: 0.000200
56
+ 2023-08-07 08:53:13,426 - INFO - [Train] step: 18399, loss: 0.112717, lr: 0.000200
57
+ 2023-08-07 08:54:35,491 - INFO - [Train] step: 18799, loss: 0.123487, lr: 0.000200
58
+ 2023-08-07 08:55:57,467 - INFO - [Train] step: 19199, loss: 0.147328, lr: 0.000200
59
+ 2023-08-07 08:57:19,790 - INFO - [Train] step: 19599, loss: 0.096656, lr: 0.000200
60
+ 2023-08-07 08:58:42,138 - INFO - [Train] step: 19999, loss: 0.102965, lr: 0.000200
61
+ 2023-08-07 09:00:26,563 - INFO - [Train] step: 20399, loss: 0.155700, lr: 0.000200
62
+ 2023-08-07 09:01:48,902 - INFO - [Train] step: 20799, loss: 0.129004, lr: 0.000200
63
+ 2023-08-07 09:03:11,239 - INFO - [Train] step: 21199, loss: 0.109404, lr: 0.000200
64
+ 2023-08-07 09:04:33,571 - INFO - [Train] step: 21599, loss: 0.107027, lr: 0.000200
65
+ 2023-08-07 09:05:55,701 - INFO - [Train] step: 21999, loss: 0.107544, lr: 0.000200
66
+ 2023-08-07 09:07:18,032 - INFO - [Train] step: 22399, loss: 0.105604, lr: 0.000200
67
+ 2023-08-07 09:08:39,934 - INFO - [Train] step: 22799, loss: 0.122270, lr: 0.000200
68
+ 2023-08-07 09:10:02,291 - INFO - [Train] step: 23199, loss: 0.128101, lr: 0.000200
69
+ 2023-08-07 09:11:24,398 - INFO - [Train] step: 23599, loss: 0.094124, lr: 0.000200
70
+ 2023-08-07 09:12:46,315 - INFO - [Train] step: 23999, loss: 0.125352, lr: 0.000200
71
+ 2023-08-07 09:14:07,859 - INFO - [Train] step: 24399, loss: 0.095635, lr: 0.000200
72
+ 2023-08-07 09:15:29,952 - INFO - [Train] step: 24799, loss: 0.083003, lr: 0.000200
73
+ 2023-08-07 09:17:12,575 - INFO - [Train] step: 25199, loss: 0.090902, lr: 0.000200
74
+ 2023-08-07 09:18:34,666 - INFO - [Train] step: 25599, loss: 0.093669, lr: 0.000200
75
+ 2023-08-07 09:19:57,022 - INFO - [Train] step: 25999, loss: 0.115352, lr: 0.000200
76
+ 2023-08-07 09:21:18,751 - INFO - [Train] step: 26399, loss: 0.113375, lr: 0.000200
77
+ 2023-08-07 09:22:40,442 - INFO - [Train] step: 26799, loss: 0.122325, lr: 0.000200
78
+ 2023-08-07 09:24:02,440 - INFO - [Train] step: 27199, loss: 0.116420, lr: 0.000200
79
+ 2023-08-07 09:25:24,509 - INFO - [Train] step: 27599, loss: 0.139663, lr: 0.000200
80
+ 2023-08-07 09:26:46,422 - INFO - [Train] step: 27999, loss: 0.103903, lr: 0.000200
81
+ 2023-08-07 09:28:08,147 - INFO - [Train] step: 28399, loss: 0.116586, lr: 0.000200
82
+ 2023-08-07 09:29:30,259 - INFO - [Train] step: 28799, loss: 0.152031, lr: 0.000200
83
+ 2023-08-07 09:30:52,145 - INFO - [Train] step: 29199, loss: 0.099812, lr: 0.000200
84
+ 2023-08-07 09:32:14,272 - INFO - [Train] step: 29599, loss: 0.112326, lr: 0.000200
85
+ 2023-08-07 09:33:36,637 - INFO - [Train] step: 29999, loss: 0.126363, lr: 0.000200
86
+ 2023-08-07 09:35:21,244 - INFO - [Train] step: 30399, loss: 0.105871, lr: 0.000200
87
+ 2023-08-07 09:36:43,452 - INFO - [Train] step: 30799, loss: 0.130065, lr: 0.000200
88
+ 2023-08-07 09:38:05,688 - INFO - [Train] step: 31199, loss: 0.105673, lr: 0.000200
89
+ 2023-08-07 09:39:28,109 - INFO - [Train] step: 31599, loss: 0.102385, lr: 0.000200
90
+ 2023-08-07 09:40:50,177 - INFO - [Train] step: 31999, loss: 0.130559, lr: 0.000200
91
+ 2023-08-07 09:42:12,305 - INFO - [Train] step: 32399, loss: 0.096475, lr: 0.000200
92
+ 2023-08-07 09:43:34,656 - INFO - [Train] step: 32799, loss: 0.101694, lr: 0.000200
93
+ 2023-08-07 09:44:56,970 - INFO - [Train] step: 33199, loss: 0.115387, lr: 0.000200
94
+ 2023-08-07 09:46:19,111 - INFO - [Train] step: 33599, loss: 0.127949, lr: 0.000200
95
+ 2023-08-07 09:47:41,435 - INFO - [Train] step: 33999, loss: 0.112779, lr: 0.000200
96
+ 2023-08-07 09:49:03,543 - INFO - [Train] step: 34399, loss: 0.134133, lr: 0.000200
97
+ 2023-08-07 09:50:25,707 - INFO - [Train] step: 34799, loss: 0.111700, lr: 0.000200
98
+ 2023-08-07 09:52:09,128 - INFO - [Train] step: 35199, loss: 0.135414, lr: 0.000200
99
+ 2023-08-07 09:53:31,407 - INFO - [Train] step: 35599, loss: 0.119928, lr: 0.000200
100
+ 2023-08-07 09:54:53,372 - INFO - [Train] step: 35999, loss: 0.115652, lr: 0.000200
101
+ 2023-08-07 09:56:15,305 - INFO - [Train] step: 36399, loss: 0.100819, lr: 0.000200
102
+ 2023-08-07 09:57:37,592 - INFO - [Train] step: 36799, loss: 0.120232, lr: 0.000200
103
+ 2023-08-07 09:58:59,540 - INFO - [Train] step: 37199, loss: 0.142750, lr: 0.000200
104
+ 2023-08-07 10:00:21,641 - INFO - [Train] step: 37599, loss: 0.070434, lr: 0.000200
105
+ 2023-08-07 10:01:43,790 - INFO - [Train] step: 37999, loss: 0.112740, lr: 0.000200
106
+ 2023-08-07 10:03:05,922 - INFO - [Train] step: 38399, loss: 0.112250, lr: 0.000200
107
+ 2023-08-07 10:04:28,247 - INFO - [Train] step: 38799, loss: 0.079933, lr: 0.000200
108
+ 2023-08-07 10:05:50,122 - INFO - [Train] step: 39199, loss: 0.113054, lr: 0.000200
109
+ 2023-08-07 10:07:12,298 - INFO - [Train] step: 39599, loss: 0.109953, lr: 0.000200
110
+ 2023-08-07 10:08:34,516 - INFO - [Train] step: 39999, loss: 0.130182, lr: 0.000200
111
+ 2023-08-07 10:10:17,846 - INFO - [Train] step: 40399, loss: 0.149816, lr: 0.000200
112
+ 2023-08-07 10:11:39,571 - INFO - [Train] step: 40799, loss: 0.101177, lr: 0.000200
113
+ 2023-08-07 10:13:01,846 - INFO - [Train] step: 41199, loss: 0.100172, lr: 0.000200
114
+ 2023-08-07 10:14:23,798 - INFO - [Train] step: 41599, loss: 0.104291, lr: 0.000200
115
+ 2023-08-07 10:15:46,138 - INFO - [Train] step: 41999, loss: 0.104690, lr: 0.000200
116
+ 2023-08-07 10:17:08,442 - INFO - [Train] step: 42399, loss: 0.112855, lr: 0.000200
117
+ 2023-08-07 10:18:30,572 - INFO - [Train] step: 42799, loss: 0.095969, lr: 0.000200
118
+ 2023-08-07 10:19:52,666 - INFO - [Train] step: 43199, loss: 0.083825, lr: 0.000200
119
+ 2023-08-07 10:21:14,840 - INFO - [Train] step: 43599, loss: 0.133080, lr: 0.000200
120
+ 2023-08-07 10:22:36,781 - INFO - [Train] step: 43999, loss: 0.069590, lr: 0.000200
121
+ 2023-08-07 10:23:59,100 - INFO - [Train] step: 44399, loss: 0.090524, lr: 0.000200
122
+ 2023-08-07 10:25:21,013 - INFO - [Train] step: 44799, loss: 0.128244, lr: 0.000200
123
+ 2023-08-07 10:27:04,202 - INFO - [Train] step: 45199, loss: 0.101210, lr: 0.000200
124
+ 2023-08-07 10:28:25,700 - INFO - [Train] step: 45599, loss: 0.100796, lr: 0.000200
125
+ 2023-08-07 10:29:47,874 - INFO - [Train] step: 45999, loss: 0.091977, lr: 0.000200
126
+ 2023-08-07 10:31:09,976 - INFO - [Train] step: 46399, loss: 0.088182, lr: 0.000200
127
+ 2023-08-07 10:32:32,410 - INFO - [Train] step: 46799, loss: 0.121288, lr: 0.000200
128
+ 2023-08-07 10:33:54,859 - INFO - [Train] step: 47199, loss: 0.167538, lr: 0.000200
129
+ 2023-08-07 10:35:16,362 - INFO - [Train] step: 47599, loss: 0.156691, lr: 0.000200
130
+ 2023-08-07 10:36:38,279 - INFO - [Train] step: 47999, loss: 0.122817, lr: 0.000200
131
+ 2023-08-07 10:38:00,182 - INFO - [Train] step: 48399, loss: 0.098979, lr: 0.000200
132
+ 2023-08-07 10:39:22,149 - INFO - [Train] step: 48799, loss: 0.118480, lr: 0.000200
133
+ 2023-08-07 10:40:44,057 - INFO - [Train] step: 49199, loss: 0.093471, lr: 0.000200
134
+ 2023-08-07 10:42:06,400 - INFO - [Train] step: 49599, loss: 0.108082, lr: 0.000200
135
+ 2023-08-07 10:43:28,706 - INFO - [Train] step: 49999, loss: 0.122123, lr: 0.000200
136
+ 2023-08-07 10:45:13,391 - INFO - [Train] step: 50399, loss: 0.127382, lr: 0.000200
137
+ 2023-08-07 10:46:35,084 - INFO - [Train] step: 50799, loss: 0.136278, lr: 0.000200
138
+ 2023-08-07 10:47:57,204 - INFO - [Train] step: 51199, loss: 0.110470, lr: 0.000200
139
+ 2023-08-07 10:49:19,309 - INFO - [Train] step: 51599, loss: 0.142462, lr: 0.000200
140
+ 2023-08-07 10:50:41,676 - INFO - [Train] step: 51999, loss: 0.110496, lr: 0.000200
141
+ 2023-08-07 10:52:03,801 - INFO - [Train] step: 52399, loss: 0.119802, lr: 0.000200
142
+ 2023-08-07 10:53:25,499 - INFO - [Train] step: 52799, loss: 0.120931, lr: 0.000200
143
+ 2023-08-07 10:54:47,625 - INFO - [Train] step: 53199, loss: 0.130146, lr: 0.000200
144
+ 2023-08-07 10:56:09,767 - INFO - [Train] step: 53599, loss: 0.099482, lr: 0.000200
145
+ 2023-08-07 10:57:31,440 - INFO - [Train] step: 53999, loss: 0.117266, lr: 0.000200
146
+ 2023-08-07 10:58:53,373 - INFO - [Train] step: 54399, loss: 0.087286, lr: 0.000200
147
+ 2023-08-07 11:00:15,523 - INFO - [Train] step: 54799, loss: 0.131403, lr: 0.000200
148
+ 2023-08-07 11:01:59,765 - INFO - [Train] step: 55199, loss: 0.104190, lr: 0.000200
149
+ 2023-08-07 11:03:21,449 - INFO - [Train] step: 55599, loss: 0.118375, lr: 0.000200
150
+ 2023-08-07 11:04:43,375 - INFO - [Train] step: 55999, loss: 0.091755, lr: 0.000200
151
+ 2023-08-07 11:06:05,712 - INFO - [Train] step: 56399, loss: 0.109616, lr: 0.000200
152
+ 2023-08-07 11:07:28,055 - INFO - [Train] step: 56799, loss: 0.124397, lr: 0.000200
153
+ 2023-08-07 11:08:50,180 - INFO - [Train] step: 57199, loss: 0.108280, lr: 0.000200
154
+ 2023-08-07 11:10:12,507 - INFO - [Train] step: 57599, loss: 0.092169, lr: 0.000200
155
+ 2023-08-07 11:11:34,645 - INFO - [Train] step: 57999, loss: 0.083924, lr: 0.000200
156
+ 2023-08-07 11:12:56,983 - INFO - [Train] step: 58399, loss: 0.134117, lr: 0.000200
157
+ 2023-08-07 11:14:19,303 - INFO - [Train] step: 58799, loss: 0.106008, lr: 0.000200
158
+ 2023-08-07 11:15:41,433 - INFO - [Train] step: 59199, loss: 0.109932, lr: 0.000200
159
+ 2023-08-07 11:17:03,770 - INFO - [Train] step: 59599, loss: 0.122537, lr: 0.000200
160
+ 2023-08-07 11:18:26,093 - INFO - [Train] step: 59999, loss: 0.091691, lr: 0.000200
161
+ 2023-08-07 11:20:11,139 - INFO - [Train] step: 60399, loss: 0.125825, lr: 0.000200
162
+ 2023-08-07 11:21:33,071 - INFO - [Train] step: 60799, loss: 0.094548, lr: 0.000200
163
+ 2023-08-07 11:22:54,599 - INFO - [Train] step: 61199, loss: 0.104905, lr: 0.000200
164
+ 2023-08-07 11:24:16,730 - INFO - [Train] step: 61599, loss: 0.133830, lr: 0.000200
165
+ 2023-08-07 11:25:38,899 - INFO - [Train] step: 61999, loss: 0.112731, lr: 0.000200
166
+ 2023-08-07 11:27:01,077 - INFO - [Train] step: 62399, loss: 0.100290, lr: 0.000200
167
+ 2023-08-07 11:28:23,330 - INFO - [Train] step: 62799, loss: 0.128646, lr: 0.000200
168
+ 2023-08-07 11:29:45,494 - INFO - [Train] step: 63199, loss: 0.107791, lr: 0.000200
169
+ 2023-08-07 11:31:06,766 - INFO - [Train] step: 63599, loss: 0.107345, lr: 0.000200
170
+ 2023-08-07 11:32:29,095 - INFO - [Train] step: 63999, loss: 0.119336, lr: 0.000200
171
+ 2023-08-07 11:33:51,232 - INFO - [Train] step: 64399, loss: 0.120495, lr: 0.000200
172
+ 2023-08-07 11:35:13,370 - INFO - [Train] step: 64799, loss: 0.115367, lr: 0.000200
173
+ 2023-08-07 11:36:57,599 - INFO - [Train] step: 65199, loss: 0.108854, lr: 0.000200
174
+ 2023-08-07 11:38:19,702 - INFO - [Train] step: 65599, loss: 0.169773, lr: 0.000200
175
+ 2023-08-07 11:39:42,067 - INFO - [Train] step: 65999, loss: 0.147495, lr: 0.000200
176
+ 2023-08-07 11:41:04,404 - INFO - [Train] step: 66399, loss: 0.105130, lr: 0.000200
177
+ 2023-08-07 11:42:26,107 - INFO - [Train] step: 66799, loss: 0.126324, lr: 0.000200
178
+ 2023-08-07 11:43:48,447 - INFO - [Train] step: 67199, loss: 0.127703, lr: 0.000200
179
+ 2023-08-07 11:45:10,773 - INFO - [Train] step: 67599, loss: 0.138544, lr: 0.000200
180
+ 2023-08-07 11:46:32,664 - INFO - [Train] step: 67999, loss: 0.120105, lr: 0.000200
181
+ 2023-08-07 11:47:54,421 - INFO - [Train] step: 68399, loss: 0.126496, lr: 0.000200
182
+ 2023-08-07 11:49:16,629 - INFO - [Train] step: 68799, loss: 0.121904, lr: 0.000200
183
+ 2023-08-07 11:50:38,465 - INFO - [Train] step: 69199, loss: 0.094325, lr: 0.000200
184
+ 2023-08-07 11:52:00,802 - INFO - [Train] step: 69599, loss: 0.095884, lr: 0.000200
185
+ 2023-08-07 11:53:22,706 - INFO - [Train] step: 69999, loss: 0.151581, lr: 0.000200
186
+ 2023-08-07 11:55:09,013 - INFO - [Train] step: 70399, loss: 0.124470, lr: 0.000200
187
+ 2023-08-07 11:56:30,934 - INFO - [Train] step: 70799, loss: 0.121488, lr: 0.000200
188
+ 2023-08-07 11:57:52,827 - INFO - [Train] step: 71199, loss: 0.106889, lr: 0.000200
189
+ 2023-08-07 11:59:14,349 - INFO - [Train] step: 71599, loss: 0.130704, lr: 0.000200
190
+ 2023-08-07 12:00:36,423 - INFO - [Train] step: 71999, loss: 0.087750, lr: 0.000200
191
+ 2023-08-07 12:01:58,383 - INFO - [Train] step: 72399, loss: 0.095367, lr: 0.000200
192
+ 2023-08-07 12:03:20,257 - INFO - [Train] step: 72799, loss: 0.096545, lr: 0.000200
193
+ 2023-08-07 12:04:42,188 - INFO - [Train] step: 73199, loss: 0.099555, lr: 0.000200
194
+ 2023-08-07 12:06:04,337 - INFO - [Train] step: 73599, loss: 0.145404, lr: 0.000200
195
+ 2023-08-07 12:07:26,269 - INFO - [Train] step: 73999, loss: 0.103730, lr: 0.000200
196
+ 2023-08-07 12:08:48,551 - INFO - [Train] step: 74399, loss: 0.143072, lr: 0.000200
197
+ 2023-08-07 12:10:10,691 - INFO - [Train] step: 74799, loss: 0.095879, lr: 0.000200
198
+ 2023-08-07 12:11:54,112 - INFO - [Train] step: 75199, loss: 0.130812, lr: 0.000200
199
+ 2023-08-07 12:13:16,265 - INFO - [Train] step: 75599, loss: 0.079059, lr: 0.000200
200
+ 2023-08-07 12:14:38,388 - INFO - [Train] step: 75999, loss: 0.107061, lr: 0.000200
201
+ 2023-08-07 12:16:00,316 - INFO - [Train] step: 76399, loss: 0.097370, lr: 0.000200
202
+ 2023-08-07 12:17:21,601 - INFO - [Train] step: 76799, loss: 0.115616, lr: 0.000200
203
+ 2023-08-07 12:18:43,734 - INFO - [Train] step: 77199, loss: 0.115731, lr: 0.000200
204
+ 2023-08-07 12:20:05,703 - INFO - [Train] step: 77599, loss: 0.126601, lr: 0.000200
205
+ 2023-08-07 12:21:27,856 - INFO - [Train] step: 77999, loss: 0.099151, lr: 0.000200
206
+ 2023-08-07 12:22:50,083 - INFO - [Train] step: 78399, loss: 0.098982, lr: 0.000200
207
+ 2023-08-07 12:24:12,235 - INFO - [Train] step: 78799, loss: 0.108005, lr: 0.000200
208
+ 2023-08-07 12:25:34,144 - INFO - [Train] step: 79199, loss: 0.110650, lr: 0.000200
209
+ 2023-08-07 12:26:56,289 - INFO - [Train] step: 79599, loss: 0.077124, lr: 0.000200
210
+ 2023-08-07 12:28:18,061 - INFO - [Train] step: 79999, loss: 0.119822, lr: 0.000200
211
+ 2023-08-07 12:30:03,862 - INFO - [Train] step: 80399, loss: 0.117670, lr: 0.000200
212
+ 2023-08-07 12:31:25,371 - INFO - [Train] step: 80799, loss: 0.116992, lr: 0.000200
213
+ 2023-08-07 12:32:47,496 - INFO - [Train] step: 81199, loss: 0.079119, lr: 0.000200
214
+ 2023-08-07 12:34:09,635 - INFO - [Train] step: 81599, loss: 0.106365, lr: 0.000200
215
+ 2023-08-07 12:35:31,786 - INFO - [Train] step: 81999, loss: 0.120059, lr: 0.000200
216
+ 2023-08-07 12:36:53,317 - INFO - [Train] step: 82399, loss: 0.086176, lr: 0.000200
217
+ 2023-08-07 12:38:15,182 - INFO - [Train] step: 82799, loss: 0.096099, lr: 0.000200
218
+ 2023-08-07 12:39:37,105 - INFO - [Train] step: 83199, loss: 0.084154, lr: 0.000200
219
+ 2023-08-07 12:40:59,243 - INFO - [Train] step: 83599, loss: 0.121638, lr: 0.000200
220
+ 2023-08-07 12:42:21,342 - INFO - [Train] step: 83999, loss: 0.117321, lr: 0.000200
221
+ 2023-08-07 12:43:43,387 - INFO - [Train] step: 84399, loss: 0.106803, lr: 0.000200
222
+ 2023-08-07 12:45:05,578 - INFO - [Train] step: 84799, loss: 0.083176, lr: 0.000200
223
+ 2023-08-07 12:46:49,266 - INFO - [Train] step: 85199, loss: 0.087706, lr: 0.000200
224
+ 2023-08-07 12:48:11,556 - INFO - [Train] step: 85599, loss: 0.111652, lr: 0.000200
225
+ 2023-08-07 12:49:33,502 - INFO - [Train] step: 85999, loss: 0.075117, lr: 0.000200
226
+ 2023-08-07 12:50:55,682 - INFO - [Train] step: 86399, loss: 0.094607, lr: 0.000200
227
+ 2023-08-07 12:52:17,342 - INFO - [Train] step: 86799, loss: 0.163171, lr: 0.000200
228
+ 2023-08-07 12:53:38,848 - INFO - [Train] step: 87199, loss: 0.125229, lr: 0.000200
229
+ 2023-08-07 12:55:01,179 - INFO - [Train] step: 87599, loss: 0.128466, lr: 0.000200
230
+ 2023-08-07 12:56:23,496 - INFO - [Train] step: 87999, loss: 0.108964, lr: 0.000200
231
+ 2023-08-07 12:57:45,834 - INFO - [Train] step: 88399, loss: 0.099056, lr: 0.000200
232
+ 2023-08-07 12:59:07,740 - INFO - [Train] step: 88799, loss: 0.084511, lr: 0.000200
233
+ 2023-08-07 13:00:29,869 - INFO - [Train] step: 89199, loss: 0.104628, lr: 0.000200
234
+ 2023-08-07 13:01:51,611 - INFO - [Train] step: 89599, loss: 0.075985, lr: 0.000200
235
+ 2023-08-07 13:03:13,539 - INFO - [Train] step: 89999, loss: 0.107070, lr: 0.000200
236
+ 2023-08-07 13:04:58,186 - INFO - [Train] step: 90399, loss: 0.135741, lr: 0.000200
237
+ 2023-08-07 13:06:20,294 - INFO - [Train] step: 90799, loss: 0.130711, lr: 0.000200
238
+ 2023-08-07 13:07:42,443 - INFO - [Train] step: 91199, loss: 0.087756, lr: 0.000200
239
+ 2023-08-07 13:09:04,365 - INFO - [Train] step: 91599, loss: 0.117803, lr: 0.000200
240
+ 2023-08-07 13:10:26,489 - INFO - [Train] step: 91999, loss: 0.070383, lr: 0.000200
241
+ 2023-08-07 13:11:48,803 - INFO - [Train] step: 92399, loss: 0.110226, lr: 0.000200
242
+ 2023-08-07 13:13:11,136 - INFO - [Train] step: 92799, loss: 0.104816, lr: 0.000200
243
+ 2023-08-07 13:14:33,057 - INFO - [Train] step: 93199, loss: 0.130824, lr: 0.000200
244
+ 2023-08-07 13:15:55,293 - INFO - [Train] step: 93599, loss: 0.091706, lr: 0.000200
245
+ 2023-08-07 13:17:17,910 - INFO - [Train] step: 93999, loss: 0.108431, lr: 0.000200
246
+ 2023-08-07 13:18:40,259 - INFO - [Train] step: 94399, loss: 0.095255, lr: 0.000200
247
+ 2023-08-07 13:20:01,989 - INFO - [Train] step: 94799, loss: 0.128445, lr: 0.000200
248
+ 2023-08-07 13:21:45,582 - INFO - [Train] step: 95199, loss: 0.136346, lr: 0.000200
249
+ 2023-08-07 13:23:07,352 - INFO - [Train] step: 95599, loss: 0.108398, lr: 0.000200
250
+ 2023-08-07 13:24:29,052 - INFO - [Train] step: 95999, loss: 0.102946, lr: 0.000200
251
+ 2023-08-07 13:25:51,135 - INFO - [Train] step: 96399, loss: 0.115570, lr: 0.000200
252
+ 2023-08-07 13:27:13,484 - INFO - [Train] step: 96799, loss: 0.134736, lr: 0.000200
253
+ 2023-08-07 13:28:35,801 - INFO - [Train] step: 97199, loss: 0.121527, lr: 0.000200
254
+ 2023-08-07 13:29:56,938 - INFO - [Train] step: 97599, loss: 0.087732, lr: 0.000200
255
+ 2023-08-07 13:31:19,072 - INFO - [Train] step: 97999, loss: 0.106258, lr: 0.000200
256
+ 2023-08-07 13:32:41,389 - INFO - [Train] step: 98399, loss: 0.102467, lr: 0.000200
257
+ 2023-08-07 13:34:03,501 - INFO - [Train] step: 98799, loss: 0.112881, lr: 0.000200
258
+ 2023-08-07 13:35:25,325 - INFO - [Train] step: 99199, loss: 0.099730, lr: 0.000200
259
+ 2023-08-07 13:36:47,359 - INFO - [Train] step: 99599, loss: 0.132975, lr: 0.000200
260
+ 2023-08-07 13:38:09,688 - INFO - [Train] step: 99999, loss: 0.109757, lr: 0.000200
261
+ 2023-08-07 13:39:55,393 - INFO - [Train] step: 100399, loss: 0.119519, lr: 0.000200
262
+ 2023-08-07 13:41:17,492 - INFO - [Train] step: 100799, loss: 0.095384, lr: 0.000200
263
+ 2023-08-07 13:42:39,812 - INFO - [Train] step: 101199, loss: 0.094317, lr: 0.000200
264
+ 2023-08-07 13:44:01,951 - INFO - [Train] step: 101599, loss: 0.077808, lr: 0.000200
265
+ 2023-08-07 13:45:24,292 - INFO - [Train] step: 101999, loss: 0.131864, lr: 0.000200
266
+ 2023-08-07 13:46:46,178 - INFO - [Train] step: 102399, loss: 0.104115, lr: 0.000200
267
+ 2023-08-07 13:48:08,526 - INFO - [Train] step: 102799, loss: 0.118189, lr: 0.000200
268
+ 2023-08-07 13:49:30,230 - INFO - [Train] step: 103199, loss: 0.124610, lr: 0.000200
269
+ 2023-08-07 13:50:52,359 - INFO - [Train] step: 103599, loss: 0.106008, lr: 0.000200
270
+ 2023-08-07 13:52:14,291 - INFO - [Train] step: 103999, loss: 0.086552, lr: 0.000200
271
+ 2023-08-07 13:53:36,466 - INFO - [Train] step: 104399, loss: 0.091638, lr: 0.000200
272
+ 2023-08-07 13:54:57,932 - INFO - [Train] step: 104799, loss: 0.113715, lr: 0.000200
273
+ 2023-08-07 13:56:42,356 - INFO - [Train] step: 105199, loss: 0.101307, lr: 0.000200
274
+ 2023-08-07 13:58:04,313 - INFO - [Train] step: 105599, loss: 0.086030, lr: 0.000200
275
+ 2023-08-07 13:59:25,996 - INFO - [Train] step: 105999, loss: 0.110670, lr: 0.000200
276
+ 2023-08-07 14:00:48,345 - INFO - [Train] step: 106399, loss: 0.105311, lr: 0.000200
277
+ 2023-08-07 14:02:10,689 - INFO - [Train] step: 106799, loss: 0.143710, lr: 0.000200
278
+ 2023-08-07 14:03:32,975 - INFO - [Train] step: 107199, loss: 0.104342, lr: 0.000200
279
+ 2023-08-07 14:04:54,932 - INFO - [Train] step: 107599, loss: 0.128174, lr: 0.000200
280
+ 2023-08-07 14:06:16,847 - INFO - [Train] step: 107999, loss: 0.108958, lr: 0.000200
281
+ 2023-08-07 14:07:38,959 - INFO - [Train] step: 108399, loss: 0.117044, lr: 0.000200
282
+ 2023-08-07 14:09:01,105 - INFO - [Train] step: 108799, loss: 0.104867, lr: 0.000200
283
+ 2023-08-07 14:10:23,319 - INFO - [Train] step: 109199, loss: 0.118993, lr: 0.000200
284
+ 2023-08-07 14:11:45,736 - INFO - [Train] step: 109599, loss: 0.107032, lr: 0.000200
285
+ 2023-08-07 14:13:08,073 - INFO - [Train] step: 109999, loss: 0.119035, lr: 0.000200
286
+ 2023-08-07 14:14:53,768 - INFO - [Train] step: 110399, loss: 0.123649, lr: 0.000200
287
+ 2023-08-07 14:16:15,699 - INFO - [Train] step: 110799, loss: 0.115585, lr: 0.000200
288
+ 2023-08-07 14:17:38,027 - INFO - [Train] step: 111199, loss: 0.114956, lr: 0.000200
289
+ 2023-08-07 14:19:00,161 - INFO - [Train] step: 111599, loss: 0.074508, lr: 0.000200
290
+ 2023-08-07 14:20:22,478 - INFO - [Train] step: 111999, loss: 0.136996, lr: 0.000200
291
+ 2023-08-07 14:21:44,817 - INFO - [Train] step: 112399, loss: 0.118663, lr: 0.000200
292
+ 2023-08-07 14:23:06,937 - INFO - [Train] step: 112799, loss: 0.153161, lr: 0.000200
293
+ 2023-08-07 14:24:28,859 - INFO - [Train] step: 113199, loss: 0.111933, lr: 0.000200
294
+ 2023-08-07 14:25:50,374 - INFO - [Train] step: 113599, loss: 0.082598, lr: 0.000200
295
+ 2023-08-07 14:27:12,053 - INFO - [Train] step: 113999, loss: 0.120279, lr: 0.000200
296
+ 2023-08-07 14:28:34,077 - INFO - [Train] step: 114399, loss: 0.075852, lr: 0.000200
297
+ 2023-08-07 14:29:56,128 - INFO - [Train] step: 114799, loss: 0.094710, lr: 0.000200
298
+ 2023-08-07 14:31:40,171 - INFO - [Train] step: 115199, loss: 0.124519, lr: 0.000200
299
+ 2023-08-07 14:33:02,143 - INFO - [Train] step: 115599, loss: 0.103630, lr: 0.000200
300
+ 2023-08-07 14:34:24,194 - INFO - [Train] step: 115999, loss: 0.104483, lr: 0.000200
301
+ 2023-08-07 14:35:46,364 - INFO - [Train] step: 116399, loss: 0.133401, lr: 0.000200
302
+ 2023-08-07 14:37:08,208 - INFO - [Train] step: 116799, loss: 0.128127, lr: 0.000200
303
+ 2023-08-07 14:38:29,774 - INFO - [Train] step: 117199, loss: 0.116914, lr: 0.000200
304
+ 2023-08-07 14:39:52,111 - INFO - [Train] step: 117599, loss: 0.115837, lr: 0.000200
305
+ 2023-08-07 14:41:14,440 - INFO - [Train] step: 117999, loss: 0.135401, lr: 0.000200
306
+ 2023-08-07 14:42:35,960 - INFO - [Train] step: 118399, loss: 0.123450, lr: 0.000200
307
+ 2023-08-07 14:43:57,870 - INFO - [Train] step: 118799, loss: 0.125168, lr: 0.000200
308
+ 2023-08-07 14:45:19,982 - INFO - [Train] step: 119199, loss: 0.102492, lr: 0.000200
309
+ 2023-08-07 14:46:42,301 - INFO - [Train] step: 119599, loss: 0.093711, lr: 0.000200
310
+ 2023-08-07 14:48:04,397 - INFO - [Train] step: 119999, loss: 0.128934, lr: 0.000200
311
+ 2023-08-07 14:49:48,874 - INFO - [Train] step: 120399, loss: 0.119873, lr: 0.000200
312
+ 2023-08-07 14:51:10,594 - INFO - [Train] step: 120799, loss: 0.112037, lr: 0.000200
313
+ 2023-08-07 14:52:32,688 - INFO - [Train] step: 121199, loss: 0.108190, lr: 0.000200
314
+ 2023-08-07 14:53:55,035 - INFO - [Train] step: 121599, loss: 0.115544, lr: 0.000200
315
+ 2023-08-07 14:55:16,909 - INFO - [Train] step: 121999, loss: 0.091772, lr: 0.000200
316
+ 2023-08-07 14:56:38,769 - INFO - [Train] step: 122399, loss: 0.166007, lr: 0.000200
317
+ 2023-08-07 14:58:00,993 - INFO - [Train] step: 122799, loss: 0.102445, lr: 0.000200
318
+ 2023-08-07 14:59:22,894 - INFO - [Train] step: 123199, loss: 0.104988, lr: 0.000200
319
+ 2023-08-07 15:00:45,026 - INFO - [Train] step: 123599, loss: 0.103103, lr: 0.000200
320
+ 2023-08-07 15:02:07,329 - INFO - [Train] step: 123999, loss: 0.094463, lr: 0.000200
321
+ 2023-08-07 15:03:29,469 - INFO - [Train] step: 124399, loss: 0.111020, lr: 0.000200
322
+ 2023-08-07 15:04:51,914 - INFO - [Train] step: 124799, loss: 0.133209, lr: 0.000200
323
+ 2023-08-07 15:06:36,109 - INFO - [Train] step: 125199, loss: 0.117668, lr: 0.000200
324
+ 2023-08-07 15:07:57,989 - INFO - [Train] step: 125599, loss: 0.109193, lr: 0.000200
325
+ 2023-08-07 15:09:19,295 - INFO - [Train] step: 125999, loss: 0.131822, lr: 0.000200
326
+ 2023-08-07 15:10:41,415 - INFO - [Train] step: 126399, loss: 0.105944, lr: 0.000200
327
+ 2023-08-07 15:12:03,779 - INFO - [Train] step: 126799, loss: 0.104178, lr: 0.000200
328
+ 2023-08-07 15:13:26,058 - INFO - [Train] step: 127199, loss: 0.101726, lr: 0.000200
329
+ 2023-08-07 15:14:47,982 - INFO - [Train] step: 127599, loss: 0.156785, lr: 0.000200
330
+ 2023-08-07 15:16:09,905 - INFO - [Train] step: 127999, loss: 0.099558, lr: 0.000200
331
+ 2023-08-07 15:17:32,029 - INFO - [Train] step: 128399, loss: 0.121172, lr: 0.000200
332
+ 2023-08-07 15:18:54,346 - INFO - [Train] step: 128799, loss: 0.088598, lr: 0.000200
333
+ 2023-08-07 15:20:16,545 - INFO - [Train] step: 129199, loss: 0.097450, lr: 0.000200
334
+ 2023-08-07 15:21:38,819 - INFO - [Train] step: 129599, loss: 0.118880, lr: 0.000200
335
+ 2023-08-07 15:23:01,154 - INFO - [Train] step: 129999, loss: 0.115626, lr: 0.000200
336
+ 2023-08-07 15:24:47,652 - INFO - [Train] step: 130399, loss: 0.078181, lr: 0.000200
337
+ 2023-08-07 15:26:09,359 - INFO - [Train] step: 130799, loss: 0.106813, lr: 0.000200
338
+ 2023-08-07 15:27:31,281 - INFO - [Train] step: 131199, loss: 0.139148, lr: 0.000200
339
+ 2023-08-07 15:28:53,612 - INFO - [Train] step: 131599, loss: 0.125892, lr: 0.000200
340
+ 2023-08-07 15:30:15,717 - INFO - [Train] step: 131999, loss: 0.087311, lr: 0.000200
341
+ 2023-08-07 15:31:38,050 - INFO - [Train] step: 132399, loss: 0.097564, lr: 0.000200
342
+ 2023-08-07 15:33:00,183 - INFO - [Train] step: 132799, loss: 0.105457, lr: 0.000200
343
+ 2023-08-07 15:34:22,328 - INFO - [Train] step: 133199, loss: 0.082838, lr: 0.000200
344
+ 2023-08-07 15:35:44,676 - INFO - [Train] step: 133599, loss: 0.124125, lr: 0.000200
345
+ 2023-08-07 15:37:06,618 - INFO - [Train] step: 133999, loss: 0.119264, lr: 0.000200
346
+ 2023-08-07 15:38:28,496 - INFO - [Train] step: 134399, loss: 0.106483, lr: 0.000200
347
+ 2023-08-07 15:39:50,665 - INFO - [Train] step: 134799, loss: 0.108060, lr: 0.000200
348
+ 2023-08-07 15:41:34,928 - INFO - [Train] step: 135199, loss: 0.080738, lr: 0.000200
349
+ 2023-08-07 15:42:56,809 - INFO - [Train] step: 135599, loss: 0.093693, lr: 0.000200
350
+ 2023-08-07 15:44:18,543 - INFO - [Train] step: 135999, loss: 0.121478, lr: 0.000200
351
+ 2023-08-07 15:45:40,902 - INFO - [Train] step: 136399, loss: 0.106023, lr: 0.000200
352
+ 2023-08-07 15:47:03,256 - INFO - [Train] step: 136799, loss: 0.101689, lr: 0.000200
353
+ 2023-08-07 15:48:25,606 - INFO - [Train] step: 137199, loss: 0.112455, lr: 0.000200
354
+ 2023-08-07 15:49:47,114 - INFO - [Train] step: 137599, loss: 0.139700, lr: 0.000200
355
+ 2023-08-07 15:51:08,849 - INFO - [Train] step: 137999, loss: 0.087761, lr: 0.000200
356
+ 2023-08-07 15:52:30,964 - INFO - [Train] step: 138399, loss: 0.159153, lr: 0.000200
357
+ 2023-08-07 15:53:53,280 - INFO - [Train] step: 138799, loss: 0.083335, lr: 0.000200
358
+ 2023-08-07 15:55:15,426 - INFO - [Train] step: 139199, loss: 0.101326, lr: 0.000200
359
+ 2023-08-07 15:56:37,107 - INFO - [Train] step: 139599, loss: 0.127182, lr: 0.000200
360
+ 2023-08-07 15:57:59,075 - INFO - [Train] step: 139999, loss: 0.096868, lr: 0.000200
361
+ 2023-08-07 15:59:43,592 - INFO - [Train] step: 140399, loss: 0.113807, lr: 0.000200
362
+ 2023-08-07 16:01:05,462 - INFO - [Train] step: 140799, loss: 0.093394, lr: 0.000200
363
+ 2023-08-07 16:02:27,530 - INFO - [Train] step: 141199, loss: 0.102109, lr: 0.000200
364
+ 2023-08-07 16:03:49,820 - INFO - [Train] step: 141599, loss: 0.113728, lr: 0.000200
365
+ 2023-08-07 16:05:12,163 - INFO - [Train] step: 141999, loss: 0.130736, lr: 0.000200
366
+ 2023-08-07 16:06:34,274 - INFO - [Train] step: 142399, loss: 0.124864, lr: 0.000200
367
+ 2023-08-07 16:07:56,520 - INFO - [Train] step: 142799, loss: 0.097180, lr: 0.000200
368
+ 2023-08-07 16:09:18,718 - INFO - [Train] step: 143199, loss: 0.115004, lr: 0.000200
369
+ 2023-08-07 16:10:40,848 - INFO - [Train] step: 143599, loss: 0.113431, lr: 0.000200
370
+ 2023-08-07 16:12:02,837 - INFO - [Train] step: 143999, loss: 0.136714, lr: 0.000200
371
+ 2023-08-07 16:13:25,091 - INFO - [Train] step: 144399, loss: 0.104968, lr: 0.000200
372
+ 2023-08-07 16:14:47,021 - INFO - [Train] step: 144799, loss: 0.129130, lr: 0.000200
373
+ 2023-08-07 16:16:30,050 - INFO - [Train] step: 145199, loss: 0.090315, lr: 0.000200
374
+ 2023-08-07 16:17:52,154 - INFO - [Train] step: 145599, loss: 0.094249, lr: 0.000200
375
+ 2023-08-07 16:19:14,528 - INFO - [Train] step: 145999, loss: 0.085897, lr: 0.000200
376
+ 2023-08-07 16:20:36,643 - INFO - [Train] step: 146399, loss: 0.135044, lr: 0.000200
377
+ 2023-08-07 16:21:58,717 - INFO - [Train] step: 146799, loss: 0.162018, lr: 0.000200
378
+ 2023-08-07 16:23:20,478 - INFO - [Train] step: 147199, loss: 0.157382, lr: 0.000200
379
+ 2023-08-07 16:24:42,399 - INFO - [Train] step: 147599, loss: 0.096387, lr: 0.000200
380
+ 2023-08-07 16:26:04,108 - INFO - [Train] step: 147999, loss: 0.146949, lr: 0.000200
381
+ 2023-08-07 16:27:26,450 - INFO - [Train] step: 148399, loss: 0.117472, lr: 0.000200
382
+ 2023-08-07 16:28:48,560 - INFO - [Train] step: 148799, loss: 0.073945, lr: 0.000200
383
+ 2023-08-07 16:30:10,900 - INFO - [Train] step: 149199, loss: 0.106740, lr: 0.000200
384
+ 2023-08-07 16:31:32,589 - INFO - [Train] step: 149599, loss: 0.080322, lr: 0.000200
385
+ 2023-08-07 16:32:54,112 - INFO - [Train] step: 149999, loss: 0.136559, lr: 0.000200
386
+ 2023-08-07 16:34:39,170 - INFO - [Train] step: 150399, loss: 0.126293, lr: 0.000200
387
+ 2023-08-07 16:36:01,082 - INFO - [Train] step: 150799, loss: 0.114251, lr: 0.000200
388
+ 2023-08-07 16:37:22,783 - INFO - [Train] step: 151199, loss: 0.096746, lr: 0.000200
389
+ 2023-08-07 16:38:44,927 - INFO - [Train] step: 151599, loss: 0.108269, lr: 0.000200
390
+ 2023-08-07 16:40:06,611 - INFO - [Train] step: 151999, loss: 0.086108, lr: 0.000200
391
+ 2023-08-07 16:41:28,580 - INFO - [Train] step: 152399, loss: 0.112391, lr: 0.000200
392
+ 2023-08-07 16:42:50,904 - INFO - [Train] step: 152799, loss: 0.101756, lr: 0.000200
393
+ 2023-08-07 16:44:13,030 - INFO - [Train] step: 153199, loss: 0.136925, lr: 0.000200
394
+ 2023-08-07 16:45:34,567 - INFO - [Train] step: 153599, loss: 0.077758, lr: 0.000200
395
+ 2023-08-07 16:46:56,040 - INFO - [Train] step: 153999, loss: 0.122852, lr: 0.000200
396
+ 2023-08-07 16:48:18,189 - INFO - [Train] step: 154399, loss: 0.087679, lr: 0.000200
397
+ 2023-08-07 16:49:40,284 - INFO - [Train] step: 154799, loss: 0.094326, lr: 0.000200
398
+ 2023-08-07 16:51:24,907 - INFO - [Train] step: 155199, loss: 0.128245, lr: 0.000200
399
+ 2023-08-07 16:52:47,084 - INFO - [Train] step: 155599, loss: 0.123441, lr: 0.000200
400
+ 2023-08-07 16:54:09,313 - INFO - [Train] step: 155999, loss: 0.129279, lr: 0.000200
401
+ 2023-08-07 16:55:31,350 - INFO - [Train] step: 156399, loss: 0.132648, lr: 0.000200
402
+ 2023-08-07 16:56:53,699 - INFO - [Train] step: 156799, loss: 0.138483, lr: 0.000200
403
+ 2023-08-07 16:58:15,574 - INFO - [Train] step: 157199, loss: 0.082590, lr: 0.000200
404
+ 2023-08-07 16:59:37,897 - INFO - [Train] step: 157599, loss: 0.108296, lr: 0.000200
405
+ 2023-08-07 17:01:00,210 - INFO - [Train] step: 157999, loss: 0.116951, lr: 0.000200
406
+ 2023-08-07 17:02:22,364 - INFO - [Train] step: 158399, loss: 0.103571, lr: 0.000200
407
+ 2023-08-07 17:03:44,491 - INFO - [Train] step: 158799, loss: 0.096296, lr: 0.000200
408
+ 2023-08-07 17:05:06,286 - INFO - [Train] step: 159199, loss: 0.123831, lr: 0.000200
409
+ 2023-08-07 17:06:28,540 - INFO - [Train] step: 159599, loss: 0.100907, lr: 0.000200
410
+ 2023-08-07 17:07:50,459 - INFO - [Train] step: 159999, loss: 0.135853, lr: 0.000200
411
+ 2023-08-07 17:09:34,690 - INFO - [Train] step: 160399, loss: 0.101097, lr: 0.000200
412
+ 2023-08-07 17:10:57,052 - INFO - [Train] step: 160799, loss: 0.121195, lr: 0.000200
413
+ 2023-08-07 17:12:19,350 - INFO - [Train] step: 161199, loss: 0.102940, lr: 0.000200
414
+ 2023-08-07 17:13:41,486 - INFO - [Train] step: 161599, loss: 0.102387, lr: 0.000200
415
+ 2023-08-07 17:15:02,802 - INFO - [Train] step: 161999, loss: 0.124292, lr: 0.000200
416
+ 2023-08-07 17:16:25,109 - INFO - [Train] step: 162399, loss: 0.115599, lr: 0.000200
417
+ 2023-08-07 17:17:47,453 - INFO - [Train] step: 162799, loss: 0.142791, lr: 0.000200
418
+ 2023-08-07 17:19:09,598 - INFO - [Train] step: 163199, loss: 0.090653, lr: 0.000200
419
+ 2023-08-07 17:20:31,786 - INFO - [Train] step: 163599, loss: 0.110355, lr: 0.000200
420
+ 2023-08-07 17:21:54,047 - INFO - [Train] step: 163999, loss: 0.131236, lr: 0.000200
421
+ 2023-08-07 17:23:16,339 - INFO - [Train] step: 164399, loss: 0.153339, lr: 0.000200
422
+ 2023-08-07 17:24:38,703 - INFO - [Train] step: 164799, loss: 0.109122, lr: 0.000200
423
+ 2023-08-07 17:26:22,327 - INFO - [Train] step: 165199, loss: 0.088501, lr: 0.000200
424
+ 2023-08-07 17:27:44,041 - INFO - [Train] step: 165599, loss: 0.105226, lr: 0.000200
425
+ 2023-08-07 17:29:06,114 - INFO - [Train] step: 165999, loss: 0.141148, lr: 0.000200
426
+ 2023-08-07 17:30:28,259 - INFO - [Train] step: 166399, loss: 0.092174, lr: 0.000200
427
+ 2023-08-07 17:31:50,574 - INFO - [Train] step: 166799, loss: 0.123589, lr: 0.000200
428
+ 2023-08-07 17:33:12,117 - INFO - [Train] step: 167199, loss: 0.117268, lr: 0.000200
429
+ 2023-08-07 17:34:33,830 - INFO - [Train] step: 167599, loss: 0.096447, lr: 0.000200
430
+ 2023-08-07 17:35:56,200 - INFO - [Train] step: 167999, loss: 0.132877, lr: 0.000200
431
+ 2023-08-07 17:37:18,305 - INFO - [Train] step: 168399, loss: 0.117735, lr: 0.000200
432
+ 2023-08-07 17:38:40,655 - INFO - [Train] step: 168799, loss: 0.079842, lr: 0.000200
433
+ 2023-08-07 17:40:02,991 - INFO - [Train] step: 169199, loss: 0.090230, lr: 0.000200
434
+ 2023-08-07 17:41:24,898 - INFO - [Train] step: 169599, loss: 0.099085, lr: 0.000200
435
+ 2023-08-07 17:42:47,007 - INFO - [Train] step: 169999, loss: 0.118393, lr: 0.000200
436
+ 2023-08-07 17:44:32,501 - INFO - [Train] step: 170399, loss: 0.149631, lr: 0.000200
437
+ 2023-08-07 17:45:54,845 - INFO - [Train] step: 170799, loss: 0.094728, lr: 0.000200
438
+ 2023-08-07 17:47:16,557 - INFO - [Train] step: 171199, loss: 0.113340, lr: 0.000200
439
+ 2023-08-07 17:48:38,365 - INFO - [Train] step: 171599, loss: 0.159166, lr: 0.000200
440
+ 2023-08-07 17:50:00,848 - INFO - [Train] step: 171999, loss: 0.122926, lr: 0.000200
441
+ 2023-08-07 17:51:22,717 - INFO - [Train] step: 172399, loss: 0.095736, lr: 0.000200
442
+ 2023-08-07 17:52:44,864 - INFO - [Train] step: 172799, loss: 0.130558, lr: 0.000200
443
+ 2023-08-07 17:54:07,182 - INFO - [Train] step: 173199, loss: 0.121787, lr: 0.000200
444
+ 2023-08-07 17:55:28,680 - INFO - [Train] step: 173599, loss: 0.103055, lr: 0.000200
445
+ 2023-08-07 17:56:50,465 - INFO - [Train] step: 173999, loss: 0.102579, lr: 0.000200
446
+ 2023-08-07 17:58:12,732 - INFO - [Train] step: 174399, loss: 0.126776, lr: 0.000200
447
+ 2023-08-07 17:59:34,449 - INFO - [Train] step: 174799, loss: 0.117643, lr: 0.000200
448
+ 2023-08-07 18:01:18,510 - INFO - [Train] step: 175199, loss: 0.095255, lr: 0.000200
449
+ 2023-08-07 18:02:40,832 - INFO - [Train] step: 175599, loss: 0.103943, lr: 0.000200
450
+ 2023-08-07 18:04:02,954 - INFO - [Train] step: 175999, loss: 0.114158, lr: 0.000200
451
+ 2023-08-07 18:05:25,091 - INFO - [Train] step: 176399, loss: 0.111492, lr: 0.000200
452
+ 2023-08-07 18:06:46,954 - INFO - [Train] step: 176799, loss: 0.121293, lr: 0.000200
453
+ 2023-08-07 18:08:08,711 - INFO - [Train] step: 177199, loss: 0.090264, lr: 0.000200
454
+ 2023-08-07 18:09:30,853 - INFO - [Train] step: 177599, loss: 0.100538, lr: 0.000200
455
+ 2023-08-07 18:10:52,840 - INFO - [Train] step: 177999, loss: 0.099849, lr: 0.000200
456
+ 2023-08-07 18:12:14,466 - INFO - [Train] step: 178399, loss: 0.107672, lr: 0.000200
457
+ 2023-08-07 18:13:36,154 - INFO - [Train] step: 178799, loss: 0.144601, lr: 0.000200
458
+ 2023-08-07 18:14:58,294 - INFO - [Train] step: 179199, loss: 0.100732, lr: 0.000200
459
+ 2023-08-07 18:16:20,225 - INFO - [Train] step: 179599, loss: 0.092496, lr: 0.000200
460
+ 2023-08-07 18:17:42,326 - INFO - [Train] step: 179999, loss: 0.111023, lr: 0.000200
461
+ 2023-08-07 18:19:26,804 - INFO - [Train] step: 180399, loss: 0.148467, lr: 0.000200
462
+ 2023-08-07 18:20:49,109 - INFO - [Train] step: 180799, loss: 0.092455, lr: 0.000200
463
+ 2023-08-07 18:22:11,032 - INFO - [Train] step: 181199, loss: 0.102241, lr: 0.000200
464
+ 2023-08-07 18:23:32,939 - INFO - [Train] step: 181599, loss: 0.108256, lr: 0.000200
465
+ 2023-08-07 18:24:55,283 - INFO - [Train] step: 181999, loss: 0.094663, lr: 0.000200
466
+ 2023-08-07 18:26:17,390 - INFO - [Train] step: 182399, loss: 0.108905, lr: 0.000200
467
+ 2023-08-07 18:27:39,101 - INFO - [Train] step: 182799, loss: 0.129013, lr: 0.000200
468
+ 2023-08-07 18:29:01,238 - INFO - [Train] step: 183199, loss: 0.124781, lr: 0.000200
469
+ 2023-08-07 18:30:22,754 - INFO - [Train] step: 183599, loss: 0.117572, lr: 0.000200
470
+ 2023-08-07 18:31:44,269 - INFO - [Train] step: 183999, loss: 0.105549, lr: 0.000200
471
+ 2023-08-07 18:33:06,548 - INFO - [Train] step: 184399, loss: 0.105690, lr: 0.000200
472
+ 2023-08-07 18:34:28,725 - INFO - [Train] step: 184799, loss: 0.131030, lr: 0.000200
473
+ 2023-08-07 18:36:13,991 - INFO - [Train] step: 185199, loss: 0.116739, lr: 0.000200
474
+ 2023-08-07 18:37:36,333 - INFO - [Train] step: 185599, loss: 0.138103, lr: 0.000200
475
+ 2023-08-07 18:38:57,994 - INFO - [Train] step: 185999, loss: 0.092167, lr: 0.000200
476
+ 2023-08-07 18:40:19,745 - INFO - [Train] step: 186399, loss: 0.137942, lr: 0.000200
477
+ 2023-08-07 18:41:41,866 - INFO - [Train] step: 186799, loss: 0.082331, lr: 0.000200
478
+ 2023-08-07 18:43:03,889 - INFO - [Train] step: 187199, loss: 0.095830, lr: 0.000200
479
+ 2023-08-07 18:44:26,118 - INFO - [Train] step: 187599, loss: 0.100209, lr: 0.000200
480
+ 2023-08-07 18:45:48,474 - INFO - [Train] step: 187999, loss: 0.114014, lr: 0.000200
481
+ 2023-08-07 18:47:10,765 - INFO - [Train] step: 188399, loss: 0.142246, lr: 0.000200
482
+ 2023-08-07 18:48:32,925 - INFO - [Train] step: 188799, loss: 0.106054, lr: 0.000200
483
+ 2023-08-07 18:49:55,076 - INFO - [Train] step: 189199, loss: 0.111874, lr: 0.000200
484
+ 2023-08-07 18:51:16,942 - INFO - [Train] step: 189599, loss: 0.101729, lr: 0.000200
485
+ 2023-08-07 18:52:38,308 - INFO - [Train] step: 189999, loss: 0.122955, lr: 0.000200
486
+ 2023-08-07 18:54:22,743 - INFO - [Train] step: 190399, loss: 0.085192, lr: 0.000200
487
+ 2023-08-07 18:55:44,874 - INFO - [Train] step: 190799, loss: 0.088300, lr: 0.000200
488
+ 2023-08-07 18:57:06,561 - INFO - [Train] step: 191199, loss: 0.127417, lr: 0.000200
489
+ 2023-08-07 18:58:28,136 - INFO - [Train] step: 191599, loss: 0.123584, lr: 0.000200
490
+ 2023-08-07 18:59:50,426 - INFO - [Train] step: 191999, loss: 0.127683, lr: 0.000200
491
+ 2023-08-07 19:01:11,914 - INFO - [Train] step: 192399, loss: 0.133990, lr: 0.000200
492
+ 2023-08-07 19:02:34,031 - INFO - [Train] step: 192799, loss: 0.120536, lr: 0.000200
493
+ 2023-08-07 19:03:56,193 - INFO - [Train] step: 193199, loss: 0.083013, lr: 0.000200
494
+ 2023-08-07 19:05:17,503 - INFO - [Train] step: 193599, loss: 0.119776, lr: 0.000200
495
+ 2023-08-07 19:06:39,407 - INFO - [Train] step: 193999, loss: 0.103435, lr: 0.000200
496
+ 2023-08-07 19:08:00,508 - INFO - [Train] step: 194399, loss: 0.118171, lr: 0.000200
497
+ 2023-08-07 19:09:22,647 - INFO - [Train] step: 194799, loss: 0.097865, lr: 0.000200
498
+ 2023-08-07 19:11:06,494 - INFO - [Train] step: 195199, loss: 0.128009, lr: 0.000200
499
+ 2023-08-07 19:12:27,998 - INFO - [Train] step: 195599, loss: 0.093132, lr: 0.000200
500
+ 2023-08-07 19:13:50,097 - INFO - [Train] step: 195999, loss: 0.092939, lr: 0.000200
501
+ 2023-08-07 19:15:11,619 - INFO - [Train] step: 196399, loss: 0.074518, lr: 0.000200
502
+ 2023-08-07 19:16:33,364 - INFO - [Train] step: 196799, loss: 0.127847, lr: 0.000200
503
+ 2023-08-07 19:17:54,869 - INFO - [Train] step: 197199, loss: 0.126053, lr: 0.000200
504
+ 2023-08-07 19:19:17,192 - INFO - [Train] step: 197599, loss: 0.104337, lr: 0.000200
505
+ 2023-08-07 19:20:39,328 - INFO - [Train] step: 197999, loss: 0.110391, lr: 0.000200
506
+ 2023-08-07 19:22:00,952 - INFO - [Train] step: 198399, loss: 0.090200, lr: 0.000200
507
+ 2023-08-07 19:23:22,319 - INFO - [Train] step: 198799, loss: 0.098856, lr: 0.000200
508
+ 2023-08-07 19:24:44,281 - INFO - [Train] step: 199199, loss: 0.121909, lr: 0.000200
509
+ 2023-08-07 19:26:06,583 - INFO - [Train] step: 199599, loss: 0.100613, lr: 0.000200
510
+ 2023-08-07 19:27:28,736 - INFO - [Train] step: 199999, loss: 0.099918, lr: 0.000200
511
+ 2023-08-07 19:29:11,578 - INFO - [Train] step: 200399, loss: 0.093422, lr: 0.000200
512
+ 2023-08-07 19:30:33,270 - INFO - [Train] step: 200799, loss: 0.113539, lr: 0.000200
513
+ 2023-08-07 19:31:54,567 - INFO - [Train] step: 201199, loss: 0.106162, lr: 0.000200
514
+ 2023-08-07 19:33:16,881 - INFO - [Train] step: 201599, loss: 0.122963, lr: 0.000200
515
+ 2023-08-07 19:34:39,231 - INFO - [Train] step: 201999, loss: 0.103843, lr: 0.000200
516
+ 2023-08-07 19:36:01,156 - INFO - [Train] step: 202399, loss: 0.112713, lr: 0.000200
517
+ 2023-08-07 19:37:23,580 - INFO - [Train] step: 202799, loss: 0.137966, lr: 0.000200
518
+ 2023-08-07 19:38:45,798 - INFO - [Train] step: 203199, loss: 0.116644, lr: 0.000200
519
+ 2023-08-07 19:40:07,701 - INFO - [Train] step: 203599, loss: 0.102900, lr: 0.000200
520
+ 2023-08-07 19:41:29,629 - INFO - [Train] step: 203999, loss: 0.122038, lr: 0.000200
521
+ 2023-08-07 19:42:51,963 - INFO - [Train] step: 204399, loss: 0.102956, lr: 0.000200
522
+ 2023-08-07 19:44:14,327 - INFO - [Train] step: 204799, loss: 0.093602, lr: 0.000200
523
+ 2023-08-07 19:45:56,730 - INFO - [Train] step: 205199, loss: 0.108089, lr: 0.000200
524
+ 2023-08-07 19:47:18,833 - INFO - [Train] step: 205599, loss: 0.126664, lr: 0.000200
525
+ 2023-08-07 19:48:40,336 - INFO - [Train] step: 205999, loss: 0.123856, lr: 0.000200
526
+ 2023-08-07 19:50:01,843 - INFO - [Train] step: 206399, loss: 0.121010, lr: 0.000200
527
+ 2023-08-07 19:51:23,989 - INFO - [Train] step: 206799, loss: 0.130720, lr: 0.000200
528
+ 2023-08-07 19:52:46,327 - INFO - [Train] step: 207199, loss: 0.127612, lr: 0.000200
529
+ 2023-08-07 19:54:07,464 - INFO - [Train] step: 207599, loss: 0.093583, lr: 0.000200
530
+ 2023-08-07 19:55:29,517 - INFO - [Train] step: 207999, loss: 0.128115, lr: 0.000200
531
+ 2023-08-07 19:56:51,687 - INFO - [Train] step: 208399, loss: 0.129089, lr: 0.000200
532
+ 2023-08-07 19:58:13,715 - INFO - [Train] step: 208799, loss: 0.110647, lr: 0.000200
533
+ 2023-08-07 19:59:35,504 - INFO - [Train] step: 209199, loss: 0.111910, lr: 0.000200
534
+ 2023-08-07 20:00:57,345 - INFO - [Train] step: 209599, loss: 0.103247, lr: 0.000200
535
+ 2023-08-07 20:02:19,349 - INFO - [Train] step: 209999, loss: 0.134336, lr: 0.000200
536
+ 2023-08-07 20:04:03,389 - INFO - [Train] step: 210399, loss: 0.091911, lr: 0.000200
537
+ 2023-08-07 20:05:25,082 - INFO - [Train] step: 210799, loss: 0.107047, lr: 0.000200
538
+ 2023-08-07 20:06:47,209 - INFO - [Train] step: 211199, loss: 0.117183, lr: 0.000200
539
+ 2023-08-07 20:08:09,135 - INFO - [Train] step: 211599, loss: 0.118698, lr: 0.000200
540
+ 2023-08-07 20:09:31,083 - INFO - [Train] step: 211999, loss: 0.113967, lr: 0.000200
541
+ 2023-08-07 20:10:53,410 - INFO - [Train] step: 212399, loss: 0.104207, lr: 0.000200
542
+ 2023-08-07 20:12:14,927 - INFO - [Train] step: 212799, loss: 0.097402, lr: 0.000200
543
+ 2023-08-07 20:13:36,608 - INFO - [Train] step: 213199, loss: 0.119846, lr: 0.000200
544
+ 2023-08-07 20:14:58,968 - INFO - [Train] step: 213599, loss: 0.083921, lr: 0.000200
545
+ 2023-08-07 20:16:21,097 - INFO - [Train] step: 213999, loss: 0.138865, lr: 0.000200
546
+ 2023-08-07 20:17:43,389 - INFO - [Train] step: 214399, loss: 0.099082, lr: 0.000200
547
+ 2023-08-07 20:19:05,557 - INFO - [Train] step: 214799, loss: 0.108562, lr: 0.000200
548
+ 2023-08-07 20:20:48,947 - INFO - [Train] step: 215199, loss: 0.100820, lr: 0.000200
549
+ 2023-08-07 20:22:11,080 - INFO - [Train] step: 215599, loss: 0.099039, lr: 0.000200
550
+ 2023-08-07 20:23:33,423 - INFO - [Train] step: 215999, loss: 0.114374, lr: 0.000200
551
+ 2023-08-07 20:24:55,531 - INFO - [Train] step: 216399, loss: 0.105309, lr: 0.000200
552
+ 2023-08-07 20:26:17,462 - INFO - [Train] step: 216799, loss: 0.153880, lr: 0.000200
553
+ 2023-08-07 20:27:39,809 - INFO - [Train] step: 217199, loss: 0.109010, lr: 0.000200
554
+ 2023-08-07 20:29:01,739 - INFO - [Train] step: 217599, loss: 0.119924, lr: 0.000200
555
+ 2023-08-07 20:30:23,449 - INFO - [Train] step: 217999, loss: 0.123794, lr: 0.000200
556
+ 2023-08-07 20:31:45,893 - INFO - [Train] step: 218399, loss: 0.113580, lr: 0.000200
557
+ 2023-08-07 20:33:08,280 - INFO - [Train] step: 218799, loss: 0.132963, lr: 0.000200
558
+ 2023-08-07 20:34:30,164 - INFO - [Train] step: 219199, loss: 0.122822, lr: 0.000200
559
+ 2023-08-07 20:35:51,602 - INFO - [Train] step: 219599, loss: 0.109602, lr: 0.000200
560
+ 2023-08-07 20:37:13,259 - INFO - [Train] step: 219999, loss: 0.091663, lr: 0.000200
561
+ 2023-08-07 20:38:58,950 - INFO - [Train] step: 220399, loss: 0.106912, lr: 0.000200
562
+ 2023-08-07 20:40:20,868 - INFO - [Train] step: 220799, loss: 0.117222, lr: 0.000200
563
+ 2023-08-07 20:41:42,586 - INFO - [Train] step: 221199, loss: 0.135882, lr: 0.000200
564
+ 2023-08-07 20:43:04,673 - INFO - [Train] step: 221599, loss: 0.085260, lr: 0.000200
565
+ 2023-08-07 20:44:26,840 - INFO - [Train] step: 221999, loss: 0.076068, lr: 0.000200
566
+ 2023-08-07 20:45:48,968 - INFO - [Train] step: 222399, loss: 0.104925, lr: 0.000200
567
+ 2023-08-07 20:47:10,725 - INFO - [Train] step: 222799, loss: 0.104246, lr: 0.000200
568
+ 2023-08-07 20:48:32,799 - INFO - [Train] step: 223199, loss: 0.114426, lr: 0.000200
569
+ 2023-08-07 20:49:54,694 - INFO - [Train] step: 223599, loss: 0.103470, lr: 0.000200
570
+ 2023-08-07 20:51:16,360 - INFO - [Train] step: 223999, loss: 0.105722, lr: 0.000200
571
+ 2023-08-07 20:52:38,560 - INFO - [Train] step: 224399, loss: 0.134950, lr: 0.000200
572
+ 2023-08-07 20:54:00,689 - INFO - [Train] step: 224799, loss: 0.088887, lr: 0.000200
573
+ 2023-08-07 20:55:45,135 - INFO - [Train] step: 225199, loss: 0.098573, lr: 0.000200
574
+ 2023-08-07 20:57:07,274 - INFO - [Train] step: 225599, loss: 0.118965, lr: 0.000200
575
+ 2023-08-07 20:58:29,578 - INFO - [Train] step: 225999, loss: 0.137455, lr: 0.000200
576
+ 2023-08-07 20:59:51,682 - INFO - [Train] step: 226399, loss: 0.103509, lr: 0.000200
577
+ 2023-08-07 21:01:13,438 - INFO - [Train] step: 226799, loss: 0.131616, lr: 0.000200
578
+ 2023-08-07 21:02:35,784 - INFO - [Train] step: 227199, loss: 0.089393, lr: 0.000200
579
+ 2023-08-07 21:03:57,704 - INFO - [Train] step: 227599, loss: 0.100664, lr: 0.000200
580
+ 2023-08-07 21:05:20,008 - INFO - [Train] step: 227999, loss: 0.097019, lr: 0.000200
581
+ 2023-08-07 21:06:42,128 - INFO - [Train] step: 228399, loss: 0.125824, lr: 0.000200
582
+ 2023-08-07 21:08:04,245 - INFO - [Train] step: 228799, loss: 0.074508, lr: 0.000200
583
+ 2023-08-07 21:09:26,273 - INFO - [Train] step: 229199, loss: 0.128657, lr: 0.000200
584
+ 2023-08-07 21:10:48,094 - INFO - [Train] step: 229599, loss: 0.119606, lr: 0.000200
585
+ 2023-08-07 21:12:09,832 - INFO - [Train] step: 229999, loss: 0.106112, lr: 0.000200
586
+ 2023-08-07 21:13:54,265 - INFO - [Train] step: 230399, loss: 0.120219, lr: 0.000200
587
+ 2023-08-07 21:15:16,394 - INFO - [Train] step: 230799, loss: 0.130889, lr: 0.000200
588
+ 2023-08-07 21:16:38,117 - INFO - [Train] step: 231199, loss: 0.148711, lr: 0.000200
589
+ 2023-08-07 21:18:00,241 - INFO - [Train] step: 231599, loss: 0.104815, lr: 0.000200
590
+ 2023-08-07 21:19:22,542 - INFO - [Train] step: 231999, loss: 0.089949, lr: 0.000200
591
+ 2023-08-07 21:20:44,895 - INFO - [Train] step: 232399, loss: 0.125149, lr: 0.000200
592
+ 2023-08-07 21:22:06,860 - INFO - [Train] step: 232799, loss: 0.092998, lr: 0.000200
593
+ 2023-08-07 21:23:29,126 - INFO - [Train] step: 233199, loss: 0.140286, lr: 0.000200
594
+ 2023-08-07 21:24:51,483 - INFO - [Train] step: 233599, loss: 0.109981, lr: 0.000200
595
+ 2023-08-07 21:26:13,503 - INFO - [Train] step: 233999, loss: 0.098034, lr: 0.000200
596
+ 2023-08-07 21:27:35,301 - INFO - [Train] step: 234399, loss: 0.092793, lr: 0.000200
597
+ 2023-08-07 21:28:57,647 - INFO - [Train] step: 234799, loss: 0.141544, lr: 0.000200
598
+ 2023-08-07 21:30:43,114 - INFO - [Train] step: 235199, loss: 0.090412, lr: 0.000200
599
+ 2023-08-07 21:32:04,815 - INFO - [Train] step: 235599, loss: 0.123882, lr: 0.000200
600
+ 2023-08-07 21:33:26,759 - INFO - [Train] step: 235999, loss: 0.127474, lr: 0.000200
601
+ 2023-08-07 21:34:48,698 - INFO - [Train] step: 236399, loss: 0.125369, lr: 0.000200
602
+ 2023-08-07 21:36:10,611 - INFO - [Train] step: 236799, loss: 0.116213, lr: 0.000200
603
+ 2023-08-07 21:37:32,274 - INFO - [Train] step: 237199, loss: 0.106958, lr: 0.000200
604
+ 2023-08-07 21:38:54,239 - INFO - [Train] step: 237599, loss: 0.095185, lr: 0.000200
605
+ 2023-08-07 21:40:16,363 - INFO - [Train] step: 237999, loss: 0.101106, lr: 0.000200
606
+ 2023-08-07 21:41:38,677 - INFO - [Train] step: 238399, loss: 0.123637, lr: 0.000200
607
+ 2023-08-07 21:43:00,614 - INFO - [Train] step: 238799, loss: 0.110368, lr: 0.000200
608
+ 2023-08-07 21:44:22,741 - INFO - [Train] step: 239199, loss: 0.131181, lr: 0.000200
609
+ 2023-08-07 21:45:45,052 - INFO - [Train] step: 239599, loss: 0.084828, lr: 0.000200
610
+ 2023-08-07 21:47:07,369 - INFO - [Train] step: 239999, loss: 0.119953, lr: 0.000200
611
+ 2023-08-07 21:48:50,640 - INFO - [Train] step: 240399, loss: 0.139034, lr: 0.000200
612
+ 2023-08-07 21:50:12,525 - INFO - [Train] step: 240799, loss: 0.103394, lr: 0.000200
613
+ 2023-08-07 21:51:34,447 - INFO - [Train] step: 241199, loss: 0.124396, lr: 0.000200
614
+ 2023-08-07 21:52:56,596 - INFO - [Train] step: 241599, loss: 0.096993, lr: 0.000200
615
+ 2023-08-07 21:54:18,702 - INFO - [Train] step: 241999, loss: 0.116652, lr: 0.000200
616
+ 2023-08-07 21:55:40,813 - INFO - [Train] step: 242399, loss: 0.117313, lr: 0.000200
617
+ 2023-08-07 21:57:03,137 - INFO - [Train] step: 242799, loss: 0.095886, lr: 0.000200
618
+ 2023-08-07 21:58:25,304 - INFO - [Train] step: 243199, loss: 0.120149, lr: 0.000200
619
+ 2023-08-07 21:59:47,213 - INFO - [Train] step: 243599, loss: 0.125677, lr: 0.000200
620
+ 2023-08-07 22:01:09,501 - INFO - [Train] step: 243999, loss: 0.073386, lr: 0.000200
621
+ 2023-08-07 22:02:31,663 - INFO - [Train] step: 244399, loss: 0.084520, lr: 0.000200
622
+ 2023-08-07 22:03:53,803 - INFO - [Train] step: 244799, loss: 0.068746, lr: 0.000200
623
+ 2023-08-07 22:05:37,267 - INFO - [Train] step: 245199, loss: 0.096776, lr: 0.000200
624
+ 2023-08-07 22:06:59,584 - INFO - [Train] step: 245599, loss: 0.123576, lr: 0.000200
625
+ 2023-08-07 22:08:22,091 - INFO - [Train] step: 245999, loss: 0.060448, lr: 0.000200
626
+ 2023-08-07 22:09:43,788 - INFO - [Train] step: 246399, loss: 0.077442, lr: 0.000200
627
+ 2023-08-07 22:11:06,110 - INFO - [Train] step: 246799, loss: 0.117696, lr: 0.000200
628
+ 2023-08-07 22:12:28,439 - INFO - [Train] step: 247199, loss: 0.122948, lr: 0.000200
629
+ 2023-08-07 22:13:50,600 - INFO - [Train] step: 247599, loss: 0.141044, lr: 0.000200
630
+ 2023-08-07 22:15:22,620 - INFO - [Train] step: 247999, loss: 0.103831, lr: 0.000200
631
+ 2023-08-07 22:16:55,858 - INFO - [Train] step: 248399, loss: 0.099113, lr: 0.000200
632
+ 2023-08-07 22:18:25,039 - INFO - [Train] step: 248799, loss: 0.116123, lr: 0.000200
633
+ 2023-08-07 22:19:52,921 - INFO - [Train] step: 249199, loss: 0.101775, lr: 0.000200
634
+ 2023-08-07 22:21:17,177 - INFO - [Train] step: 249599, loss: 0.119279, lr: 0.000200
635
+ 2023-08-07 22:22:39,788 - INFO - [Train] step: 249999, loss: 0.091824, lr: 0.000200
636
+ 2023-08-07 22:24:23,842 - INFO - [Train] step: 250399, loss: 0.105654, lr: 0.000200
637
+ 2023-08-07 22:25:46,051 - INFO - [Train] step: 250799, loss: 0.128992, lr: 0.000200
638
+ 2023-08-07 22:27:07,268 - INFO - [Train] step: 251199, loss: 0.141036, lr: 0.000200
639
+ 2023-08-07 22:28:29,590 - INFO - [Train] step: 251599, loss: 0.170446, lr: 0.000200
640
+ 2023-08-07 22:29:51,909 - INFO - [Train] step: 251999, loss: 0.108987, lr: 0.000200
641
+ 2023-08-07 22:31:14,538 - INFO - [Train] step: 252399, loss: 0.100798, lr: 0.000200
642
+ 2023-08-07 22:32:45,922 - INFO - [Train] step: 252799, loss: 0.118228, lr: 0.000200
643
+ 2023-08-07 22:34:19,412 - INFO - [Train] step: 253199, loss: 0.122678, lr: 0.000200
644
+ 2023-08-07 22:35:49,710 - INFO - [Train] step: 253599, loss: 0.101707, lr: 0.000200
645
+ 2023-08-07 22:37:16,538 - INFO - [Train] step: 253999, loss: 0.096021, lr: 0.000200
646
+ 2023-08-07 22:38:51,205 - INFO - [Train] step: 254399, loss: 0.070691, lr: 0.000200
647
+ 2023-08-07 22:40:25,431 - INFO - [Train] step: 254799, loss: 0.098387, lr: 0.000200
648
+ 2023-08-07 22:42:12,485 - INFO - [Train] step: 255199, loss: 0.111114, lr: 0.000200
649
+ 2023-08-07 22:43:35,101 - INFO - [Train] step: 255599, loss: 0.087545, lr: 0.000200
650
+ 2023-08-07 22:44:58,170 - INFO - [Train] step: 255999, loss: 0.092697, lr: 0.000200
651
+ 2023-08-07 22:46:21,719 - INFO - [Train] step: 256399, loss: 0.096473, lr: 0.000200
652
+ 2023-08-07 22:47:43,659 - INFO - [Train] step: 256799, loss: 0.122011, lr: 0.000200
653
+ 2023-08-07 22:49:05,756 - INFO - [Train] step: 257199, loss: 0.119900, lr: 0.000200
654
+ 2023-08-07 22:50:27,853 - INFO - [Train] step: 257599, loss: 0.128482, lr: 0.000200
655
+ 2023-08-07 22:51:50,018 - INFO - [Train] step: 257999, loss: 0.116662, lr: 0.000200
656
+ 2023-08-07 22:53:12,353 - INFO - [Train] step: 258399, loss: 0.108821, lr: 0.000200
657
+ 2023-08-07 22:54:34,708 - INFO - [Train] step: 258799, loss: 0.080094, lr: 0.000200
658
+ 2023-08-07 22:55:57,039 - INFO - [Train] step: 259199, loss: 0.106504, lr: 0.000200
659
+ 2023-08-07 22:57:19,150 - INFO - [Train] step: 259599, loss: 0.115011, lr: 0.000200
660
+ 2023-08-07 22:58:41,030 - INFO - [Train] step: 259999, loss: 0.111058, lr: 0.000200
661
+ 2023-08-07 23:00:25,532 - INFO - [Train] step: 260399, loss: 0.113337, lr: 0.000200
662
+ 2023-08-07 23:01:48,051 - INFO - [Train] step: 260799, loss: 0.103375, lr: 0.000200
663
+ 2023-08-07 23:03:10,082 - INFO - [Train] step: 261199, loss: 0.112022, lr: 0.000200
664
+ 2023-08-07 23:04:33,845 - INFO - [Train] step: 261599, loss: 0.107698, lr: 0.000200
665
+ 2023-08-07 23:05:55,644 - INFO - [Train] step: 261999, loss: 0.111864, lr: 0.000200
666
+ 2023-08-07 23:07:29,636 - INFO - [Train] step: 262399, loss: 0.104146, lr: 0.000200
667
+ 2023-08-07 23:09:04,637 - INFO - [Train] step: 262799, loss: 0.121387, lr: 0.000200
668
+ 2023-08-07 23:10:39,766 - INFO - [Train] step: 263199, loss: 0.089295, lr: 0.000200
669
+ 2023-08-07 23:12:04,688 - INFO - [Train] step: 263599, loss: 0.125900, lr: 0.000200
670
+ 2023-08-07 23:13:37,358 - INFO - [Train] step: 263999, loss: 0.119247, lr: 0.000200
671
+ 2023-08-07 23:15:11,067 - INFO - [Train] step: 264399, loss: 0.137994, lr: 0.000200
672
+ 2023-08-07 23:16:41,613 - INFO - [Train] step: 264799, loss: 0.123085, lr: 0.000200
673
+ 2023-08-07 23:18:32,807 - INFO - [Train] step: 265199, loss: 0.109656, lr: 0.000200
674
+ 2023-08-07 23:20:07,528 - INFO - [Train] step: 265599, loss: 0.090404, lr: 0.000200
675
+ 2023-08-07 23:21:40,199 - INFO - [Train] step: 265999, loss: 0.120225, lr: 0.000200
676
+ 2023-08-07 23:23:04,960 - INFO - [Train] step: 266399, loss: 0.092175, lr: 0.000200
677
+ 2023-08-07 23:24:39,040 - INFO - [Train] step: 266799, loss: 0.120426, lr: 0.000200
678
+ 2023-08-07 23:26:13,534 - INFO - [Train] step: 267199, loss: 0.137474, lr: 0.000200
679
+ 2023-08-07 23:27:39,611 - INFO - [Train] step: 267599, loss: 0.103276, lr: 0.000200
680
+ 2023-08-07 23:29:11,514 - INFO - [Train] step: 267999, loss: 0.118592, lr: 0.000200
681
+ 2023-08-07 23:30:45,851 - INFO - [Train] step: 268399, loss: 0.128617, lr: 0.000200
682
+ 2023-08-07 23:32:16,690 - INFO - [Train] step: 268799, loss: 0.132897, lr: 0.000200
683
+ 2023-08-07 23:33:43,079 - INFO - [Train] step: 269199, loss: 0.110867, lr: 0.000200
684
+ 2023-08-07 23:35:17,111 - INFO - [Train] step: 269599, loss: 0.133101, lr: 0.000200
685
+ 2023-08-07 23:36:51,197 - INFO - [Train] step: 269999, loss: 0.093577, lr: 0.000200
686
+ 2023-08-07 23:38:38,623 - INFO - [Train] step: 270399, loss: 0.160773, lr: 0.000200
687
+ 2023-08-07 23:40:13,996 - INFO - [Train] step: 270799, loss: 0.104640, lr: 0.000200
688
+ 2023-08-07 23:41:49,296 - INFO - [Train] step: 271199, loss: 0.117079, lr: 0.000200
689
+ 2023-08-07 23:43:22,170 - INFO - [Train] step: 271599, loss: 0.113787, lr: 0.000200
690
+ 2023-08-07 23:44:54,487 - INFO - [Train] step: 271999, loss: 0.124887, lr: 0.000200
691
+ 2023-08-07 23:46:29,620 - INFO - [Train] step: 272399, loss: 0.098357, lr: 0.000200
692
+ 2023-08-07 23:48:04,776 - INFO - [Train] step: 272799, loss: 0.099946, lr: 0.000200
693
+ 2023-08-07 23:49:32,913 - INFO - [Train] step: 273199, loss: 0.129804, lr: 0.000200
694
+ 2023-08-07 23:51:08,185 - INFO - [Train] step: 273599, loss: 0.114058, lr: 0.000200
695
+ 2023-08-07 23:52:44,069 - INFO - [Train] step: 273999, loss: 0.101482, lr: 0.000200
696
+ 2023-08-07 23:54:11,869 - INFO - [Train] step: 274399, loss: 0.103703, lr: 0.000200
697
+ 2023-08-07 23:55:47,246 - INFO - [Train] step: 274799, loss: 0.095251, lr: 0.000200
698
+ 2023-08-07 23:57:44,993 - INFO - [Train] step: 275199, loss: 0.112879, lr: 0.000200
699
+ 2023-08-07 23:59:16,346 - INFO - [Train] step: 275599, loss: 0.100263, lr: 0.000200
700
+ 2023-08-08 00:00:51,678 - INFO - [Train] step: 275999, loss: 0.090636, lr: 0.000200
701
+ 2023-08-08 00:02:27,485 - INFO - [Train] step: 276399, loss: 0.094517, lr: 0.000200
702
+ 2023-08-08 00:04:02,799 - INFO - [Train] step: 276799, loss: 0.140384, lr: 0.000200
703
+ 2023-08-08 00:05:31,864 - INFO - [Train] step: 277199, loss: 0.084483, lr: 0.000200
704
+ 2023-08-08 00:07:07,584 - INFO - [Train] step: 277599, loss: 0.104991, lr: 0.000200
705
+ 2023-08-08 00:08:43,490 - INFO - [Train] step: 277999, loss: 0.080642, lr: 0.000200
706
+ 2023-08-08 00:10:14,199 - INFO - [Train] step: 278399, loss: 0.099485, lr: 0.000200
707
+ 2023-08-08 00:11:44,941 - INFO - [Train] step: 278799, loss: 0.071400, lr: 0.000200
708
+ 2023-08-08 00:13:20,981 - INFO - [Train] step: 279199, loss: 0.108450, lr: 0.000200
709
+ 2023-08-08 00:14:56,311 - INFO - [Train] step: 279599, loss: 0.116792, lr: 0.000200
710
+ 2023-08-08 00:16:26,637 - INFO - [Train] step: 279999, loss: 0.118108, lr: 0.000200
711
+ 2023-08-08 00:18:26,057 - INFO - [Train] step: 280399, loss: 0.087363, lr: 0.000200
712
+ 2023-08-08 00:20:00,966 - INFO - [Train] step: 280799, loss: 0.163901, lr: 0.000200
713
+ 2023-08-08 00:21:34,019 - INFO - [Train] step: 281199, loss: 0.113656, lr: 0.000200
714
+ 2023-08-08 00:23:09,681 - INFO - [Train] step: 281599, loss: 0.111682, lr: 0.000200
715
+ 2023-08-08 00:24:44,518 - INFO - [Train] step: 281999, loss: 0.082123, lr: 0.000200
716
+ 2023-08-08 00:26:20,299 - INFO - [Train] step: 282399, loss: 0.094872, lr: 0.000200
717
+ 2023-08-08 00:27:52,812 - INFO - [Train] step: 282799, loss: 0.118291, lr: 0.000200
718
+ 2023-08-08 00:29:29,009 - INFO - [Train] step: 283199, loss: 0.083886, lr: 0.000200
719
+ 2023-08-08 00:31:05,215 - INFO - [Train] step: 283599, loss: 0.099051, lr: 0.000200
720
+ 2023-08-08 00:32:33,246 - INFO - [Train] step: 283999, loss: 0.118002, lr: 0.000200
721
+ 2023-08-08 00:34:09,361 - INFO - [Train] step: 284399, loss: 0.132477, lr: 0.000200
722
+ 2023-08-08 00:35:44,289 - INFO - [Train] step: 284799, loss: 0.124727, lr: 0.000200
723
+ 2023-08-08 00:37:40,821 - INFO - [Train] step: 285199, loss: 0.114215, lr: 0.000200
724
+ 2023-08-08 00:39:15,164 - INFO - [Train] step: 285599, loss: 0.119803, lr: 0.000200
725
+ 2023-08-08 00:40:50,470 - INFO - [Train] step: 285999, loss: 0.139484, lr: 0.000200
726
+ 2023-08-08 00:42:26,091 - INFO - [Train] step: 286399, loss: 0.129269, lr: 0.000200
727
+ 2023-08-08 00:43:56,441 - INFO - [Train] step: 286799, loss: 0.125098, lr: 0.000200
728
+ 2023-08-08 00:45:31,844 - INFO - [Train] step: 287199, loss: 0.106770, lr: 0.000200
729
+ 2023-08-08 00:47:07,967 - INFO - [Train] step: 287599, loss: 0.121613, lr: 0.000200
730
+ 2023-08-08 00:48:33,806 - INFO - [Train] step: 287999, loss: 0.120980, lr: 0.000200
731
+ 2023-08-08 00:50:09,070 - INFO - [Train] step: 288399, loss: 0.134276, lr: 0.000200
732
+ 2023-08-08 00:51:45,046 - INFO - [Train] step: 288799, loss: 0.082459, lr: 0.000200
733
+ 2023-08-08 00:53:14,335 - INFO - [Train] step: 289199, loss: 0.091214, lr: 0.000200
734
+ 2023-08-08 00:54:49,562 - INFO - [Train] step: 289599, loss: 0.106117, lr: 0.000200
735
+ 2023-08-08 00:56:25,336 - INFO - [Train] step: 289999, loss: 0.102086, lr: 0.000200
736
+ 2023-08-08 00:58:20,444 - INFO - [Train] step: 290399, loss: 0.115763, lr: 0.000200
737
+ 2023-08-08 00:59:54,729 - INFO - [Train] step: 290799, loss: 0.074310, lr: 0.000200
738
+ 2023-08-08 01:01:30,845 - INFO - [Train] step: 291199, loss: 0.070340, lr: 0.000200
739
+ 2023-08-08 01:03:06,663 - INFO - [Train] step: 291599, loss: 0.105800, lr: 0.000200
740
+ 2023-08-08 01:04:39,021 - INFO - [Train] step: 291999, loss: 0.094702, lr: 0.000200
741
+ 2023-08-08 01:06:14,457 - INFO - [Train] step: 292399, loss: 0.106377, lr: 0.000200
742
+ 2023-08-08 01:07:50,208 - INFO - [Train] step: 292799, loss: 0.115168, lr: 0.000200
743
+ 2023-08-08 01:09:23,197 - INFO - [Train] step: 293199, loss: 0.137461, lr: 0.000200
744
+ 2023-08-08 01:10:55,565 - INFO - [Train] step: 293599, loss: 0.140380, lr: 0.000200
745
+ 2023-08-08 01:12:31,067 - INFO - [Train] step: 293999, loss: 0.099768, lr: 0.000200
746
+ 2023-08-08 01:14:05,136 - INFO - [Train] step: 294399, loss: 0.110645, lr: 0.000200
747
+ 2023-08-08 01:15:32,927 - INFO - [Train] step: 294799, loss: 0.119441, lr: 0.000200
748
+ 2023-08-08 01:17:28,985 - INFO - [Train] step: 295199, loss: 0.109740, lr: 0.000200
749
+ 2023-08-08 01:19:05,154 - INFO - [Train] step: 295599, loss: 0.109121, lr: 0.000200
750
+ 2023-08-08 01:20:34,042 - INFO - [Train] step: 295999, loss: 0.153376, lr: 0.000200
751
+ 2023-08-08 01:22:09,280 - INFO - [Train] step: 296399, loss: 0.148037, lr: 0.000200
752
+ 2023-08-08 01:23:45,818 - INFO - [Train] step: 296799, loss: 0.101609, lr: 0.000200
753
+ 2023-08-08 01:25:16,438 - INFO - [Train] step: 297199, loss: 0.083564, lr: 0.000200
754
+ 2023-08-08 01:26:49,770 - INFO - [Train] step: 297599, loss: 0.089507, lr: 0.000200
755
+ 2023-08-08 01:28:25,930 - INFO - [Train] step: 297999, loss: 0.103438, lr: 0.000200
756
+ 2023-08-08 01:30:01,904 - INFO - [Train] step: 298399, loss: 0.120304, lr: 0.000200
757
+ 2023-08-08 01:31:31,271 - INFO - [Train] step: 298799, loss: 0.101671, lr: 0.000200
758
+ 2023-08-08 01:33:06,951 - INFO - [Train] step: 299199, loss: 0.090928, lr: 0.000200
759
+ 2023-08-08 01:34:42,893 - INFO - [Train] step: 299599, loss: 0.096609, lr: 0.000200
760
+ 2023-08-08 01:36:13,167 - INFO - [Train] step: 299999, loss: 0.105133, lr: 0.000200
761
+ 2023-08-08 01:38:10,193 - INFO - [Train] step: 300399, loss: 0.153599, lr: 0.000200
762
+ 2023-08-08 01:39:45,934 - INFO - [Train] step: 300799, loss: 0.099348, lr: 0.000200
763
+ 2023-08-08 01:41:18,613 - INFO - [Train] step: 301199, loss: 0.131756, lr: 0.000200
764
+ 2023-08-08 01:42:52,537 - INFO - [Train] step: 301599, loss: 0.103970, lr: 0.000200
765
+ 2023-08-08 01:44:28,877 - INFO - [Train] step: 301999, loss: 0.101250, lr: 0.000200
766
+ 2023-08-08 01:46:04,245 - INFO - [Train] step: 302399, loss: 0.118983, lr: 0.000200
767
+ 2023-08-08 01:47:37,308 - INFO - [Train] step: 302799, loss: 0.105215, lr: 0.000200
768
+ 2023-08-08 01:49:12,548 - INFO - [Train] step: 303199, loss: 0.086182, lr: 0.000200
769
+ 2023-08-08 01:50:47,913 - INFO - [Train] step: 303599, loss: 0.119117, lr: 0.000200
770
+ 2023-08-08 01:52:22,498 - INFO - [Train] step: 303999, loss: 0.104346, lr: 0.000200
771
+ 2023-08-08 01:53:54,134 - INFO - [Train] step: 304399, loss: 0.112351, lr: 0.000200
772
+ 2023-08-08 01:55:29,212 - INFO - [Train] step: 304799, loss: 0.100717, lr: 0.000200
773
+ 2023-08-08 01:57:25,982 - INFO - [Train] step: 305199, loss: 0.102997, lr: 0.000200
774
+ 2023-08-08 01:58:52,766 - INFO - [Train] step: 305599, loss: 0.091558, lr: 0.000200
775
+ 2023-08-08 02:00:27,824 - INFO - [Train] step: 305999, loss: 0.124631, lr: 0.000200
776
+ 2023-08-08 02:02:04,243 - INFO - [Train] step: 306399, loss: 0.100726, lr: 0.000200
777
+ 2023-08-08 02:03:32,469 - INFO - [Train] step: 306799, loss: 0.119213, lr: 0.000200
778
+ 2023-08-08 02:05:07,745 - INFO - [Train] step: 307199, loss: 0.090630, lr: 0.000200
779
+ 2023-08-08 02:06:44,002 - INFO - [Train] step: 307599, loss: 0.120802, lr: 0.000200
780
+ 2023-08-08 02:08:16,554 - INFO - [Train] step: 307999, loss: 0.124660, lr: 0.000200
781
+ 2023-08-08 02:09:50,417 - INFO - [Train] step: 308399, loss: 0.105882, lr: 0.000200
782
+ 2023-08-08 02:11:26,544 - INFO - [Train] step: 308799, loss: 0.130760, lr: 0.000200
783
+ 2023-08-08 02:13:02,429 - INFO - [Train] step: 309199, loss: 0.117145, lr: 0.000200
784
+ 2023-08-08 02:14:33,895 - INFO - [Train] step: 309599, loss: 0.101169, lr: 0.000200
785
+ 2023-08-08 02:16:09,586 - INFO - [Train] step: 309999, loss: 0.099388, lr: 0.000200
786
+ 2023-08-08 02:18:08,655 - INFO - [Train] step: 310399, loss: 0.126765, lr: 0.000200
787
+ 2023-08-08 02:19:39,624 - INFO - [Train] step: 310799, loss: 0.111021, lr: 0.000200
788
+ 2023-08-08 02:21:11,973 - INFO - [Train] step: 311199, loss: 0.088242, lr: 0.000200
789
+ 2023-08-08 02:22:48,098 - INFO - [Train] step: 311599, loss: 0.123908, lr: 0.000200
790
+ 2023-08-08 02:24:21,602 - INFO - [Train] step: 311999, loss: 0.102665, lr: 0.000200
791
+ 2023-08-08 02:25:52,741 - INFO - [Train] step: 312399, loss: 0.102778, lr: 0.000200
792
+ 2023-08-08 02:27:28,798 - INFO - [Train] step: 312799, loss: 0.149749, lr: 0.000200
793
+ 2023-08-08 02:29:04,047 - INFO - [Train] step: 313199, loss: 0.132170, lr: 0.000200
794
+ 2023-08-08 02:30:35,359 - INFO - [Train] step: 313599, loss: 0.115316, lr: 0.000200
795
+ 2023-08-08 02:32:11,247 - INFO - [Train] step: 313999, loss: 0.149668, lr: 0.000200
796
+ 2023-08-08 02:33:47,158 - INFO - [Train] step: 314399, loss: 0.194232, lr: 0.000200
797
+ 2023-08-08 02:35:22,470 - INFO - [Train] step: 314799, loss: 0.107666, lr: 0.000200
798
+ 2023-08-08 02:37:18,296 - INFO - [Train] step: 315199, loss: 0.127770, lr: 0.000200
799
+ 2023-08-08 02:38:53,857 - INFO - [Train] step: 315599, loss: 0.111345, lr: 0.000200
800
+ 2023-08-08 02:40:29,955 - INFO - [Train] step: 315999, loss: 0.101629, lr: 0.000200
801
+ 2023-08-08 02:42:00,133 - INFO - [Train] step: 316399, loss: 0.113506, lr: 0.000200
802
+ 2023-08-08 02:43:35,836 - INFO - [Train] step: 316799, loss: 0.123914, lr: 0.000200
803
+ 2023-08-08 02:45:11,927 - INFO - [Train] step: 317199, loss: 0.099489, lr: 0.000200
804
+ 2023-08-08 02:46:39,574 - INFO - [Train] step: 317599, loss: 0.145891, lr: 0.000200
805
+ 2023-08-08 02:48:14,756 - INFO - [Train] step: 317999, loss: 0.104523, lr: 0.000200
806
+ 2023-08-08 02:49:50,460 - INFO - [Train] step: 318399, loss: 0.085938, lr: 0.000200
807
+ 2023-08-08 02:51:20,467 - INFO - [Train] step: 318799, loss: 0.133758, lr: 0.000200
808
+ 2023-08-08 02:52:53,229 - INFO - [Train] step: 319199, loss: 0.104609, lr: 0.000200
809
+ 2023-08-08 02:54:29,308 - INFO - [Train] step: 319599, loss: 0.105176, lr: 0.000200
810
+ 2023-08-08 02:56:04,610 - INFO - [Train] step: 319999, loss: 0.126959, lr: 0.000200
811
+ 2023-08-08 02:58:02,211 - INFO - [Train] step: 320399, loss: 0.145715, lr: 0.000200
812
+ 2023-08-08 02:59:38,056 - INFO - [Train] step: 320799, loss: 0.091949, lr: 0.000200
813
+ 2023-08-08 03:01:13,267 - INFO - [Train] step: 321199, loss: 0.100993, lr: 0.000200
814
+ 2023-08-08 03:02:46,036 - INFO - [Train] step: 321599, loss: 0.137251, lr: 0.000200
815
+ 2023-08-08 03:04:19,839 - INFO - [Train] step: 321999, loss: 0.129178, lr: 0.000200
816
+ 2023-08-08 03:05:55,137 - INFO - [Train] step: 322399, loss: 0.153326, lr: 0.000200
817
+ 2023-08-08 03:07:30,612 - INFO - [Train] step: 322799, loss: 0.091629, lr: 0.000200
818
+ 2023-08-08 03:09:01,173 - INFO - [Train] step: 323199, loss: 0.108642, lr: 0.000200
819
+ 2023-08-08 03:10:36,595 - INFO - [Train] step: 323599, loss: 0.114612, lr: 0.000200
820
+ 2023-08-08 03:12:12,520 - INFO - [Train] step: 323999, loss: 0.121940, lr: 0.000200
821
+ 2023-08-08 03:13:40,234 - INFO - [Train] step: 324399, loss: 0.092551, lr: 0.000200
822
+ 2023-08-08 03:15:15,922 - INFO - [Train] step: 324799, loss: 0.127496, lr: 0.000200
823
+ 2023-08-08 03:17:14,156 - INFO - [Train] step: 325199, loss: 0.104814, lr: 0.000200
824
+ 2023-08-08 03:18:43,526 - INFO - [Train] step: 325599, loss: 0.101345, lr: 0.000200
825
+ 2023-08-08 03:20:18,290 - INFO - [Train] step: 325999, loss: 0.105136, lr: 0.000200
826
+ 2023-08-08 03:21:54,474 - INFO - [Train] step: 326399, loss: 0.152865, lr: 0.000200
827
+ 2023-08-08 03:23:30,489 - INFO - [Train] step: 326799, loss: 0.107397, lr: 0.000200
828
+ 2023-08-08 03:25:02,374 - INFO - [Train] step: 327199, loss: 0.129319, lr: 0.000200
829
+ 2023-08-08 03:26:38,370 - INFO - [Train] step: 327599, loss: 0.118777, lr: 0.000200
830
+ 2023-08-08 03:28:14,374 - INFO - [Train] step: 327999, loss: 0.134705, lr: 0.000200
831
+ 2023-08-08 03:29:46,274 - INFO - [Train] step: 328399, loss: 0.106616, lr: 0.000200
832
+ 2023-08-08 03:31:19,521 - INFO - [Train] step: 328799, loss: 0.093173, lr: 0.000200
833
+ 2023-08-08 03:32:55,013 - INFO - [Train] step: 329199, loss: 0.070722, lr: 0.000200
834
+ 2023-08-08 03:34:29,704 - INFO - [Train] step: 329599, loss: 0.104294, lr: 0.000200
835
+ 2023-08-08 03:35:58,359 - INFO - [Train] step: 329999, loss: 0.113696, lr: 0.000200
836
+ 2023-08-08 03:37:56,870 - INFO - [Train] step: 330399, loss: 0.107551, lr: 0.000200
837
+ 2023-08-08 03:39:32,632 - INFO - [Train] step: 330799, loss: 0.108850, lr: 0.000200
838
+ 2023-08-08 03:41:01,181 - INFO - [Train] step: 331199, loss: 0.107326, lr: 0.000200
839
+ 2023-08-08 03:42:36,356 - INFO - [Train] step: 331599, loss: 0.126462, lr: 0.000200
840
+ 2023-08-08 03:44:12,217 - INFO - [Train] step: 331999, loss: 0.096204, lr: 0.000200
841
+ 2023-08-08 03:45:43,514 - INFO - [Train] step: 332399, loss: 0.120802, lr: 0.000200
842
+ 2023-08-08 03:47:18,949 - INFO - [Train] step: 332799, loss: 0.110875, lr: 0.000200
843
+ 2023-08-08 03:48:54,829 - INFO - [Train] step: 333199, loss: 0.106790, lr: 0.000200
844
+ 2023-08-08 03:50:29,275 - INFO - [Train] step: 333599, loss: 0.113384, lr: 0.000200
845
+ 2023-08-08 03:52:02,707 - INFO - [Train] step: 333999, loss: 0.109059, lr: 0.000200
846
+ 2023-08-08 03:53:38,815 - INFO - [Train] step: 334399, loss: 0.132590, lr: 0.000200
847
+ 2023-08-08 03:55:15,013 - INFO - [Train] step: 334799, loss: 0.106684, lr: 0.000200
ddpm_pred_x0_cifar10/output-2023-08-09-23-34-33.log ADDED
The diff for this file is too large to render. See raw diff
 
ddpm_pred_x0_cifar10/samples.zip ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:837da269f961683da6c52f5764b004fb327017fdcd659015188d73cf9916ec62
3
+ size 22176790
ddpm_pred_x0_cifar10/tensorboard/events.out.tfevents.1691408944.aa-ESC4000-G4.24711.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a50c8670058c4d5de173835ee0b65f1edcba92c9f6fe4289f740a027d5a44276
3
+ size 32800768
ddpm_pred_x0_cifar10/tensorboard/events.out.tfevents.1691638477.aa-ESC4000-G4.12367.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:989546dece66e89f2f1c8dd167e876af39d79ab34309a77a419829ad8ca53eb6
3
+ size 47040088