Upload ddpm_pred_x0_cifar10 checkpoints and training logs
Browse files- ddpm_pred_x0_cifar10/ckpt/step079999/ema.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step079999/meta.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step079999/model.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step079999/optimizer.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step159999/ema.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step159999/meta.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step159999/model.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step159999/optimizer.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step239999/ema.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step239999/meta.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step239999/model.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step239999/optimizer.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step319999/ema.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step319999/meta.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step319999/model.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step319999/optimizer.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step399999/ema.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step399999/meta.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step399999/model.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step399999/optimizer.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step479999/ema.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step479999/meta.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step479999/model.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step479999/optimizer.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step559999/ema.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step559999/meta.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step559999/model.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step559999/optimizer.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step639999/ema.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step639999/meta.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step639999/model.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step639999/optimizer.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step719999/ema.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step719999/meta.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step719999/model.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step719999/optimizer.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step799999/ema.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step799999/meta.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step799999/model.pt +3 -0
- ddpm_pred_x0_cifar10/ckpt/step799999/optimizer.pt +3 -0
- ddpm_pred_x0_cifar10/config-2023-08-07-07-49-01.yaml +54 -0
- ddpm_pred_x0_cifar10/config-2023-08-09-23-34-33.yaml +54 -0
- ddpm_pred_x0_cifar10/output-2023-08-07-07-49-01.log +847 -0
- ddpm_pred_x0_cifar10/output-2023-08-09-23-34-33.log +0 -0
- ddpm_pred_x0_cifar10/samples.zip +3 -0
- ddpm_pred_x0_cifar10/tensorboard/events.out.tfevents.1691408944.aa-ESC4000-G4.24711.0 +3 -0
- ddpm_pred_x0_cifar10/tensorboard/events.out.tfevents.1691638477.aa-ESC4000-G4.12367.0 +3 -0
ddpm_pred_x0_cifar10/ckpt/step079999/ema.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e666f49c500ce6fe4861467bee440518c18d87877644156d1929b3cd910ca071
|
3 |
+
size 143067273
|
ddpm_pred_x0_cifar10/ckpt/step079999/meta.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:460ecfe9af7c3c2749db8e2d57285af468303df85ec0957a7e666d8bf7bc5387
|
3 |
+
size 425
|
ddpm_pred_x0_cifar10/ckpt/step079999/model.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6e097b35bfae47bcba0ac9c90c9214779d6e22b00bdc3384066ce87297accb27
|
3 |
+
size 143084317
|
ddpm_pred_x0_cifar10/ckpt/step079999/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1eabd446b8ce194d9c702f4b3ee89024aef5c41b3643a91d50aad6db7521b304
|
3 |
+
size 286245381
|
ddpm_pred_x0_cifar10/ckpt/step159999/ema.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bfdef0e713407ba8ee5658c3ade1035d7591287357c1c63d63996d1838dde02a
|
3 |
+
size 143067273
|
ddpm_pred_x0_cifar10/ckpt/step159999/meta.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0c084ac1d346ecee3fb1e4d361399e0c2850bdba7e412be1f08db0ae69cd059c
|
3 |
+
size 425
|
ddpm_pred_x0_cifar10/ckpt/step159999/model.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d925d5eef0fde7149524c922f6c5b1e525a73512672f96d552a5729dab1c3f83
|
3 |
+
size 143084317
|
ddpm_pred_x0_cifar10/ckpt/step159999/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8462a49bfd582335fd535ff259208a4acb62a3e579cfc5c5a342f2ab57de082c
|
3 |
+
size 286245381
|
ddpm_pred_x0_cifar10/ckpt/step239999/ema.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0d31c3e953df4e8b1a6b959754e3c55b4d0d80f4ac4131cedac599ee05732fd2
|
3 |
+
size 143067273
|
ddpm_pred_x0_cifar10/ckpt/step239999/meta.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4df47f45c956c46e67d34dd50aa6b4413b30d5ca1c3db1a0c6534d2659d8afbf
|
3 |
+
size 425
|
ddpm_pred_x0_cifar10/ckpt/step239999/model.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c11cb487cdb3d63f70da5a94cba34858fb717c644ec28f190d5403ab9e74ad36
|
3 |
+
size 143084317
|
ddpm_pred_x0_cifar10/ckpt/step239999/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0f336dff86b18025f20ef321646b06a5b3b1f74f8f2630f92702ca164b7705ce
|
3 |
+
size 286245381
|
ddpm_pred_x0_cifar10/ckpt/step319999/ema.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3786b19277c131b381805135684027edc2c2958bd2de43a40aa114d5f70986e2
|
3 |
+
size 143067273
|
ddpm_pred_x0_cifar10/ckpt/step319999/meta.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0f5499fc4a910fdf81fe9bd42925dde65531a71584d5cba241e210753ed32a71
|
3 |
+
size 425
|
ddpm_pred_x0_cifar10/ckpt/step319999/model.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3fa82ef1e1960e0b6ed5942d9a31bd9b312ede685b1b1838b0e487be64990f98
|
3 |
+
size 143084317
|
ddpm_pred_x0_cifar10/ckpt/step319999/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b2f0bac98db2393c14d39e118ffba87cad1a4bb3b1f879404e95266865ebeb3a
|
3 |
+
size 286245381
|
ddpm_pred_x0_cifar10/ckpt/step399999/ema.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b3c4b7c0d40c02cc673509ec66ea15b5c03530c5f793a78d202fefd88eb10dab
|
3 |
+
size 143067273
|
ddpm_pred_x0_cifar10/ckpt/step399999/meta.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e9f6dc55e9ad20a9d9fb49c912a1c2df80668233313c10de9d33705ec250ea9c
|
3 |
+
size 425
|
ddpm_pred_x0_cifar10/ckpt/step399999/model.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:727309417f4cd17eeb975340b4da653adb831697a80b93b5a1b64623c8857f04
|
3 |
+
size 143084317
|
ddpm_pred_x0_cifar10/ckpt/step399999/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dab9f540504edc3d3c6e508368fe8bdae815c489c4d19e7c88db90a9e9800354
|
3 |
+
size 286249989
|
ddpm_pred_x0_cifar10/ckpt/step479999/ema.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3e28c7f24ea88ff0b7694ae73812fc3b79519c69396d3266dcce19bf1c9f9c3c
|
3 |
+
size 143067273
|
ddpm_pred_x0_cifar10/ckpt/step479999/meta.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:985bdda8a39e9fb74da57b328f1891a570fad0d423cdd81c18e3de102b8ebd4f
|
3 |
+
size 425
|
ddpm_pred_x0_cifar10/ckpt/step479999/model.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7c8f86ed4de0ea7286f39b8a443f1f4671d9a3c5305048a9209c0e76debee0eb
|
3 |
+
size 143084317
|
ddpm_pred_x0_cifar10/ckpt/step479999/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7d05595006895fe164173ba4d5b1025408495854bdeac076135623bd24b5ee28
|
3 |
+
size 286249989
|
ddpm_pred_x0_cifar10/ckpt/step559999/ema.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d60723e582b6b5fbcbb3eb8b8d40a637cf4fcf83b907a9e23e406405c440cf11
|
3 |
+
size 143067273
|
ddpm_pred_x0_cifar10/ckpt/step559999/meta.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cdc2d49c8d25a3b6ff963b03a7b56b5f32d8c3f976155b6a480b476766196f8f
|
3 |
+
size 425
|
ddpm_pred_x0_cifar10/ckpt/step559999/model.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e6635c847fd00c3329feca454b18ef3b4890f6cc3224134f67820bf0a7219544
|
3 |
+
size 143084317
|
ddpm_pred_x0_cifar10/ckpt/step559999/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2e600dd82f0b7a03a657c0ca0c27dcaa8f3011c6acc10dc99032b10f7ef22a89
|
3 |
+
size 286249989
|
ddpm_pred_x0_cifar10/ckpt/step639999/ema.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7a877b7043569d4e306033c88449bbd9a55c459be0a88e7a00611b659e43cad6
|
3 |
+
size 143067273
|
ddpm_pred_x0_cifar10/ckpt/step639999/meta.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f9174d79f6b5349c61129e6a460309b4fbb79ef89b46c7ebf3a3ef369e264a8e
|
3 |
+
size 425
|
ddpm_pred_x0_cifar10/ckpt/step639999/model.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b8193ea6179292427b4b053e21b852e718a48ef729ef5e1888c97be3e72b4cb9
|
3 |
+
size 143084317
|
ddpm_pred_x0_cifar10/ckpt/step639999/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:41611e3721463acc18ab85f35c4a1927e9246d3a9303c7f4402040e4cef9af18
|
3 |
+
size 286249989
|
ddpm_pred_x0_cifar10/ckpt/step719999/ema.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0381c4709e6109e4a5844a80c2a201890900c53c36fa7ac57cfe317f0e4679be
|
3 |
+
size 143067273
|
ddpm_pred_x0_cifar10/ckpt/step719999/meta.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:baefc936b268cd126cebe024329539b0016de73a64b053eed6cdd7374d85e0ea
|
3 |
+
size 425
|
ddpm_pred_x0_cifar10/ckpt/step719999/model.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e65dc301140cb5d09e95874a1a8c7ecd02c798c3058880bcf6d7a11c76010495
|
3 |
+
size 143084317
|
ddpm_pred_x0_cifar10/ckpt/step719999/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4ea18303173867a6b43b6dcd15dd81246180c4bf00465967c9b3f45386e8c13e
|
3 |
+
size 286249989
|
ddpm_pred_x0_cifar10/ckpt/step799999/ema.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:71474fe843c08630f00d1eb6e9bd54b0ed716514751669d3038ba2bbaabc2976
|
3 |
+
size 143067273
|
ddpm_pred_x0_cifar10/ckpt/step799999/meta.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a7fc389681c8eb2922fbeabf0c843cedbad10aaaa30455d0d2b53d1e86ad206b
|
3 |
+
size 425
|
ddpm_pred_x0_cifar10/ckpt/step799999/model.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5a9c1d6adb3a6589581d9fe08c285dbb8ac79e0d267721e7731e092128120a37
|
3 |
+
size 143084317
|
ddpm_pred_x0_cifar10/ckpt/step799999/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:79fb294e85b13efbc16852bb3196b7388455cc396661c3a19bbc446e6bed0733
|
3 |
+
size 286249989
|
ddpm_pred_x0_cifar10/config-2023-08-07-07-49-01.yaml
ADDED
@@ -0,0 +1,54 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
seed: 2022
|
2 |
+
data:
|
3 |
+
name: CIFAR-10
|
4 |
+
dataroot: ~/data/CIFAR-10/
|
5 |
+
img_size: 32
|
6 |
+
img_channels: 3
|
7 |
+
num_classes: 10
|
8 |
+
dataloader:
|
9 |
+
num_workers: 4
|
10 |
+
pin_memory: true
|
11 |
+
prefetch_factor: 2
|
12 |
+
micro_batch: 0
|
13 |
+
model:
|
14 |
+
target: models.unet.UNet
|
15 |
+
params:
|
16 |
+
in_channels: 3
|
17 |
+
out_channels: 3
|
18 |
+
dim: 128
|
19 |
+
dim_mults:
|
20 |
+
- 1
|
21 |
+
- 2
|
22 |
+
- 2
|
23 |
+
- 2
|
24 |
+
use_attn:
|
25 |
+
- false
|
26 |
+
- true
|
27 |
+
- false
|
28 |
+
- false
|
29 |
+
num_res_blocks: 2
|
30 |
+
n_heads: 1
|
31 |
+
dropout: 0.1
|
32 |
+
ema_decay: 0.9999
|
33 |
+
ema_gradual: true
|
34 |
+
diffusion:
|
35 |
+
target: diffusions.ddpm.DDPM
|
36 |
+
params:
|
37 |
+
total_steps: 1000
|
38 |
+
beta_schedule: linear
|
39 |
+
beta_start: 0.0001
|
40 |
+
beta_end: 0.02
|
41 |
+
objective: pred_x0
|
42 |
+
var_type: fixed_large
|
43 |
+
train:
|
44 |
+
n_steps: 800000
|
45 |
+
batch_size: 128
|
46 |
+
clip_grad_norm: 1.0
|
47 |
+
print_freq: 400
|
48 |
+
save_freq: 10000
|
49 |
+
sample_freq: 5000
|
50 |
+
n_samples: 64
|
51 |
+
optim:
|
52 |
+
target: torch.optim.Adam
|
53 |
+
params:
|
54 |
+
lr: 0.0002
|
ddpm_pred_x0_cifar10/config-2023-08-09-23-34-33.yaml
ADDED
@@ -0,0 +1,54 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
seed: 2022
|
2 |
+
data:
|
3 |
+
name: CIFAR-10
|
4 |
+
dataroot: ~/data/CIFAR-10/
|
5 |
+
img_size: 32
|
6 |
+
img_channels: 3
|
7 |
+
num_classes: 10
|
8 |
+
dataloader:
|
9 |
+
num_workers: 4
|
10 |
+
pin_memory: true
|
11 |
+
prefetch_factor: 2
|
12 |
+
micro_batch: 0
|
13 |
+
model:
|
14 |
+
target: models.unet.UNet
|
15 |
+
params:
|
16 |
+
in_channels: 3
|
17 |
+
out_channels: 3
|
18 |
+
dim: 128
|
19 |
+
dim_mults:
|
20 |
+
- 1
|
21 |
+
- 2
|
22 |
+
- 2
|
23 |
+
- 2
|
24 |
+
use_attn:
|
25 |
+
- false
|
26 |
+
- true
|
27 |
+
- false
|
28 |
+
- false
|
29 |
+
num_res_blocks: 2
|
30 |
+
n_heads: 1
|
31 |
+
dropout: 0.1
|
32 |
+
ema_decay: 0.9999
|
33 |
+
ema_gradual: true
|
34 |
+
diffusion:
|
35 |
+
target: diffusions.ddpm.DDPM
|
36 |
+
params:
|
37 |
+
total_steps: 1000
|
38 |
+
beta_schedule: linear
|
39 |
+
beta_start: 0.0001
|
40 |
+
beta_end: 0.02
|
41 |
+
objective: pred_x0
|
42 |
+
var_type: fixed_large
|
43 |
+
train:
|
44 |
+
n_steps: 800000
|
45 |
+
batch_size: 128
|
46 |
+
clip_grad_norm: 1.0
|
47 |
+
print_freq: 400
|
48 |
+
save_freq: 10000
|
49 |
+
sample_freq: 5000
|
50 |
+
n_samples: 64
|
51 |
+
optim:
|
52 |
+
target: torch.optim.Adam
|
53 |
+
params:
|
54 |
+
lr: 0.0002
|
ddpm_pred_x0_cifar10/output-2023-08-07-07-49-01.log
ADDED
@@ -0,0 +1,847 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
2023-08-07 07:49:04,646 - INFO - Experiment directory: ./runs/ddpm_cifar10_pred_x0
|
2 |
+
2023-08-07 07:49:04,647 - INFO - Number of processes: 4
|
3 |
+
2023-08-07 07:49:04,647 - INFO - Distributed type: DistributedType.MULTI_GPU
|
4 |
+
2023-08-07 07:49:04,647 - INFO - Mixed precision: no
|
5 |
+
2023-08-07 07:49:04,647 - INFO - ==============================
|
6 |
+
2023-08-07 07:49:06,157 - INFO - Size of training set: 50000
|
7 |
+
2023-08-07 07:49:06,157 - INFO - Batch size per process: 32
|
8 |
+
2023-08-07 07:49:06,157 - INFO - Total batch size: 128
|
9 |
+
2023-08-07 07:49:06,157 - INFO - ==============================
|
10 |
+
2023-08-07 07:49:06,976 - INFO - Start training...
|
11 |
+
2023-08-07 07:50:31,019 - INFO - [Train] step: 399, loss: 0.119830, lr: 0.000200
|
12 |
+
2023-08-07 07:51:53,317 - INFO - [Train] step: 799, loss: 0.150414, lr: 0.000200
|
13 |
+
2023-08-07 07:53:15,251 - INFO - [Train] step: 1199, loss: 0.125284, lr: 0.000200
|
14 |
+
2023-08-07 07:54:37,181 - INFO - [Train] step: 1599, loss: 0.163405, lr: 0.000200
|
15 |
+
2023-08-07 07:55:59,305 - INFO - [Train] step: 1999, loss: 0.124573, lr: 0.000200
|
16 |
+
2023-08-07 07:57:21,488 - INFO - [Train] step: 2399, loss: 0.121558, lr: 0.000200
|
17 |
+
2023-08-07 07:58:43,758 - INFO - [Train] step: 2799, loss: 0.140495, lr: 0.000200
|
18 |
+
2023-08-07 08:00:06,082 - INFO - [Train] step: 3199, loss: 0.120343, lr: 0.000200
|
19 |
+
2023-08-07 08:01:28,422 - INFO - [Train] step: 3599, loss: 0.126704, lr: 0.000200
|
20 |
+
2023-08-07 08:02:50,514 - INFO - [Train] step: 3999, loss: 0.129013, lr: 0.000200
|
21 |
+
2023-08-07 08:04:12,450 - INFO - [Train] step: 4399, loss: 0.131795, lr: 0.000200
|
22 |
+
2023-08-07 08:05:34,597 - INFO - [Train] step: 4799, loss: 0.125565, lr: 0.000200
|
23 |
+
2023-08-07 08:07:17,806 - INFO - [Train] step: 5199, loss: 0.090839, lr: 0.000200
|
24 |
+
2023-08-07 08:08:40,145 - INFO - [Train] step: 5599, loss: 0.109404, lr: 0.000200
|
25 |
+
2023-08-07 08:10:02,054 - INFO - [Train] step: 5999, loss: 0.123738, lr: 0.000200
|
26 |
+
2023-08-07 08:11:23,822 - INFO - [Train] step: 6399, loss: 0.119509, lr: 0.000200
|
27 |
+
2023-08-07 08:12:45,922 - INFO - [Train] step: 6799, loss: 0.099912, lr: 0.000200
|
28 |
+
2023-08-07 08:14:07,996 - INFO - [Train] step: 7199, loss: 0.087818, lr: 0.000200
|
29 |
+
2023-08-07 08:15:30,343 - INFO - [Train] step: 7599, loss: 0.119198, lr: 0.000200
|
30 |
+
2023-08-07 08:16:52,457 - INFO - [Train] step: 7999, loss: 0.115591, lr: 0.000200
|
31 |
+
2023-08-07 08:18:14,790 - INFO - [Train] step: 8399, loss: 0.099107, lr: 0.000200
|
32 |
+
2023-08-07 08:19:36,938 - INFO - [Train] step: 8799, loss: 0.148510, lr: 0.000200
|
33 |
+
2023-08-07 08:20:59,065 - INFO - [Train] step: 9199, loss: 0.136262, lr: 0.000200
|
34 |
+
2023-08-07 08:22:21,136 - INFO - [Train] step: 9599, loss: 0.087655, lr: 0.000200
|
35 |
+
2023-08-07 08:23:42,670 - INFO - [Train] step: 9999, loss: 0.128666, lr: 0.000200
|
36 |
+
2023-08-07 08:25:28,794 - INFO - [Train] step: 10399, loss: 0.117899, lr: 0.000200
|
37 |
+
2023-08-07 08:26:50,704 - INFO - [Train] step: 10799, loss: 0.098106, lr: 0.000200
|
38 |
+
2023-08-07 08:28:12,833 - INFO - [Train] step: 11199, loss: 0.104961, lr: 0.000200
|
39 |
+
2023-08-07 08:29:35,175 - INFO - [Train] step: 11599, loss: 0.106024, lr: 0.000200
|
40 |
+
2023-08-07 08:30:57,087 - INFO - [Train] step: 11999, loss: 0.095780, lr: 0.000200
|
41 |
+
2023-08-07 08:32:19,193 - INFO - [Train] step: 12399, loss: 0.110405, lr: 0.000200
|
42 |
+
2023-08-07 08:33:41,542 - INFO - [Train] step: 12799, loss: 0.116221, lr: 0.000200
|
43 |
+
2023-08-07 08:35:03,633 - INFO - [Train] step: 13199, loss: 0.097295, lr: 0.000200
|
44 |
+
2023-08-07 08:36:25,794 - INFO - [Train] step: 13599, loss: 0.146900, lr: 0.000200
|
45 |
+
2023-08-07 08:37:48,075 - INFO - [Train] step: 13999, loss: 0.122996, lr: 0.000200
|
46 |
+
2023-08-07 08:39:10,261 - INFO - [Train] step: 14399, loss: 0.133118, lr: 0.000200
|
47 |
+
2023-08-07 08:40:32,561 - INFO - [Train] step: 14799, loss: 0.077836, lr: 0.000200
|
48 |
+
2023-08-07 08:42:17,045 - INFO - [Train] step: 15199, loss: 0.129952, lr: 0.000200
|
49 |
+
2023-08-07 08:43:39,221 - INFO - [Train] step: 15599, loss: 0.110417, lr: 0.000200
|
50 |
+
2023-08-07 08:45:01,943 - INFO - [Train] step: 15999, loss: 0.099892, lr: 0.000200
|
51 |
+
2023-08-07 08:46:24,017 - INFO - [Train] step: 16399, loss: 0.156371, lr: 0.000200
|
52 |
+
2023-08-07 08:47:46,151 - INFO - [Train] step: 16799, loss: 0.099175, lr: 0.000200
|
53 |
+
2023-08-07 08:49:07,453 - INFO - [Train] step: 17199, loss: 0.133834, lr: 0.000200
|
54 |
+
2023-08-07 08:50:28,968 - INFO - [Train] step: 17599, loss: 0.135953, lr: 0.000200
|
55 |
+
2023-08-07 08:51:51,301 - INFO - [Train] step: 17999, loss: 0.120979, lr: 0.000200
|
56 |
+
2023-08-07 08:53:13,426 - INFO - [Train] step: 18399, loss: 0.112717, lr: 0.000200
|
57 |
+
2023-08-07 08:54:35,491 - INFO - [Train] step: 18799, loss: 0.123487, lr: 0.000200
|
58 |
+
2023-08-07 08:55:57,467 - INFO - [Train] step: 19199, loss: 0.147328, lr: 0.000200
|
59 |
+
2023-08-07 08:57:19,790 - INFO - [Train] step: 19599, loss: 0.096656, lr: 0.000200
|
60 |
+
2023-08-07 08:58:42,138 - INFO - [Train] step: 19999, loss: 0.102965, lr: 0.000200
|
61 |
+
2023-08-07 09:00:26,563 - INFO - [Train] step: 20399, loss: 0.155700, lr: 0.000200
|
62 |
+
2023-08-07 09:01:48,902 - INFO - [Train] step: 20799, loss: 0.129004, lr: 0.000200
|
63 |
+
2023-08-07 09:03:11,239 - INFO - [Train] step: 21199, loss: 0.109404, lr: 0.000200
|
64 |
+
2023-08-07 09:04:33,571 - INFO - [Train] step: 21599, loss: 0.107027, lr: 0.000200
|
65 |
+
2023-08-07 09:05:55,701 - INFO - [Train] step: 21999, loss: 0.107544, lr: 0.000200
|
66 |
+
2023-08-07 09:07:18,032 - INFO - [Train] step: 22399, loss: 0.105604, lr: 0.000200
|
67 |
+
2023-08-07 09:08:39,934 - INFO - [Train] step: 22799, loss: 0.122270, lr: 0.000200
|
68 |
+
2023-08-07 09:10:02,291 - INFO - [Train] step: 23199, loss: 0.128101, lr: 0.000200
|
69 |
+
2023-08-07 09:11:24,398 - INFO - [Train] step: 23599, loss: 0.094124, lr: 0.000200
|
70 |
+
2023-08-07 09:12:46,315 - INFO - [Train] step: 23999, loss: 0.125352, lr: 0.000200
|
71 |
+
2023-08-07 09:14:07,859 - INFO - [Train] step: 24399, loss: 0.095635, lr: 0.000200
|
72 |
+
2023-08-07 09:15:29,952 - INFO - [Train] step: 24799, loss: 0.083003, lr: 0.000200
|
73 |
+
2023-08-07 09:17:12,575 - INFO - [Train] step: 25199, loss: 0.090902, lr: 0.000200
|
74 |
+
2023-08-07 09:18:34,666 - INFO - [Train] step: 25599, loss: 0.093669, lr: 0.000200
|
75 |
+
2023-08-07 09:19:57,022 - INFO - [Train] step: 25999, loss: 0.115352, lr: 0.000200
|
76 |
+
2023-08-07 09:21:18,751 - INFO - [Train] step: 26399, loss: 0.113375, lr: 0.000200
|
77 |
+
2023-08-07 09:22:40,442 - INFO - [Train] step: 26799, loss: 0.122325, lr: 0.000200
|
78 |
+
2023-08-07 09:24:02,440 - INFO - [Train] step: 27199, loss: 0.116420, lr: 0.000200
|
79 |
+
2023-08-07 09:25:24,509 - INFO - [Train] step: 27599, loss: 0.139663, lr: 0.000200
|
80 |
+
2023-08-07 09:26:46,422 - INFO - [Train] step: 27999, loss: 0.103903, lr: 0.000200
|
81 |
+
2023-08-07 09:28:08,147 - INFO - [Train] step: 28399, loss: 0.116586, lr: 0.000200
|
82 |
+
2023-08-07 09:29:30,259 - INFO - [Train] step: 28799, loss: 0.152031, lr: 0.000200
|
83 |
+
2023-08-07 09:30:52,145 - INFO - [Train] step: 29199, loss: 0.099812, lr: 0.000200
|
84 |
+
2023-08-07 09:32:14,272 - INFO - [Train] step: 29599, loss: 0.112326, lr: 0.000200
|
85 |
+
2023-08-07 09:33:36,637 - INFO - [Train] step: 29999, loss: 0.126363, lr: 0.000200
|
86 |
+
2023-08-07 09:35:21,244 - INFO - [Train] step: 30399, loss: 0.105871, lr: 0.000200
|
87 |
+
2023-08-07 09:36:43,452 - INFO - [Train] step: 30799, loss: 0.130065, lr: 0.000200
|
88 |
+
2023-08-07 09:38:05,688 - INFO - [Train] step: 31199, loss: 0.105673, lr: 0.000200
|
89 |
+
2023-08-07 09:39:28,109 - INFO - [Train] step: 31599, loss: 0.102385, lr: 0.000200
|
90 |
+
2023-08-07 09:40:50,177 - INFO - [Train] step: 31999, loss: 0.130559, lr: 0.000200
|
91 |
+
2023-08-07 09:42:12,305 - INFO - [Train] step: 32399, loss: 0.096475, lr: 0.000200
|
92 |
+
2023-08-07 09:43:34,656 - INFO - [Train] step: 32799, loss: 0.101694, lr: 0.000200
|
93 |
+
2023-08-07 09:44:56,970 - INFO - [Train] step: 33199, loss: 0.115387, lr: 0.000200
|
94 |
+
2023-08-07 09:46:19,111 - INFO - [Train] step: 33599, loss: 0.127949, lr: 0.000200
|
95 |
+
2023-08-07 09:47:41,435 - INFO - [Train] step: 33999, loss: 0.112779, lr: 0.000200
|
96 |
+
2023-08-07 09:49:03,543 - INFO - [Train] step: 34399, loss: 0.134133, lr: 0.000200
|
97 |
+
2023-08-07 09:50:25,707 - INFO - [Train] step: 34799, loss: 0.111700, lr: 0.000200
|
98 |
+
2023-08-07 09:52:09,128 - INFO - [Train] step: 35199, loss: 0.135414, lr: 0.000200
|
99 |
+
2023-08-07 09:53:31,407 - INFO - [Train] step: 35599, loss: 0.119928, lr: 0.000200
|
100 |
+
2023-08-07 09:54:53,372 - INFO - [Train] step: 35999, loss: 0.115652, lr: 0.000200
|
101 |
+
2023-08-07 09:56:15,305 - INFO - [Train] step: 36399, loss: 0.100819, lr: 0.000200
|
102 |
+
2023-08-07 09:57:37,592 - INFO - [Train] step: 36799, loss: 0.120232, lr: 0.000200
|
103 |
+
2023-08-07 09:58:59,540 - INFO - [Train] step: 37199, loss: 0.142750, lr: 0.000200
|
104 |
+
2023-08-07 10:00:21,641 - INFO - [Train] step: 37599, loss: 0.070434, lr: 0.000200
|
105 |
+
2023-08-07 10:01:43,790 - INFO - [Train] step: 37999, loss: 0.112740, lr: 0.000200
|
106 |
+
2023-08-07 10:03:05,922 - INFO - [Train] step: 38399, loss: 0.112250, lr: 0.000200
|
107 |
+
2023-08-07 10:04:28,247 - INFO - [Train] step: 38799, loss: 0.079933, lr: 0.000200
|
108 |
+
2023-08-07 10:05:50,122 - INFO - [Train] step: 39199, loss: 0.113054, lr: 0.000200
|
109 |
+
2023-08-07 10:07:12,298 - INFO - [Train] step: 39599, loss: 0.109953, lr: 0.000200
|
110 |
+
2023-08-07 10:08:34,516 - INFO - [Train] step: 39999, loss: 0.130182, lr: 0.000200
|
111 |
+
2023-08-07 10:10:17,846 - INFO - [Train] step: 40399, loss: 0.149816, lr: 0.000200
|
112 |
+
2023-08-07 10:11:39,571 - INFO - [Train] step: 40799, loss: 0.101177, lr: 0.000200
|
113 |
+
2023-08-07 10:13:01,846 - INFO - [Train] step: 41199, loss: 0.100172, lr: 0.000200
|
114 |
+
2023-08-07 10:14:23,798 - INFO - [Train] step: 41599, loss: 0.104291, lr: 0.000200
|
115 |
+
2023-08-07 10:15:46,138 - INFO - [Train] step: 41999, loss: 0.104690, lr: 0.000200
|
116 |
+
2023-08-07 10:17:08,442 - INFO - [Train] step: 42399, loss: 0.112855, lr: 0.000200
|
117 |
+
2023-08-07 10:18:30,572 - INFO - [Train] step: 42799, loss: 0.095969, lr: 0.000200
|
118 |
+
2023-08-07 10:19:52,666 - INFO - [Train] step: 43199, loss: 0.083825, lr: 0.000200
|
119 |
+
2023-08-07 10:21:14,840 - INFO - [Train] step: 43599, loss: 0.133080, lr: 0.000200
|
120 |
+
2023-08-07 10:22:36,781 - INFO - [Train] step: 43999, loss: 0.069590, lr: 0.000200
|
121 |
+
2023-08-07 10:23:59,100 - INFO - [Train] step: 44399, loss: 0.090524, lr: 0.000200
|
122 |
+
2023-08-07 10:25:21,013 - INFO - [Train] step: 44799, loss: 0.128244, lr: 0.000200
|
123 |
+
2023-08-07 10:27:04,202 - INFO - [Train] step: 45199, loss: 0.101210, lr: 0.000200
|
124 |
+
2023-08-07 10:28:25,700 - INFO - [Train] step: 45599, loss: 0.100796, lr: 0.000200
|
125 |
+
2023-08-07 10:29:47,874 - INFO - [Train] step: 45999, loss: 0.091977, lr: 0.000200
|
126 |
+
2023-08-07 10:31:09,976 - INFO - [Train] step: 46399, loss: 0.088182, lr: 0.000200
|
127 |
+
2023-08-07 10:32:32,410 - INFO - [Train] step: 46799, loss: 0.121288, lr: 0.000200
|
128 |
+
2023-08-07 10:33:54,859 - INFO - [Train] step: 47199, loss: 0.167538, lr: 0.000200
|
129 |
+
2023-08-07 10:35:16,362 - INFO - [Train] step: 47599, loss: 0.156691, lr: 0.000200
|
130 |
+
2023-08-07 10:36:38,279 - INFO - [Train] step: 47999, loss: 0.122817, lr: 0.000200
|
131 |
+
2023-08-07 10:38:00,182 - INFO - [Train] step: 48399, loss: 0.098979, lr: 0.000200
|
132 |
+
2023-08-07 10:39:22,149 - INFO - [Train] step: 48799, loss: 0.118480, lr: 0.000200
|
133 |
+
2023-08-07 10:40:44,057 - INFO - [Train] step: 49199, loss: 0.093471, lr: 0.000200
|
134 |
+
2023-08-07 10:42:06,400 - INFO - [Train] step: 49599, loss: 0.108082, lr: 0.000200
|
135 |
+
2023-08-07 10:43:28,706 - INFO - [Train] step: 49999, loss: 0.122123, lr: 0.000200
|
136 |
+
2023-08-07 10:45:13,391 - INFO - [Train] step: 50399, loss: 0.127382, lr: 0.000200
|
137 |
+
2023-08-07 10:46:35,084 - INFO - [Train] step: 50799, loss: 0.136278, lr: 0.000200
|
138 |
+
2023-08-07 10:47:57,204 - INFO - [Train] step: 51199, loss: 0.110470, lr: 0.000200
|
139 |
+
2023-08-07 10:49:19,309 - INFO - [Train] step: 51599, loss: 0.142462, lr: 0.000200
|
140 |
+
2023-08-07 10:50:41,676 - INFO - [Train] step: 51999, loss: 0.110496, lr: 0.000200
|
141 |
+
2023-08-07 10:52:03,801 - INFO - [Train] step: 52399, loss: 0.119802, lr: 0.000200
|
142 |
+
2023-08-07 10:53:25,499 - INFO - [Train] step: 52799, loss: 0.120931, lr: 0.000200
|
143 |
+
2023-08-07 10:54:47,625 - INFO - [Train] step: 53199, loss: 0.130146, lr: 0.000200
|
144 |
+
2023-08-07 10:56:09,767 - INFO - [Train] step: 53599, loss: 0.099482, lr: 0.000200
|
145 |
+
2023-08-07 10:57:31,440 - INFO - [Train] step: 53999, loss: 0.117266, lr: 0.000200
|
146 |
+
2023-08-07 10:58:53,373 - INFO - [Train] step: 54399, loss: 0.087286, lr: 0.000200
|
147 |
+
2023-08-07 11:00:15,523 - INFO - [Train] step: 54799, loss: 0.131403, lr: 0.000200
|
148 |
+
2023-08-07 11:01:59,765 - INFO - [Train] step: 55199, loss: 0.104190, lr: 0.000200
|
149 |
+
2023-08-07 11:03:21,449 - INFO - [Train] step: 55599, loss: 0.118375, lr: 0.000200
|
150 |
+
2023-08-07 11:04:43,375 - INFO - [Train] step: 55999, loss: 0.091755, lr: 0.000200
|
151 |
+
2023-08-07 11:06:05,712 - INFO - [Train] step: 56399, loss: 0.109616, lr: 0.000200
|
152 |
+
2023-08-07 11:07:28,055 - INFO - [Train] step: 56799, loss: 0.124397, lr: 0.000200
|
153 |
+
2023-08-07 11:08:50,180 - INFO - [Train] step: 57199, loss: 0.108280, lr: 0.000200
|
154 |
+
2023-08-07 11:10:12,507 - INFO - [Train] step: 57599, loss: 0.092169, lr: 0.000200
|
155 |
+
2023-08-07 11:11:34,645 - INFO - [Train] step: 57999, loss: 0.083924, lr: 0.000200
|
156 |
+
2023-08-07 11:12:56,983 - INFO - [Train] step: 58399, loss: 0.134117, lr: 0.000200
|
157 |
+
2023-08-07 11:14:19,303 - INFO - [Train] step: 58799, loss: 0.106008, lr: 0.000200
|
158 |
+
2023-08-07 11:15:41,433 - INFO - [Train] step: 59199, loss: 0.109932, lr: 0.000200
|
159 |
+
2023-08-07 11:17:03,770 - INFO - [Train] step: 59599, loss: 0.122537, lr: 0.000200
|
160 |
+
2023-08-07 11:18:26,093 - INFO - [Train] step: 59999, loss: 0.091691, lr: 0.000200
|
161 |
+
2023-08-07 11:20:11,139 - INFO - [Train] step: 60399, loss: 0.125825, lr: 0.000200
|
162 |
+
2023-08-07 11:21:33,071 - INFO - [Train] step: 60799, loss: 0.094548, lr: 0.000200
|
163 |
+
2023-08-07 11:22:54,599 - INFO - [Train] step: 61199, loss: 0.104905, lr: 0.000200
|
164 |
+
2023-08-07 11:24:16,730 - INFO - [Train] step: 61599, loss: 0.133830, lr: 0.000200
|
165 |
+
2023-08-07 11:25:38,899 - INFO - [Train] step: 61999, loss: 0.112731, lr: 0.000200
|
166 |
+
2023-08-07 11:27:01,077 - INFO - [Train] step: 62399, loss: 0.100290, lr: 0.000200
|
167 |
+
2023-08-07 11:28:23,330 - INFO - [Train] step: 62799, loss: 0.128646, lr: 0.000200
|
168 |
+
2023-08-07 11:29:45,494 - INFO - [Train] step: 63199, loss: 0.107791, lr: 0.000200
|
169 |
+
2023-08-07 11:31:06,766 - INFO - [Train] step: 63599, loss: 0.107345, lr: 0.000200
|
170 |
+
2023-08-07 11:32:29,095 - INFO - [Train] step: 63999, loss: 0.119336, lr: 0.000200
|
171 |
+
2023-08-07 11:33:51,232 - INFO - [Train] step: 64399, loss: 0.120495, lr: 0.000200
|
172 |
+
2023-08-07 11:35:13,370 - INFO - [Train] step: 64799, loss: 0.115367, lr: 0.000200
|
173 |
+
2023-08-07 11:36:57,599 - INFO - [Train] step: 65199, loss: 0.108854, lr: 0.000200
|
174 |
+
2023-08-07 11:38:19,702 - INFO - [Train] step: 65599, loss: 0.169773, lr: 0.000200
|
175 |
+
2023-08-07 11:39:42,067 - INFO - [Train] step: 65999, loss: 0.147495, lr: 0.000200
|
176 |
+
2023-08-07 11:41:04,404 - INFO - [Train] step: 66399, loss: 0.105130, lr: 0.000200
|
177 |
+
2023-08-07 11:42:26,107 - INFO - [Train] step: 66799, loss: 0.126324, lr: 0.000200
|
178 |
+
2023-08-07 11:43:48,447 - INFO - [Train] step: 67199, loss: 0.127703, lr: 0.000200
|
179 |
+
2023-08-07 11:45:10,773 - INFO - [Train] step: 67599, loss: 0.138544, lr: 0.000200
|
180 |
+
2023-08-07 11:46:32,664 - INFO - [Train] step: 67999, loss: 0.120105, lr: 0.000200
|
181 |
+
2023-08-07 11:47:54,421 - INFO - [Train] step: 68399, loss: 0.126496, lr: 0.000200
|
182 |
+
2023-08-07 11:49:16,629 - INFO - [Train] step: 68799, loss: 0.121904, lr: 0.000200
|
183 |
+
2023-08-07 11:50:38,465 - INFO - [Train] step: 69199, loss: 0.094325, lr: 0.000200
|
184 |
+
2023-08-07 11:52:00,802 - INFO - [Train] step: 69599, loss: 0.095884, lr: 0.000200
|
185 |
+
2023-08-07 11:53:22,706 - INFO - [Train] step: 69999, loss: 0.151581, lr: 0.000200
|
186 |
+
2023-08-07 11:55:09,013 - INFO - [Train] step: 70399, loss: 0.124470, lr: 0.000200
|
187 |
+
2023-08-07 11:56:30,934 - INFO - [Train] step: 70799, loss: 0.121488, lr: 0.000200
|
188 |
+
2023-08-07 11:57:52,827 - INFO - [Train] step: 71199, loss: 0.106889, lr: 0.000200
|
189 |
+
2023-08-07 11:59:14,349 - INFO - [Train] step: 71599, loss: 0.130704, lr: 0.000200
|
190 |
+
2023-08-07 12:00:36,423 - INFO - [Train] step: 71999, loss: 0.087750, lr: 0.000200
|
191 |
+
2023-08-07 12:01:58,383 - INFO - [Train] step: 72399, loss: 0.095367, lr: 0.000200
|
192 |
+
2023-08-07 12:03:20,257 - INFO - [Train] step: 72799, loss: 0.096545, lr: 0.000200
|
193 |
+
2023-08-07 12:04:42,188 - INFO - [Train] step: 73199, loss: 0.099555, lr: 0.000200
|
194 |
+
2023-08-07 12:06:04,337 - INFO - [Train] step: 73599, loss: 0.145404, lr: 0.000200
|
195 |
+
2023-08-07 12:07:26,269 - INFO - [Train] step: 73999, loss: 0.103730, lr: 0.000200
|
196 |
+
2023-08-07 12:08:48,551 - INFO - [Train] step: 74399, loss: 0.143072, lr: 0.000200
|
197 |
+
2023-08-07 12:10:10,691 - INFO - [Train] step: 74799, loss: 0.095879, lr: 0.000200
|
198 |
+
2023-08-07 12:11:54,112 - INFO - [Train] step: 75199, loss: 0.130812, lr: 0.000200
|
199 |
+
2023-08-07 12:13:16,265 - INFO - [Train] step: 75599, loss: 0.079059, lr: 0.000200
|
200 |
+
2023-08-07 12:14:38,388 - INFO - [Train] step: 75999, loss: 0.107061, lr: 0.000200
|
201 |
+
2023-08-07 12:16:00,316 - INFO - [Train] step: 76399, loss: 0.097370, lr: 0.000200
|
202 |
+
2023-08-07 12:17:21,601 - INFO - [Train] step: 76799, loss: 0.115616, lr: 0.000200
|
203 |
+
2023-08-07 12:18:43,734 - INFO - [Train] step: 77199, loss: 0.115731, lr: 0.000200
|
204 |
+
2023-08-07 12:20:05,703 - INFO - [Train] step: 77599, loss: 0.126601, lr: 0.000200
|
205 |
+
2023-08-07 12:21:27,856 - INFO - [Train] step: 77999, loss: 0.099151, lr: 0.000200
|
206 |
+
2023-08-07 12:22:50,083 - INFO - [Train] step: 78399, loss: 0.098982, lr: 0.000200
|
207 |
+
2023-08-07 12:24:12,235 - INFO - [Train] step: 78799, loss: 0.108005, lr: 0.000200
|
208 |
+
2023-08-07 12:25:34,144 - INFO - [Train] step: 79199, loss: 0.110650, lr: 0.000200
|
209 |
+
2023-08-07 12:26:56,289 - INFO - [Train] step: 79599, loss: 0.077124, lr: 0.000200
|
210 |
+
2023-08-07 12:28:18,061 - INFO - [Train] step: 79999, loss: 0.119822, lr: 0.000200
|
211 |
+
2023-08-07 12:30:03,862 - INFO - [Train] step: 80399, loss: 0.117670, lr: 0.000200
|
212 |
+
2023-08-07 12:31:25,371 - INFO - [Train] step: 80799, loss: 0.116992, lr: 0.000200
|
213 |
+
2023-08-07 12:32:47,496 - INFO - [Train] step: 81199, loss: 0.079119, lr: 0.000200
|
214 |
+
2023-08-07 12:34:09,635 - INFO - [Train] step: 81599, loss: 0.106365, lr: 0.000200
|
215 |
+
2023-08-07 12:35:31,786 - INFO - [Train] step: 81999, loss: 0.120059, lr: 0.000200
|
216 |
+
2023-08-07 12:36:53,317 - INFO - [Train] step: 82399, loss: 0.086176, lr: 0.000200
|
217 |
+
2023-08-07 12:38:15,182 - INFO - [Train] step: 82799, loss: 0.096099, lr: 0.000200
|
218 |
+
2023-08-07 12:39:37,105 - INFO - [Train] step: 83199, loss: 0.084154, lr: 0.000200
|
219 |
+
2023-08-07 12:40:59,243 - INFO - [Train] step: 83599, loss: 0.121638, lr: 0.000200
|
220 |
+
2023-08-07 12:42:21,342 - INFO - [Train] step: 83999, loss: 0.117321, lr: 0.000200
|
221 |
+
2023-08-07 12:43:43,387 - INFO - [Train] step: 84399, loss: 0.106803, lr: 0.000200
|
222 |
+
2023-08-07 12:45:05,578 - INFO - [Train] step: 84799, loss: 0.083176, lr: 0.000200
|
223 |
+
2023-08-07 12:46:49,266 - INFO - [Train] step: 85199, loss: 0.087706, lr: 0.000200
|
224 |
+
2023-08-07 12:48:11,556 - INFO - [Train] step: 85599, loss: 0.111652, lr: 0.000200
|
225 |
+
2023-08-07 12:49:33,502 - INFO - [Train] step: 85999, loss: 0.075117, lr: 0.000200
|
226 |
+
2023-08-07 12:50:55,682 - INFO - [Train] step: 86399, loss: 0.094607, lr: 0.000200
|
227 |
+
2023-08-07 12:52:17,342 - INFO - [Train] step: 86799, loss: 0.163171, lr: 0.000200
|
228 |
+
2023-08-07 12:53:38,848 - INFO - [Train] step: 87199, loss: 0.125229, lr: 0.000200
|
229 |
+
2023-08-07 12:55:01,179 - INFO - [Train] step: 87599, loss: 0.128466, lr: 0.000200
|
230 |
+
2023-08-07 12:56:23,496 - INFO - [Train] step: 87999, loss: 0.108964, lr: 0.000200
|
231 |
+
2023-08-07 12:57:45,834 - INFO - [Train] step: 88399, loss: 0.099056, lr: 0.000200
|
232 |
+
2023-08-07 12:59:07,740 - INFO - [Train] step: 88799, loss: 0.084511, lr: 0.000200
|
233 |
+
2023-08-07 13:00:29,869 - INFO - [Train] step: 89199, loss: 0.104628, lr: 0.000200
|
234 |
+
2023-08-07 13:01:51,611 - INFO - [Train] step: 89599, loss: 0.075985, lr: 0.000200
|
235 |
+
2023-08-07 13:03:13,539 - INFO - [Train] step: 89999, loss: 0.107070, lr: 0.000200
|
236 |
+
2023-08-07 13:04:58,186 - INFO - [Train] step: 90399, loss: 0.135741, lr: 0.000200
|
237 |
+
2023-08-07 13:06:20,294 - INFO - [Train] step: 90799, loss: 0.130711, lr: 0.000200
|
238 |
+
2023-08-07 13:07:42,443 - INFO - [Train] step: 91199, loss: 0.087756, lr: 0.000200
|
239 |
+
2023-08-07 13:09:04,365 - INFO - [Train] step: 91599, loss: 0.117803, lr: 0.000200
|
240 |
+
2023-08-07 13:10:26,489 - INFO - [Train] step: 91999, loss: 0.070383, lr: 0.000200
|
241 |
+
2023-08-07 13:11:48,803 - INFO - [Train] step: 92399, loss: 0.110226, lr: 0.000200
|
242 |
+
2023-08-07 13:13:11,136 - INFO - [Train] step: 92799, loss: 0.104816, lr: 0.000200
|
243 |
+
2023-08-07 13:14:33,057 - INFO - [Train] step: 93199, loss: 0.130824, lr: 0.000200
|
244 |
+
2023-08-07 13:15:55,293 - INFO - [Train] step: 93599, loss: 0.091706, lr: 0.000200
|
245 |
+
2023-08-07 13:17:17,910 - INFO - [Train] step: 93999, loss: 0.108431, lr: 0.000200
|
246 |
+
2023-08-07 13:18:40,259 - INFO - [Train] step: 94399, loss: 0.095255, lr: 0.000200
|
247 |
+
2023-08-07 13:20:01,989 - INFO - [Train] step: 94799, loss: 0.128445, lr: 0.000200
|
248 |
+
2023-08-07 13:21:45,582 - INFO - [Train] step: 95199, loss: 0.136346, lr: 0.000200
|
249 |
+
2023-08-07 13:23:07,352 - INFO - [Train] step: 95599, loss: 0.108398, lr: 0.000200
|
250 |
+
2023-08-07 13:24:29,052 - INFO - [Train] step: 95999, loss: 0.102946, lr: 0.000200
|
251 |
+
2023-08-07 13:25:51,135 - INFO - [Train] step: 96399, loss: 0.115570, lr: 0.000200
|
252 |
+
2023-08-07 13:27:13,484 - INFO - [Train] step: 96799, loss: 0.134736, lr: 0.000200
|
253 |
+
2023-08-07 13:28:35,801 - INFO - [Train] step: 97199, loss: 0.121527, lr: 0.000200
|
254 |
+
2023-08-07 13:29:56,938 - INFO - [Train] step: 97599, loss: 0.087732, lr: 0.000200
|
255 |
+
2023-08-07 13:31:19,072 - INFO - [Train] step: 97999, loss: 0.106258, lr: 0.000200
|
256 |
+
2023-08-07 13:32:41,389 - INFO - [Train] step: 98399, loss: 0.102467, lr: 0.000200
|
257 |
+
2023-08-07 13:34:03,501 - INFO - [Train] step: 98799, loss: 0.112881, lr: 0.000200
|
258 |
+
2023-08-07 13:35:25,325 - INFO - [Train] step: 99199, loss: 0.099730, lr: 0.000200
|
259 |
+
2023-08-07 13:36:47,359 - INFO - [Train] step: 99599, loss: 0.132975, lr: 0.000200
|
260 |
+
2023-08-07 13:38:09,688 - INFO - [Train] step: 99999, loss: 0.109757, lr: 0.000200
|
261 |
+
2023-08-07 13:39:55,393 - INFO - [Train] step: 100399, loss: 0.119519, lr: 0.000200
|
262 |
+
2023-08-07 13:41:17,492 - INFO - [Train] step: 100799, loss: 0.095384, lr: 0.000200
|
263 |
+
2023-08-07 13:42:39,812 - INFO - [Train] step: 101199, loss: 0.094317, lr: 0.000200
|
264 |
+
2023-08-07 13:44:01,951 - INFO - [Train] step: 101599, loss: 0.077808, lr: 0.000200
|
265 |
+
2023-08-07 13:45:24,292 - INFO - [Train] step: 101999, loss: 0.131864, lr: 0.000200
|
266 |
+
2023-08-07 13:46:46,178 - INFO - [Train] step: 102399, loss: 0.104115, lr: 0.000200
|
267 |
+
2023-08-07 13:48:08,526 - INFO - [Train] step: 102799, loss: 0.118189, lr: 0.000200
|
268 |
+
2023-08-07 13:49:30,230 - INFO - [Train] step: 103199, loss: 0.124610, lr: 0.000200
|
269 |
+
2023-08-07 13:50:52,359 - INFO - [Train] step: 103599, loss: 0.106008, lr: 0.000200
|
270 |
+
2023-08-07 13:52:14,291 - INFO - [Train] step: 103999, loss: 0.086552, lr: 0.000200
|
271 |
+
2023-08-07 13:53:36,466 - INFO - [Train] step: 104399, loss: 0.091638, lr: 0.000200
|
272 |
+
2023-08-07 13:54:57,932 - INFO - [Train] step: 104799, loss: 0.113715, lr: 0.000200
|
273 |
+
2023-08-07 13:56:42,356 - INFO - [Train] step: 105199, loss: 0.101307, lr: 0.000200
|
274 |
+
2023-08-07 13:58:04,313 - INFO - [Train] step: 105599, loss: 0.086030, lr: 0.000200
|
275 |
+
2023-08-07 13:59:25,996 - INFO - [Train] step: 105999, loss: 0.110670, lr: 0.000200
|
276 |
+
2023-08-07 14:00:48,345 - INFO - [Train] step: 106399, loss: 0.105311, lr: 0.000200
|
277 |
+
2023-08-07 14:02:10,689 - INFO - [Train] step: 106799, loss: 0.143710, lr: 0.000200
|
278 |
+
2023-08-07 14:03:32,975 - INFO - [Train] step: 107199, loss: 0.104342, lr: 0.000200
|
279 |
+
2023-08-07 14:04:54,932 - INFO - [Train] step: 107599, loss: 0.128174, lr: 0.000200
|
280 |
+
2023-08-07 14:06:16,847 - INFO - [Train] step: 107999, loss: 0.108958, lr: 0.000200
|
281 |
+
2023-08-07 14:07:38,959 - INFO - [Train] step: 108399, loss: 0.117044, lr: 0.000200
|
282 |
+
2023-08-07 14:09:01,105 - INFO - [Train] step: 108799, loss: 0.104867, lr: 0.000200
|
283 |
+
2023-08-07 14:10:23,319 - INFO - [Train] step: 109199, loss: 0.118993, lr: 0.000200
|
284 |
+
2023-08-07 14:11:45,736 - INFO - [Train] step: 109599, loss: 0.107032, lr: 0.000200
|
285 |
+
2023-08-07 14:13:08,073 - INFO - [Train] step: 109999, loss: 0.119035, lr: 0.000200
|
286 |
+
2023-08-07 14:14:53,768 - INFO - [Train] step: 110399, loss: 0.123649, lr: 0.000200
|
287 |
+
2023-08-07 14:16:15,699 - INFO - [Train] step: 110799, loss: 0.115585, lr: 0.000200
|
288 |
+
2023-08-07 14:17:38,027 - INFO - [Train] step: 111199, loss: 0.114956, lr: 0.000200
|
289 |
+
2023-08-07 14:19:00,161 - INFO - [Train] step: 111599, loss: 0.074508, lr: 0.000200
|
290 |
+
2023-08-07 14:20:22,478 - INFO - [Train] step: 111999, loss: 0.136996, lr: 0.000200
|
291 |
+
2023-08-07 14:21:44,817 - INFO - [Train] step: 112399, loss: 0.118663, lr: 0.000200
|
292 |
+
2023-08-07 14:23:06,937 - INFO - [Train] step: 112799, loss: 0.153161, lr: 0.000200
|
293 |
+
2023-08-07 14:24:28,859 - INFO - [Train] step: 113199, loss: 0.111933, lr: 0.000200
|
294 |
+
2023-08-07 14:25:50,374 - INFO - [Train] step: 113599, loss: 0.082598, lr: 0.000200
|
295 |
+
2023-08-07 14:27:12,053 - INFO - [Train] step: 113999, loss: 0.120279, lr: 0.000200
|
296 |
+
2023-08-07 14:28:34,077 - INFO - [Train] step: 114399, loss: 0.075852, lr: 0.000200
|
297 |
+
2023-08-07 14:29:56,128 - INFO - [Train] step: 114799, loss: 0.094710, lr: 0.000200
|
298 |
+
2023-08-07 14:31:40,171 - INFO - [Train] step: 115199, loss: 0.124519, lr: 0.000200
|
299 |
+
2023-08-07 14:33:02,143 - INFO - [Train] step: 115599, loss: 0.103630, lr: 0.000200
|
300 |
+
2023-08-07 14:34:24,194 - INFO - [Train] step: 115999, loss: 0.104483, lr: 0.000200
|
301 |
+
2023-08-07 14:35:46,364 - INFO - [Train] step: 116399, loss: 0.133401, lr: 0.000200
|
302 |
+
2023-08-07 14:37:08,208 - INFO - [Train] step: 116799, loss: 0.128127, lr: 0.000200
|
303 |
+
2023-08-07 14:38:29,774 - INFO - [Train] step: 117199, loss: 0.116914, lr: 0.000200
|
304 |
+
2023-08-07 14:39:52,111 - INFO - [Train] step: 117599, loss: 0.115837, lr: 0.000200
|
305 |
+
2023-08-07 14:41:14,440 - INFO - [Train] step: 117999, loss: 0.135401, lr: 0.000200
|
306 |
+
2023-08-07 14:42:35,960 - INFO - [Train] step: 118399, loss: 0.123450, lr: 0.000200
|
307 |
+
2023-08-07 14:43:57,870 - INFO - [Train] step: 118799, loss: 0.125168, lr: 0.000200
|
308 |
+
2023-08-07 14:45:19,982 - INFO - [Train] step: 119199, loss: 0.102492, lr: 0.000200
|
309 |
+
2023-08-07 14:46:42,301 - INFO - [Train] step: 119599, loss: 0.093711, lr: 0.000200
|
310 |
+
2023-08-07 14:48:04,397 - INFO - [Train] step: 119999, loss: 0.128934, lr: 0.000200
|
311 |
+
2023-08-07 14:49:48,874 - INFO - [Train] step: 120399, loss: 0.119873, lr: 0.000200
|
312 |
+
2023-08-07 14:51:10,594 - INFO - [Train] step: 120799, loss: 0.112037, lr: 0.000200
|
313 |
+
2023-08-07 14:52:32,688 - INFO - [Train] step: 121199, loss: 0.108190, lr: 0.000200
|
314 |
+
2023-08-07 14:53:55,035 - INFO - [Train] step: 121599, loss: 0.115544, lr: 0.000200
|
315 |
+
2023-08-07 14:55:16,909 - INFO - [Train] step: 121999, loss: 0.091772, lr: 0.000200
|
316 |
+
2023-08-07 14:56:38,769 - INFO - [Train] step: 122399, loss: 0.166007, lr: 0.000200
|
317 |
+
2023-08-07 14:58:00,993 - INFO - [Train] step: 122799, loss: 0.102445, lr: 0.000200
|
318 |
+
2023-08-07 14:59:22,894 - INFO - [Train] step: 123199, loss: 0.104988, lr: 0.000200
|
319 |
+
2023-08-07 15:00:45,026 - INFO - [Train] step: 123599, loss: 0.103103, lr: 0.000200
|
320 |
+
2023-08-07 15:02:07,329 - INFO - [Train] step: 123999, loss: 0.094463, lr: 0.000200
|
321 |
+
2023-08-07 15:03:29,469 - INFO - [Train] step: 124399, loss: 0.111020, lr: 0.000200
|
322 |
+
2023-08-07 15:04:51,914 - INFO - [Train] step: 124799, loss: 0.133209, lr: 0.000200
|
323 |
+
2023-08-07 15:06:36,109 - INFO - [Train] step: 125199, loss: 0.117668, lr: 0.000200
|
324 |
+
2023-08-07 15:07:57,989 - INFO - [Train] step: 125599, loss: 0.109193, lr: 0.000200
|
325 |
+
2023-08-07 15:09:19,295 - INFO - [Train] step: 125999, loss: 0.131822, lr: 0.000200
|
326 |
+
2023-08-07 15:10:41,415 - INFO - [Train] step: 126399, loss: 0.105944, lr: 0.000200
|
327 |
+
2023-08-07 15:12:03,779 - INFO - [Train] step: 126799, loss: 0.104178, lr: 0.000200
|
328 |
+
2023-08-07 15:13:26,058 - INFO - [Train] step: 127199, loss: 0.101726, lr: 0.000200
|
329 |
+
2023-08-07 15:14:47,982 - INFO - [Train] step: 127599, loss: 0.156785, lr: 0.000200
|
330 |
+
2023-08-07 15:16:09,905 - INFO - [Train] step: 127999, loss: 0.099558, lr: 0.000200
|
331 |
+
2023-08-07 15:17:32,029 - INFO - [Train] step: 128399, loss: 0.121172, lr: 0.000200
|
332 |
+
2023-08-07 15:18:54,346 - INFO - [Train] step: 128799, loss: 0.088598, lr: 0.000200
|
333 |
+
2023-08-07 15:20:16,545 - INFO - [Train] step: 129199, loss: 0.097450, lr: 0.000200
|
334 |
+
2023-08-07 15:21:38,819 - INFO - [Train] step: 129599, loss: 0.118880, lr: 0.000200
|
335 |
+
2023-08-07 15:23:01,154 - INFO - [Train] step: 129999, loss: 0.115626, lr: 0.000200
|
336 |
+
2023-08-07 15:24:47,652 - INFO - [Train] step: 130399, loss: 0.078181, lr: 0.000200
|
337 |
+
2023-08-07 15:26:09,359 - INFO - [Train] step: 130799, loss: 0.106813, lr: 0.000200
|
338 |
+
2023-08-07 15:27:31,281 - INFO - [Train] step: 131199, loss: 0.139148, lr: 0.000200
|
339 |
+
2023-08-07 15:28:53,612 - INFO - [Train] step: 131599, loss: 0.125892, lr: 0.000200
|
340 |
+
2023-08-07 15:30:15,717 - INFO - [Train] step: 131999, loss: 0.087311, lr: 0.000200
|
341 |
+
2023-08-07 15:31:38,050 - INFO - [Train] step: 132399, loss: 0.097564, lr: 0.000200
|
342 |
+
2023-08-07 15:33:00,183 - INFO - [Train] step: 132799, loss: 0.105457, lr: 0.000200
|
343 |
+
2023-08-07 15:34:22,328 - INFO - [Train] step: 133199, loss: 0.082838, lr: 0.000200
|
344 |
+
2023-08-07 15:35:44,676 - INFO - [Train] step: 133599, loss: 0.124125, lr: 0.000200
|
345 |
+
2023-08-07 15:37:06,618 - INFO - [Train] step: 133999, loss: 0.119264, lr: 0.000200
|
346 |
+
2023-08-07 15:38:28,496 - INFO - [Train] step: 134399, loss: 0.106483, lr: 0.000200
|
347 |
+
2023-08-07 15:39:50,665 - INFO - [Train] step: 134799, loss: 0.108060, lr: 0.000200
|
348 |
+
2023-08-07 15:41:34,928 - INFO - [Train] step: 135199, loss: 0.080738, lr: 0.000200
|
349 |
+
2023-08-07 15:42:56,809 - INFO - [Train] step: 135599, loss: 0.093693, lr: 0.000200
|
350 |
+
2023-08-07 15:44:18,543 - INFO - [Train] step: 135999, loss: 0.121478, lr: 0.000200
|
351 |
+
2023-08-07 15:45:40,902 - INFO - [Train] step: 136399, loss: 0.106023, lr: 0.000200
|
352 |
+
2023-08-07 15:47:03,256 - INFO - [Train] step: 136799, loss: 0.101689, lr: 0.000200
|
353 |
+
2023-08-07 15:48:25,606 - INFO - [Train] step: 137199, loss: 0.112455, lr: 0.000200
|
354 |
+
2023-08-07 15:49:47,114 - INFO - [Train] step: 137599, loss: 0.139700, lr: 0.000200
|
355 |
+
2023-08-07 15:51:08,849 - INFO - [Train] step: 137999, loss: 0.087761, lr: 0.000200
|
356 |
+
2023-08-07 15:52:30,964 - INFO - [Train] step: 138399, loss: 0.159153, lr: 0.000200
|
357 |
+
2023-08-07 15:53:53,280 - INFO - [Train] step: 138799, loss: 0.083335, lr: 0.000200
|
358 |
+
2023-08-07 15:55:15,426 - INFO - [Train] step: 139199, loss: 0.101326, lr: 0.000200
|
359 |
+
2023-08-07 15:56:37,107 - INFO - [Train] step: 139599, loss: 0.127182, lr: 0.000200
|
360 |
+
2023-08-07 15:57:59,075 - INFO - [Train] step: 139999, loss: 0.096868, lr: 0.000200
|
361 |
+
2023-08-07 15:59:43,592 - INFO - [Train] step: 140399, loss: 0.113807, lr: 0.000200
|
362 |
+
2023-08-07 16:01:05,462 - INFO - [Train] step: 140799, loss: 0.093394, lr: 0.000200
|
363 |
+
2023-08-07 16:02:27,530 - INFO - [Train] step: 141199, loss: 0.102109, lr: 0.000200
|
364 |
+
2023-08-07 16:03:49,820 - INFO - [Train] step: 141599, loss: 0.113728, lr: 0.000200
|
365 |
+
2023-08-07 16:05:12,163 - INFO - [Train] step: 141999, loss: 0.130736, lr: 0.000200
|
366 |
+
2023-08-07 16:06:34,274 - INFO - [Train] step: 142399, loss: 0.124864, lr: 0.000200
|
367 |
+
2023-08-07 16:07:56,520 - INFO - [Train] step: 142799, loss: 0.097180, lr: 0.000200
|
368 |
+
2023-08-07 16:09:18,718 - INFO - [Train] step: 143199, loss: 0.115004, lr: 0.000200
|
369 |
+
2023-08-07 16:10:40,848 - INFO - [Train] step: 143599, loss: 0.113431, lr: 0.000200
|
370 |
+
2023-08-07 16:12:02,837 - INFO - [Train] step: 143999, loss: 0.136714, lr: 0.000200
|
371 |
+
2023-08-07 16:13:25,091 - INFO - [Train] step: 144399, loss: 0.104968, lr: 0.000200
|
372 |
+
2023-08-07 16:14:47,021 - INFO - [Train] step: 144799, loss: 0.129130, lr: 0.000200
|
373 |
+
2023-08-07 16:16:30,050 - INFO - [Train] step: 145199, loss: 0.090315, lr: 0.000200
|
374 |
+
2023-08-07 16:17:52,154 - INFO - [Train] step: 145599, loss: 0.094249, lr: 0.000200
|
375 |
+
2023-08-07 16:19:14,528 - INFO - [Train] step: 145999, loss: 0.085897, lr: 0.000200
|
376 |
+
2023-08-07 16:20:36,643 - INFO - [Train] step: 146399, loss: 0.135044, lr: 0.000200
|
377 |
+
2023-08-07 16:21:58,717 - INFO - [Train] step: 146799, loss: 0.162018, lr: 0.000200
|
378 |
+
2023-08-07 16:23:20,478 - INFO - [Train] step: 147199, loss: 0.157382, lr: 0.000200
|
379 |
+
2023-08-07 16:24:42,399 - INFO - [Train] step: 147599, loss: 0.096387, lr: 0.000200
|
380 |
+
2023-08-07 16:26:04,108 - INFO - [Train] step: 147999, loss: 0.146949, lr: 0.000200
|
381 |
+
2023-08-07 16:27:26,450 - INFO - [Train] step: 148399, loss: 0.117472, lr: 0.000200
|
382 |
+
2023-08-07 16:28:48,560 - INFO - [Train] step: 148799, loss: 0.073945, lr: 0.000200
|
383 |
+
2023-08-07 16:30:10,900 - INFO - [Train] step: 149199, loss: 0.106740, lr: 0.000200
|
384 |
+
2023-08-07 16:31:32,589 - INFO - [Train] step: 149599, loss: 0.080322, lr: 0.000200
|
385 |
+
2023-08-07 16:32:54,112 - INFO - [Train] step: 149999, loss: 0.136559, lr: 0.000200
|
386 |
+
2023-08-07 16:34:39,170 - INFO - [Train] step: 150399, loss: 0.126293, lr: 0.000200
|
387 |
+
2023-08-07 16:36:01,082 - INFO - [Train] step: 150799, loss: 0.114251, lr: 0.000200
|
388 |
+
2023-08-07 16:37:22,783 - INFO - [Train] step: 151199, loss: 0.096746, lr: 0.000200
|
389 |
+
2023-08-07 16:38:44,927 - INFO - [Train] step: 151599, loss: 0.108269, lr: 0.000200
|
390 |
+
2023-08-07 16:40:06,611 - INFO - [Train] step: 151999, loss: 0.086108, lr: 0.000200
|
391 |
+
2023-08-07 16:41:28,580 - INFO - [Train] step: 152399, loss: 0.112391, lr: 0.000200
|
392 |
+
2023-08-07 16:42:50,904 - INFO - [Train] step: 152799, loss: 0.101756, lr: 0.000200
|
393 |
+
2023-08-07 16:44:13,030 - INFO - [Train] step: 153199, loss: 0.136925, lr: 0.000200
|
394 |
+
2023-08-07 16:45:34,567 - INFO - [Train] step: 153599, loss: 0.077758, lr: 0.000200
|
395 |
+
2023-08-07 16:46:56,040 - INFO - [Train] step: 153999, loss: 0.122852, lr: 0.000200
|
396 |
+
2023-08-07 16:48:18,189 - INFO - [Train] step: 154399, loss: 0.087679, lr: 0.000200
|
397 |
+
2023-08-07 16:49:40,284 - INFO - [Train] step: 154799, loss: 0.094326, lr: 0.000200
|
398 |
+
2023-08-07 16:51:24,907 - INFO - [Train] step: 155199, loss: 0.128245, lr: 0.000200
|
399 |
+
2023-08-07 16:52:47,084 - INFO - [Train] step: 155599, loss: 0.123441, lr: 0.000200
|
400 |
+
2023-08-07 16:54:09,313 - INFO - [Train] step: 155999, loss: 0.129279, lr: 0.000200
|
401 |
+
2023-08-07 16:55:31,350 - INFO - [Train] step: 156399, loss: 0.132648, lr: 0.000200
|
402 |
+
2023-08-07 16:56:53,699 - INFO - [Train] step: 156799, loss: 0.138483, lr: 0.000200
|
403 |
+
2023-08-07 16:58:15,574 - INFO - [Train] step: 157199, loss: 0.082590, lr: 0.000200
|
404 |
+
2023-08-07 16:59:37,897 - INFO - [Train] step: 157599, loss: 0.108296, lr: 0.000200
|
405 |
+
2023-08-07 17:01:00,210 - INFO - [Train] step: 157999, loss: 0.116951, lr: 0.000200
|
406 |
+
2023-08-07 17:02:22,364 - INFO - [Train] step: 158399, loss: 0.103571, lr: 0.000200
|
407 |
+
2023-08-07 17:03:44,491 - INFO - [Train] step: 158799, loss: 0.096296, lr: 0.000200
|
408 |
+
2023-08-07 17:05:06,286 - INFO - [Train] step: 159199, loss: 0.123831, lr: 0.000200
|
409 |
+
2023-08-07 17:06:28,540 - INFO - [Train] step: 159599, loss: 0.100907, lr: 0.000200
|
410 |
+
2023-08-07 17:07:50,459 - INFO - [Train] step: 159999, loss: 0.135853, lr: 0.000200
|
411 |
+
2023-08-07 17:09:34,690 - INFO - [Train] step: 160399, loss: 0.101097, lr: 0.000200
|
412 |
+
2023-08-07 17:10:57,052 - INFO - [Train] step: 160799, loss: 0.121195, lr: 0.000200
|
413 |
+
2023-08-07 17:12:19,350 - INFO - [Train] step: 161199, loss: 0.102940, lr: 0.000200
|
414 |
+
2023-08-07 17:13:41,486 - INFO - [Train] step: 161599, loss: 0.102387, lr: 0.000200
|
415 |
+
2023-08-07 17:15:02,802 - INFO - [Train] step: 161999, loss: 0.124292, lr: 0.000200
|
416 |
+
2023-08-07 17:16:25,109 - INFO - [Train] step: 162399, loss: 0.115599, lr: 0.000200
|
417 |
+
2023-08-07 17:17:47,453 - INFO - [Train] step: 162799, loss: 0.142791, lr: 0.000200
|
418 |
+
2023-08-07 17:19:09,598 - INFO - [Train] step: 163199, loss: 0.090653, lr: 0.000200
|
419 |
+
2023-08-07 17:20:31,786 - INFO - [Train] step: 163599, loss: 0.110355, lr: 0.000200
|
420 |
+
2023-08-07 17:21:54,047 - INFO - [Train] step: 163999, loss: 0.131236, lr: 0.000200
|
421 |
+
2023-08-07 17:23:16,339 - INFO - [Train] step: 164399, loss: 0.153339, lr: 0.000200
|
422 |
+
2023-08-07 17:24:38,703 - INFO - [Train] step: 164799, loss: 0.109122, lr: 0.000200
|
423 |
+
2023-08-07 17:26:22,327 - INFO - [Train] step: 165199, loss: 0.088501, lr: 0.000200
|
424 |
+
2023-08-07 17:27:44,041 - INFO - [Train] step: 165599, loss: 0.105226, lr: 0.000200
|
425 |
+
2023-08-07 17:29:06,114 - INFO - [Train] step: 165999, loss: 0.141148, lr: 0.000200
|
426 |
+
2023-08-07 17:30:28,259 - INFO - [Train] step: 166399, loss: 0.092174, lr: 0.000200
|
427 |
+
2023-08-07 17:31:50,574 - INFO - [Train] step: 166799, loss: 0.123589, lr: 0.000200
|
428 |
+
2023-08-07 17:33:12,117 - INFO - [Train] step: 167199, loss: 0.117268, lr: 0.000200
|
429 |
+
2023-08-07 17:34:33,830 - INFO - [Train] step: 167599, loss: 0.096447, lr: 0.000200
|
430 |
+
2023-08-07 17:35:56,200 - INFO - [Train] step: 167999, loss: 0.132877, lr: 0.000200
|
431 |
+
2023-08-07 17:37:18,305 - INFO - [Train] step: 168399, loss: 0.117735, lr: 0.000200
|
432 |
+
2023-08-07 17:38:40,655 - INFO - [Train] step: 168799, loss: 0.079842, lr: 0.000200
|
433 |
+
2023-08-07 17:40:02,991 - INFO - [Train] step: 169199, loss: 0.090230, lr: 0.000200
|
434 |
+
2023-08-07 17:41:24,898 - INFO - [Train] step: 169599, loss: 0.099085, lr: 0.000200
|
435 |
+
2023-08-07 17:42:47,007 - INFO - [Train] step: 169999, loss: 0.118393, lr: 0.000200
|
436 |
+
2023-08-07 17:44:32,501 - INFO - [Train] step: 170399, loss: 0.149631, lr: 0.000200
|
437 |
+
2023-08-07 17:45:54,845 - INFO - [Train] step: 170799, loss: 0.094728, lr: 0.000200
|
438 |
+
2023-08-07 17:47:16,557 - INFO - [Train] step: 171199, loss: 0.113340, lr: 0.000200
|
439 |
+
2023-08-07 17:48:38,365 - INFO - [Train] step: 171599, loss: 0.159166, lr: 0.000200
|
440 |
+
2023-08-07 17:50:00,848 - INFO - [Train] step: 171999, loss: 0.122926, lr: 0.000200
|
441 |
+
2023-08-07 17:51:22,717 - INFO - [Train] step: 172399, loss: 0.095736, lr: 0.000200
|
442 |
+
2023-08-07 17:52:44,864 - INFO - [Train] step: 172799, loss: 0.130558, lr: 0.000200
|
443 |
+
2023-08-07 17:54:07,182 - INFO - [Train] step: 173199, loss: 0.121787, lr: 0.000200
|
444 |
+
2023-08-07 17:55:28,680 - INFO - [Train] step: 173599, loss: 0.103055, lr: 0.000200
|
445 |
+
2023-08-07 17:56:50,465 - INFO - [Train] step: 173999, loss: 0.102579, lr: 0.000200
|
446 |
+
2023-08-07 17:58:12,732 - INFO - [Train] step: 174399, loss: 0.126776, lr: 0.000200
|
447 |
+
2023-08-07 17:59:34,449 - INFO - [Train] step: 174799, loss: 0.117643, lr: 0.000200
|
448 |
+
2023-08-07 18:01:18,510 - INFO - [Train] step: 175199, loss: 0.095255, lr: 0.000200
|
449 |
+
2023-08-07 18:02:40,832 - INFO - [Train] step: 175599, loss: 0.103943, lr: 0.000200
|
450 |
+
2023-08-07 18:04:02,954 - INFO - [Train] step: 175999, loss: 0.114158, lr: 0.000200
|
451 |
+
2023-08-07 18:05:25,091 - INFO - [Train] step: 176399, loss: 0.111492, lr: 0.000200
|
452 |
+
2023-08-07 18:06:46,954 - INFO - [Train] step: 176799, loss: 0.121293, lr: 0.000200
|
453 |
+
2023-08-07 18:08:08,711 - INFO - [Train] step: 177199, loss: 0.090264, lr: 0.000200
|
454 |
+
2023-08-07 18:09:30,853 - INFO - [Train] step: 177599, loss: 0.100538, lr: 0.000200
|
455 |
+
2023-08-07 18:10:52,840 - INFO - [Train] step: 177999, loss: 0.099849, lr: 0.000200
|
456 |
+
2023-08-07 18:12:14,466 - INFO - [Train] step: 178399, loss: 0.107672, lr: 0.000200
|
457 |
+
2023-08-07 18:13:36,154 - INFO - [Train] step: 178799, loss: 0.144601, lr: 0.000200
|
458 |
+
2023-08-07 18:14:58,294 - INFO - [Train] step: 179199, loss: 0.100732, lr: 0.000200
|
459 |
+
2023-08-07 18:16:20,225 - INFO - [Train] step: 179599, loss: 0.092496, lr: 0.000200
|
460 |
+
2023-08-07 18:17:42,326 - INFO - [Train] step: 179999, loss: 0.111023, lr: 0.000200
|
461 |
+
2023-08-07 18:19:26,804 - INFO - [Train] step: 180399, loss: 0.148467, lr: 0.000200
|
462 |
+
2023-08-07 18:20:49,109 - INFO - [Train] step: 180799, loss: 0.092455, lr: 0.000200
|
463 |
+
2023-08-07 18:22:11,032 - INFO - [Train] step: 181199, loss: 0.102241, lr: 0.000200
|
464 |
+
2023-08-07 18:23:32,939 - INFO - [Train] step: 181599, loss: 0.108256, lr: 0.000200
|
465 |
+
2023-08-07 18:24:55,283 - INFO - [Train] step: 181999, loss: 0.094663, lr: 0.000200
|
466 |
+
2023-08-07 18:26:17,390 - INFO - [Train] step: 182399, loss: 0.108905, lr: 0.000200
|
467 |
+
2023-08-07 18:27:39,101 - INFO - [Train] step: 182799, loss: 0.129013, lr: 0.000200
|
468 |
+
2023-08-07 18:29:01,238 - INFO - [Train] step: 183199, loss: 0.124781, lr: 0.000200
|
469 |
+
2023-08-07 18:30:22,754 - INFO - [Train] step: 183599, loss: 0.117572, lr: 0.000200
|
470 |
+
2023-08-07 18:31:44,269 - INFO - [Train] step: 183999, loss: 0.105549, lr: 0.000200
|
471 |
+
2023-08-07 18:33:06,548 - INFO - [Train] step: 184399, loss: 0.105690, lr: 0.000200
|
472 |
+
2023-08-07 18:34:28,725 - INFO - [Train] step: 184799, loss: 0.131030, lr: 0.000200
|
473 |
+
2023-08-07 18:36:13,991 - INFO - [Train] step: 185199, loss: 0.116739, lr: 0.000200
|
474 |
+
2023-08-07 18:37:36,333 - INFO - [Train] step: 185599, loss: 0.138103, lr: 0.000200
|
475 |
+
2023-08-07 18:38:57,994 - INFO - [Train] step: 185999, loss: 0.092167, lr: 0.000200
|
476 |
+
2023-08-07 18:40:19,745 - INFO - [Train] step: 186399, loss: 0.137942, lr: 0.000200
|
477 |
+
2023-08-07 18:41:41,866 - INFO - [Train] step: 186799, loss: 0.082331, lr: 0.000200
|
478 |
+
2023-08-07 18:43:03,889 - INFO - [Train] step: 187199, loss: 0.095830, lr: 0.000200
|
479 |
+
2023-08-07 18:44:26,118 - INFO - [Train] step: 187599, loss: 0.100209, lr: 0.000200
|
480 |
+
2023-08-07 18:45:48,474 - INFO - [Train] step: 187999, loss: 0.114014, lr: 0.000200
|
481 |
+
2023-08-07 18:47:10,765 - INFO - [Train] step: 188399, loss: 0.142246, lr: 0.000200
|
482 |
+
2023-08-07 18:48:32,925 - INFO - [Train] step: 188799, loss: 0.106054, lr: 0.000200
|
483 |
+
2023-08-07 18:49:55,076 - INFO - [Train] step: 189199, loss: 0.111874, lr: 0.000200
|
484 |
+
2023-08-07 18:51:16,942 - INFO - [Train] step: 189599, loss: 0.101729, lr: 0.000200
|
485 |
+
2023-08-07 18:52:38,308 - INFO - [Train] step: 189999, loss: 0.122955, lr: 0.000200
|
486 |
+
2023-08-07 18:54:22,743 - INFO - [Train] step: 190399, loss: 0.085192, lr: 0.000200
|
487 |
+
2023-08-07 18:55:44,874 - INFO - [Train] step: 190799, loss: 0.088300, lr: 0.000200
|
488 |
+
2023-08-07 18:57:06,561 - INFO - [Train] step: 191199, loss: 0.127417, lr: 0.000200
|
489 |
+
2023-08-07 18:58:28,136 - INFO - [Train] step: 191599, loss: 0.123584, lr: 0.000200
|
490 |
+
2023-08-07 18:59:50,426 - INFO - [Train] step: 191999, loss: 0.127683, lr: 0.000200
|
491 |
+
2023-08-07 19:01:11,914 - INFO - [Train] step: 192399, loss: 0.133990, lr: 0.000200
|
492 |
+
2023-08-07 19:02:34,031 - INFO - [Train] step: 192799, loss: 0.120536, lr: 0.000200
|
493 |
+
2023-08-07 19:03:56,193 - INFO - [Train] step: 193199, loss: 0.083013, lr: 0.000200
|
494 |
+
2023-08-07 19:05:17,503 - INFO - [Train] step: 193599, loss: 0.119776, lr: 0.000200
|
495 |
+
2023-08-07 19:06:39,407 - INFO - [Train] step: 193999, loss: 0.103435, lr: 0.000200
|
496 |
+
2023-08-07 19:08:00,508 - INFO - [Train] step: 194399, loss: 0.118171, lr: 0.000200
|
497 |
+
2023-08-07 19:09:22,647 - INFO - [Train] step: 194799, loss: 0.097865, lr: 0.000200
|
498 |
+
2023-08-07 19:11:06,494 - INFO - [Train] step: 195199, loss: 0.128009, lr: 0.000200
|
499 |
+
2023-08-07 19:12:27,998 - INFO - [Train] step: 195599, loss: 0.093132, lr: 0.000200
|
500 |
+
2023-08-07 19:13:50,097 - INFO - [Train] step: 195999, loss: 0.092939, lr: 0.000200
|
501 |
+
2023-08-07 19:15:11,619 - INFO - [Train] step: 196399, loss: 0.074518, lr: 0.000200
|
502 |
+
2023-08-07 19:16:33,364 - INFO - [Train] step: 196799, loss: 0.127847, lr: 0.000200
|
503 |
+
2023-08-07 19:17:54,869 - INFO - [Train] step: 197199, loss: 0.126053, lr: 0.000200
|
504 |
+
2023-08-07 19:19:17,192 - INFO - [Train] step: 197599, loss: 0.104337, lr: 0.000200
|
505 |
+
2023-08-07 19:20:39,328 - INFO - [Train] step: 197999, loss: 0.110391, lr: 0.000200
|
506 |
+
2023-08-07 19:22:00,952 - INFO - [Train] step: 198399, loss: 0.090200, lr: 0.000200
|
507 |
+
2023-08-07 19:23:22,319 - INFO - [Train] step: 198799, loss: 0.098856, lr: 0.000200
|
508 |
+
2023-08-07 19:24:44,281 - INFO - [Train] step: 199199, loss: 0.121909, lr: 0.000200
|
509 |
+
2023-08-07 19:26:06,583 - INFO - [Train] step: 199599, loss: 0.100613, lr: 0.000200
|
510 |
+
2023-08-07 19:27:28,736 - INFO - [Train] step: 199999, loss: 0.099918, lr: 0.000200
|
511 |
+
2023-08-07 19:29:11,578 - INFO - [Train] step: 200399, loss: 0.093422, lr: 0.000200
|
512 |
+
2023-08-07 19:30:33,270 - INFO - [Train] step: 200799, loss: 0.113539, lr: 0.000200
|
513 |
+
2023-08-07 19:31:54,567 - INFO - [Train] step: 201199, loss: 0.106162, lr: 0.000200
|
514 |
+
2023-08-07 19:33:16,881 - INFO - [Train] step: 201599, loss: 0.122963, lr: 0.000200
|
515 |
+
2023-08-07 19:34:39,231 - INFO - [Train] step: 201999, loss: 0.103843, lr: 0.000200
|
516 |
+
2023-08-07 19:36:01,156 - INFO - [Train] step: 202399, loss: 0.112713, lr: 0.000200
|
517 |
+
2023-08-07 19:37:23,580 - INFO - [Train] step: 202799, loss: 0.137966, lr: 0.000200
|
518 |
+
2023-08-07 19:38:45,798 - INFO - [Train] step: 203199, loss: 0.116644, lr: 0.000200
|
519 |
+
2023-08-07 19:40:07,701 - INFO - [Train] step: 203599, loss: 0.102900, lr: 0.000200
|
520 |
+
2023-08-07 19:41:29,629 - INFO - [Train] step: 203999, loss: 0.122038, lr: 0.000200
|
521 |
+
2023-08-07 19:42:51,963 - INFO - [Train] step: 204399, loss: 0.102956, lr: 0.000200
|
522 |
+
2023-08-07 19:44:14,327 - INFO - [Train] step: 204799, loss: 0.093602, lr: 0.000200
|
523 |
+
2023-08-07 19:45:56,730 - INFO - [Train] step: 205199, loss: 0.108089, lr: 0.000200
|
524 |
+
2023-08-07 19:47:18,833 - INFO - [Train] step: 205599, loss: 0.126664, lr: 0.000200
|
525 |
+
2023-08-07 19:48:40,336 - INFO - [Train] step: 205999, loss: 0.123856, lr: 0.000200
|
526 |
+
2023-08-07 19:50:01,843 - INFO - [Train] step: 206399, loss: 0.121010, lr: 0.000200
|
527 |
+
2023-08-07 19:51:23,989 - INFO - [Train] step: 206799, loss: 0.130720, lr: 0.000200
|
528 |
+
2023-08-07 19:52:46,327 - INFO - [Train] step: 207199, loss: 0.127612, lr: 0.000200
|
529 |
+
2023-08-07 19:54:07,464 - INFO - [Train] step: 207599, loss: 0.093583, lr: 0.000200
|
530 |
+
2023-08-07 19:55:29,517 - INFO - [Train] step: 207999, loss: 0.128115, lr: 0.000200
|
531 |
+
2023-08-07 19:56:51,687 - INFO - [Train] step: 208399, loss: 0.129089, lr: 0.000200
|
532 |
+
2023-08-07 19:58:13,715 - INFO - [Train] step: 208799, loss: 0.110647, lr: 0.000200
|
533 |
+
2023-08-07 19:59:35,504 - INFO - [Train] step: 209199, loss: 0.111910, lr: 0.000200
|
534 |
+
2023-08-07 20:00:57,345 - INFO - [Train] step: 209599, loss: 0.103247, lr: 0.000200
|
535 |
+
2023-08-07 20:02:19,349 - INFO - [Train] step: 209999, loss: 0.134336, lr: 0.000200
|
536 |
+
2023-08-07 20:04:03,389 - INFO - [Train] step: 210399, loss: 0.091911, lr: 0.000200
|
537 |
+
2023-08-07 20:05:25,082 - INFO - [Train] step: 210799, loss: 0.107047, lr: 0.000200
|
538 |
+
2023-08-07 20:06:47,209 - INFO - [Train] step: 211199, loss: 0.117183, lr: 0.000200
|
539 |
+
2023-08-07 20:08:09,135 - INFO - [Train] step: 211599, loss: 0.118698, lr: 0.000200
|
540 |
+
2023-08-07 20:09:31,083 - INFO - [Train] step: 211999, loss: 0.113967, lr: 0.000200
|
541 |
+
2023-08-07 20:10:53,410 - INFO - [Train] step: 212399, loss: 0.104207, lr: 0.000200
|
542 |
+
2023-08-07 20:12:14,927 - INFO - [Train] step: 212799, loss: 0.097402, lr: 0.000200
|
543 |
+
2023-08-07 20:13:36,608 - INFO - [Train] step: 213199, loss: 0.119846, lr: 0.000200
|
544 |
+
2023-08-07 20:14:58,968 - INFO - [Train] step: 213599, loss: 0.083921, lr: 0.000200
|
545 |
+
2023-08-07 20:16:21,097 - INFO - [Train] step: 213999, loss: 0.138865, lr: 0.000200
|
546 |
+
2023-08-07 20:17:43,389 - INFO - [Train] step: 214399, loss: 0.099082, lr: 0.000200
|
547 |
+
2023-08-07 20:19:05,557 - INFO - [Train] step: 214799, loss: 0.108562, lr: 0.000200
|
548 |
+
2023-08-07 20:20:48,947 - INFO - [Train] step: 215199, loss: 0.100820, lr: 0.000200
|
549 |
+
2023-08-07 20:22:11,080 - INFO - [Train] step: 215599, loss: 0.099039, lr: 0.000200
|
550 |
+
2023-08-07 20:23:33,423 - INFO - [Train] step: 215999, loss: 0.114374, lr: 0.000200
|
551 |
+
2023-08-07 20:24:55,531 - INFO - [Train] step: 216399, loss: 0.105309, lr: 0.000200
|
552 |
+
2023-08-07 20:26:17,462 - INFO - [Train] step: 216799, loss: 0.153880, lr: 0.000200
|
553 |
+
2023-08-07 20:27:39,809 - INFO - [Train] step: 217199, loss: 0.109010, lr: 0.000200
|
554 |
+
2023-08-07 20:29:01,739 - INFO - [Train] step: 217599, loss: 0.119924, lr: 0.000200
|
555 |
+
2023-08-07 20:30:23,449 - INFO - [Train] step: 217999, loss: 0.123794, lr: 0.000200
|
556 |
+
2023-08-07 20:31:45,893 - INFO - [Train] step: 218399, loss: 0.113580, lr: 0.000200
|
557 |
+
2023-08-07 20:33:08,280 - INFO - [Train] step: 218799, loss: 0.132963, lr: 0.000200
|
558 |
+
2023-08-07 20:34:30,164 - INFO - [Train] step: 219199, loss: 0.122822, lr: 0.000200
|
559 |
+
2023-08-07 20:35:51,602 - INFO - [Train] step: 219599, loss: 0.109602, lr: 0.000200
|
560 |
+
2023-08-07 20:37:13,259 - INFO - [Train] step: 219999, loss: 0.091663, lr: 0.000200
|
561 |
+
2023-08-07 20:38:58,950 - INFO - [Train] step: 220399, loss: 0.106912, lr: 0.000200
|
562 |
+
2023-08-07 20:40:20,868 - INFO - [Train] step: 220799, loss: 0.117222, lr: 0.000200
|
563 |
+
2023-08-07 20:41:42,586 - INFO - [Train] step: 221199, loss: 0.135882, lr: 0.000200
|
564 |
+
2023-08-07 20:43:04,673 - INFO - [Train] step: 221599, loss: 0.085260, lr: 0.000200
|
565 |
+
2023-08-07 20:44:26,840 - INFO - [Train] step: 221999, loss: 0.076068, lr: 0.000200
|
566 |
+
2023-08-07 20:45:48,968 - INFO - [Train] step: 222399, loss: 0.104925, lr: 0.000200
|
567 |
+
2023-08-07 20:47:10,725 - INFO - [Train] step: 222799, loss: 0.104246, lr: 0.000200
|
568 |
+
2023-08-07 20:48:32,799 - INFO - [Train] step: 223199, loss: 0.114426, lr: 0.000200
|
569 |
+
2023-08-07 20:49:54,694 - INFO - [Train] step: 223599, loss: 0.103470, lr: 0.000200
|
570 |
+
2023-08-07 20:51:16,360 - INFO - [Train] step: 223999, loss: 0.105722, lr: 0.000200
|
571 |
+
2023-08-07 20:52:38,560 - INFO - [Train] step: 224399, loss: 0.134950, lr: 0.000200
|
572 |
+
2023-08-07 20:54:00,689 - INFO - [Train] step: 224799, loss: 0.088887, lr: 0.000200
|
573 |
+
2023-08-07 20:55:45,135 - INFO - [Train] step: 225199, loss: 0.098573, lr: 0.000200
|
574 |
+
2023-08-07 20:57:07,274 - INFO - [Train] step: 225599, loss: 0.118965, lr: 0.000200
|
575 |
+
2023-08-07 20:58:29,578 - INFO - [Train] step: 225999, loss: 0.137455, lr: 0.000200
|
576 |
+
2023-08-07 20:59:51,682 - INFO - [Train] step: 226399, loss: 0.103509, lr: 0.000200
|
577 |
+
2023-08-07 21:01:13,438 - INFO - [Train] step: 226799, loss: 0.131616, lr: 0.000200
|
578 |
+
2023-08-07 21:02:35,784 - INFO - [Train] step: 227199, loss: 0.089393, lr: 0.000200
|
579 |
+
2023-08-07 21:03:57,704 - INFO - [Train] step: 227599, loss: 0.100664, lr: 0.000200
|
580 |
+
2023-08-07 21:05:20,008 - INFO - [Train] step: 227999, loss: 0.097019, lr: 0.000200
|
581 |
+
2023-08-07 21:06:42,128 - INFO - [Train] step: 228399, loss: 0.125824, lr: 0.000200
|
582 |
+
2023-08-07 21:08:04,245 - INFO - [Train] step: 228799, loss: 0.074508, lr: 0.000200
|
583 |
+
2023-08-07 21:09:26,273 - INFO - [Train] step: 229199, loss: 0.128657, lr: 0.000200
|
584 |
+
2023-08-07 21:10:48,094 - INFO - [Train] step: 229599, loss: 0.119606, lr: 0.000200
|
585 |
+
2023-08-07 21:12:09,832 - INFO - [Train] step: 229999, loss: 0.106112, lr: 0.000200
|
586 |
+
2023-08-07 21:13:54,265 - INFO - [Train] step: 230399, loss: 0.120219, lr: 0.000200
|
587 |
+
2023-08-07 21:15:16,394 - INFO - [Train] step: 230799, loss: 0.130889, lr: 0.000200
|
588 |
+
2023-08-07 21:16:38,117 - INFO - [Train] step: 231199, loss: 0.148711, lr: 0.000200
|
589 |
+
2023-08-07 21:18:00,241 - INFO - [Train] step: 231599, loss: 0.104815, lr: 0.000200
|
590 |
+
2023-08-07 21:19:22,542 - INFO - [Train] step: 231999, loss: 0.089949, lr: 0.000200
|
591 |
+
2023-08-07 21:20:44,895 - INFO - [Train] step: 232399, loss: 0.125149, lr: 0.000200
|
592 |
+
2023-08-07 21:22:06,860 - INFO - [Train] step: 232799, loss: 0.092998, lr: 0.000200
|
593 |
+
2023-08-07 21:23:29,126 - INFO - [Train] step: 233199, loss: 0.140286, lr: 0.000200
|
594 |
+
2023-08-07 21:24:51,483 - INFO - [Train] step: 233599, loss: 0.109981, lr: 0.000200
|
595 |
+
2023-08-07 21:26:13,503 - INFO - [Train] step: 233999, loss: 0.098034, lr: 0.000200
|
596 |
+
2023-08-07 21:27:35,301 - INFO - [Train] step: 234399, loss: 0.092793, lr: 0.000200
|
597 |
+
2023-08-07 21:28:57,647 - INFO - [Train] step: 234799, loss: 0.141544, lr: 0.000200
|
598 |
+
2023-08-07 21:30:43,114 - INFO - [Train] step: 235199, loss: 0.090412, lr: 0.000200
|
599 |
+
2023-08-07 21:32:04,815 - INFO - [Train] step: 235599, loss: 0.123882, lr: 0.000200
|
600 |
+
2023-08-07 21:33:26,759 - INFO - [Train] step: 235999, loss: 0.127474, lr: 0.000200
|
601 |
+
2023-08-07 21:34:48,698 - INFO - [Train] step: 236399, loss: 0.125369, lr: 0.000200
|
602 |
+
2023-08-07 21:36:10,611 - INFO - [Train] step: 236799, loss: 0.116213, lr: 0.000200
|
603 |
+
2023-08-07 21:37:32,274 - INFO - [Train] step: 237199, loss: 0.106958, lr: 0.000200
|
604 |
+
2023-08-07 21:38:54,239 - INFO - [Train] step: 237599, loss: 0.095185, lr: 0.000200
|
605 |
+
2023-08-07 21:40:16,363 - INFO - [Train] step: 237999, loss: 0.101106, lr: 0.000200
|
606 |
+
2023-08-07 21:41:38,677 - INFO - [Train] step: 238399, loss: 0.123637, lr: 0.000200
|
607 |
+
2023-08-07 21:43:00,614 - INFO - [Train] step: 238799, loss: 0.110368, lr: 0.000200
|
608 |
+
2023-08-07 21:44:22,741 - INFO - [Train] step: 239199, loss: 0.131181, lr: 0.000200
|
609 |
+
2023-08-07 21:45:45,052 - INFO - [Train] step: 239599, loss: 0.084828, lr: 0.000200
|
610 |
+
2023-08-07 21:47:07,369 - INFO - [Train] step: 239999, loss: 0.119953, lr: 0.000200
|
611 |
+
2023-08-07 21:48:50,640 - INFO - [Train] step: 240399, loss: 0.139034, lr: 0.000200
|
612 |
+
2023-08-07 21:50:12,525 - INFO - [Train] step: 240799, loss: 0.103394, lr: 0.000200
|
613 |
+
2023-08-07 21:51:34,447 - INFO - [Train] step: 241199, loss: 0.124396, lr: 0.000200
|
614 |
+
2023-08-07 21:52:56,596 - INFO - [Train] step: 241599, loss: 0.096993, lr: 0.000200
|
615 |
+
2023-08-07 21:54:18,702 - INFO - [Train] step: 241999, loss: 0.116652, lr: 0.000200
|
616 |
+
2023-08-07 21:55:40,813 - INFO - [Train] step: 242399, loss: 0.117313, lr: 0.000200
|
617 |
+
2023-08-07 21:57:03,137 - INFO - [Train] step: 242799, loss: 0.095886, lr: 0.000200
|
618 |
+
2023-08-07 21:58:25,304 - INFO - [Train] step: 243199, loss: 0.120149, lr: 0.000200
|
619 |
+
2023-08-07 21:59:47,213 - INFO - [Train] step: 243599, loss: 0.125677, lr: 0.000200
|
620 |
+
2023-08-07 22:01:09,501 - INFO - [Train] step: 243999, loss: 0.073386, lr: 0.000200
|
621 |
+
2023-08-07 22:02:31,663 - INFO - [Train] step: 244399, loss: 0.084520, lr: 0.000200
|
622 |
+
2023-08-07 22:03:53,803 - INFO - [Train] step: 244799, loss: 0.068746, lr: 0.000200
|
623 |
+
2023-08-07 22:05:37,267 - INFO - [Train] step: 245199, loss: 0.096776, lr: 0.000200
|
624 |
+
2023-08-07 22:06:59,584 - INFO - [Train] step: 245599, loss: 0.123576, lr: 0.000200
|
625 |
+
2023-08-07 22:08:22,091 - INFO - [Train] step: 245999, loss: 0.060448, lr: 0.000200
|
626 |
+
2023-08-07 22:09:43,788 - INFO - [Train] step: 246399, loss: 0.077442, lr: 0.000200
|
627 |
+
2023-08-07 22:11:06,110 - INFO - [Train] step: 246799, loss: 0.117696, lr: 0.000200
|
628 |
+
2023-08-07 22:12:28,439 - INFO - [Train] step: 247199, loss: 0.122948, lr: 0.000200
|
629 |
+
2023-08-07 22:13:50,600 - INFO - [Train] step: 247599, loss: 0.141044, lr: 0.000200
|
630 |
+
2023-08-07 22:15:22,620 - INFO - [Train] step: 247999, loss: 0.103831, lr: 0.000200
|
631 |
+
2023-08-07 22:16:55,858 - INFO - [Train] step: 248399, loss: 0.099113, lr: 0.000200
|
632 |
+
2023-08-07 22:18:25,039 - INFO - [Train] step: 248799, loss: 0.116123, lr: 0.000200
|
633 |
+
2023-08-07 22:19:52,921 - INFO - [Train] step: 249199, loss: 0.101775, lr: 0.000200
|
634 |
+
2023-08-07 22:21:17,177 - INFO - [Train] step: 249599, loss: 0.119279, lr: 0.000200
|
635 |
+
2023-08-07 22:22:39,788 - INFO - [Train] step: 249999, loss: 0.091824, lr: 0.000200
|
636 |
+
2023-08-07 22:24:23,842 - INFO - [Train] step: 250399, loss: 0.105654, lr: 0.000200
|
637 |
+
2023-08-07 22:25:46,051 - INFO - [Train] step: 250799, loss: 0.128992, lr: 0.000200
|
638 |
+
2023-08-07 22:27:07,268 - INFO - [Train] step: 251199, loss: 0.141036, lr: 0.000200
|
639 |
+
2023-08-07 22:28:29,590 - INFO - [Train] step: 251599, loss: 0.170446, lr: 0.000200
|
640 |
+
2023-08-07 22:29:51,909 - INFO - [Train] step: 251999, loss: 0.108987, lr: 0.000200
|
641 |
+
2023-08-07 22:31:14,538 - INFO - [Train] step: 252399, loss: 0.100798, lr: 0.000200
|
642 |
+
2023-08-07 22:32:45,922 - INFO - [Train] step: 252799, loss: 0.118228, lr: 0.000200
|
643 |
+
2023-08-07 22:34:19,412 - INFO - [Train] step: 253199, loss: 0.122678, lr: 0.000200
|
644 |
+
2023-08-07 22:35:49,710 - INFO - [Train] step: 253599, loss: 0.101707, lr: 0.000200
|
645 |
+
2023-08-07 22:37:16,538 - INFO - [Train] step: 253999, loss: 0.096021, lr: 0.000200
|
646 |
+
2023-08-07 22:38:51,205 - INFO - [Train] step: 254399, loss: 0.070691, lr: 0.000200
|
647 |
+
2023-08-07 22:40:25,431 - INFO - [Train] step: 254799, loss: 0.098387, lr: 0.000200
|
648 |
+
2023-08-07 22:42:12,485 - INFO - [Train] step: 255199, loss: 0.111114, lr: 0.000200
|
649 |
+
2023-08-07 22:43:35,101 - INFO - [Train] step: 255599, loss: 0.087545, lr: 0.000200
|
650 |
+
2023-08-07 22:44:58,170 - INFO - [Train] step: 255999, loss: 0.092697, lr: 0.000200
|
651 |
+
2023-08-07 22:46:21,719 - INFO - [Train] step: 256399, loss: 0.096473, lr: 0.000200
|
652 |
+
2023-08-07 22:47:43,659 - INFO - [Train] step: 256799, loss: 0.122011, lr: 0.000200
|
653 |
+
2023-08-07 22:49:05,756 - INFO - [Train] step: 257199, loss: 0.119900, lr: 0.000200
|
654 |
+
2023-08-07 22:50:27,853 - INFO - [Train] step: 257599, loss: 0.128482, lr: 0.000200
|
655 |
+
2023-08-07 22:51:50,018 - INFO - [Train] step: 257999, loss: 0.116662, lr: 0.000200
|
656 |
+
2023-08-07 22:53:12,353 - INFO - [Train] step: 258399, loss: 0.108821, lr: 0.000200
|
657 |
+
2023-08-07 22:54:34,708 - INFO - [Train] step: 258799, loss: 0.080094, lr: 0.000200
|
658 |
+
2023-08-07 22:55:57,039 - INFO - [Train] step: 259199, loss: 0.106504, lr: 0.000200
|
659 |
+
2023-08-07 22:57:19,150 - INFO - [Train] step: 259599, loss: 0.115011, lr: 0.000200
|
660 |
+
2023-08-07 22:58:41,030 - INFO - [Train] step: 259999, loss: 0.111058, lr: 0.000200
|
661 |
+
2023-08-07 23:00:25,532 - INFO - [Train] step: 260399, loss: 0.113337, lr: 0.000200
|
662 |
+
2023-08-07 23:01:48,051 - INFO - [Train] step: 260799, loss: 0.103375, lr: 0.000200
|
663 |
+
2023-08-07 23:03:10,082 - INFO - [Train] step: 261199, loss: 0.112022, lr: 0.000200
|
664 |
+
2023-08-07 23:04:33,845 - INFO - [Train] step: 261599, loss: 0.107698, lr: 0.000200
|
665 |
+
2023-08-07 23:05:55,644 - INFO - [Train] step: 261999, loss: 0.111864, lr: 0.000200
|
666 |
+
2023-08-07 23:07:29,636 - INFO - [Train] step: 262399, loss: 0.104146, lr: 0.000200
|
667 |
+
2023-08-07 23:09:04,637 - INFO - [Train] step: 262799, loss: 0.121387, lr: 0.000200
|
668 |
+
2023-08-07 23:10:39,766 - INFO - [Train] step: 263199, loss: 0.089295, lr: 0.000200
|
669 |
+
2023-08-07 23:12:04,688 - INFO - [Train] step: 263599, loss: 0.125900, lr: 0.000200
|
670 |
+
2023-08-07 23:13:37,358 - INFO - [Train] step: 263999, loss: 0.119247, lr: 0.000200
|
671 |
+
2023-08-07 23:15:11,067 - INFO - [Train] step: 264399, loss: 0.137994, lr: 0.000200
|
672 |
+
2023-08-07 23:16:41,613 - INFO - [Train] step: 264799, loss: 0.123085, lr: 0.000200
|
673 |
+
2023-08-07 23:18:32,807 - INFO - [Train] step: 265199, loss: 0.109656, lr: 0.000200
|
674 |
+
2023-08-07 23:20:07,528 - INFO - [Train] step: 265599, loss: 0.090404, lr: 0.000200
|
675 |
+
2023-08-07 23:21:40,199 - INFO - [Train] step: 265999, loss: 0.120225, lr: 0.000200
|
676 |
+
2023-08-07 23:23:04,960 - INFO - [Train] step: 266399, loss: 0.092175, lr: 0.000200
|
677 |
+
2023-08-07 23:24:39,040 - INFO - [Train] step: 266799, loss: 0.120426, lr: 0.000200
|
678 |
+
2023-08-07 23:26:13,534 - INFO - [Train] step: 267199, loss: 0.137474, lr: 0.000200
|
679 |
+
2023-08-07 23:27:39,611 - INFO - [Train] step: 267599, loss: 0.103276, lr: 0.000200
|
680 |
+
2023-08-07 23:29:11,514 - INFO - [Train] step: 267999, loss: 0.118592, lr: 0.000200
|
681 |
+
2023-08-07 23:30:45,851 - INFO - [Train] step: 268399, loss: 0.128617, lr: 0.000200
|
682 |
+
2023-08-07 23:32:16,690 - INFO - [Train] step: 268799, loss: 0.132897, lr: 0.000200
|
683 |
+
2023-08-07 23:33:43,079 - INFO - [Train] step: 269199, loss: 0.110867, lr: 0.000200
|
684 |
+
2023-08-07 23:35:17,111 - INFO - [Train] step: 269599, loss: 0.133101, lr: 0.000200
|
685 |
+
2023-08-07 23:36:51,197 - INFO - [Train] step: 269999, loss: 0.093577, lr: 0.000200
|
686 |
+
2023-08-07 23:38:38,623 - INFO - [Train] step: 270399, loss: 0.160773, lr: 0.000200
|
687 |
+
2023-08-07 23:40:13,996 - INFO - [Train] step: 270799, loss: 0.104640, lr: 0.000200
|
688 |
+
2023-08-07 23:41:49,296 - INFO - [Train] step: 271199, loss: 0.117079, lr: 0.000200
|
689 |
+
2023-08-07 23:43:22,170 - INFO - [Train] step: 271599, loss: 0.113787, lr: 0.000200
|
690 |
+
2023-08-07 23:44:54,487 - INFO - [Train] step: 271999, loss: 0.124887, lr: 0.000200
|
691 |
+
2023-08-07 23:46:29,620 - INFO - [Train] step: 272399, loss: 0.098357, lr: 0.000200
|
692 |
+
2023-08-07 23:48:04,776 - INFO - [Train] step: 272799, loss: 0.099946, lr: 0.000200
|
693 |
+
2023-08-07 23:49:32,913 - INFO - [Train] step: 273199, loss: 0.129804, lr: 0.000200
|
694 |
+
2023-08-07 23:51:08,185 - INFO - [Train] step: 273599, loss: 0.114058, lr: 0.000200
|
695 |
+
2023-08-07 23:52:44,069 - INFO - [Train] step: 273999, loss: 0.101482, lr: 0.000200
|
696 |
+
2023-08-07 23:54:11,869 - INFO - [Train] step: 274399, loss: 0.103703, lr: 0.000200
|
697 |
+
2023-08-07 23:55:47,246 - INFO - [Train] step: 274799, loss: 0.095251, lr: 0.000200
|
698 |
+
2023-08-07 23:57:44,993 - INFO - [Train] step: 275199, loss: 0.112879, lr: 0.000200
|
699 |
+
2023-08-07 23:59:16,346 - INFO - [Train] step: 275599, loss: 0.100263, lr: 0.000200
|
700 |
+
2023-08-08 00:00:51,678 - INFO - [Train] step: 275999, loss: 0.090636, lr: 0.000200
|
701 |
+
2023-08-08 00:02:27,485 - INFO - [Train] step: 276399, loss: 0.094517, lr: 0.000200
|
702 |
+
2023-08-08 00:04:02,799 - INFO - [Train] step: 276799, loss: 0.140384, lr: 0.000200
|
703 |
+
2023-08-08 00:05:31,864 - INFO - [Train] step: 277199, loss: 0.084483, lr: 0.000200
|
704 |
+
2023-08-08 00:07:07,584 - INFO - [Train] step: 277599, loss: 0.104991, lr: 0.000200
|
705 |
+
2023-08-08 00:08:43,490 - INFO - [Train] step: 277999, loss: 0.080642, lr: 0.000200
|
706 |
+
2023-08-08 00:10:14,199 - INFO - [Train] step: 278399, loss: 0.099485, lr: 0.000200
|
707 |
+
2023-08-08 00:11:44,941 - INFO - [Train] step: 278799, loss: 0.071400, lr: 0.000200
|
708 |
+
2023-08-08 00:13:20,981 - INFO - [Train] step: 279199, loss: 0.108450, lr: 0.000200
|
709 |
+
2023-08-08 00:14:56,311 - INFO - [Train] step: 279599, loss: 0.116792, lr: 0.000200
|
710 |
+
2023-08-08 00:16:26,637 - INFO - [Train] step: 279999, loss: 0.118108, lr: 0.000200
|
711 |
+
2023-08-08 00:18:26,057 - INFO - [Train] step: 280399, loss: 0.087363, lr: 0.000200
|
712 |
+
2023-08-08 00:20:00,966 - INFO - [Train] step: 280799, loss: 0.163901, lr: 0.000200
|
713 |
+
2023-08-08 00:21:34,019 - INFO - [Train] step: 281199, loss: 0.113656, lr: 0.000200
|
714 |
+
2023-08-08 00:23:09,681 - INFO - [Train] step: 281599, loss: 0.111682, lr: 0.000200
|
715 |
+
2023-08-08 00:24:44,518 - INFO - [Train] step: 281999, loss: 0.082123, lr: 0.000200
|
716 |
+
2023-08-08 00:26:20,299 - INFO - [Train] step: 282399, loss: 0.094872, lr: 0.000200
|
717 |
+
2023-08-08 00:27:52,812 - INFO - [Train] step: 282799, loss: 0.118291, lr: 0.000200
|
718 |
+
2023-08-08 00:29:29,009 - INFO - [Train] step: 283199, loss: 0.083886, lr: 0.000200
|
719 |
+
2023-08-08 00:31:05,215 - INFO - [Train] step: 283599, loss: 0.099051, lr: 0.000200
|
720 |
+
2023-08-08 00:32:33,246 - INFO - [Train] step: 283999, loss: 0.118002, lr: 0.000200
|
721 |
+
2023-08-08 00:34:09,361 - INFO - [Train] step: 284399, loss: 0.132477, lr: 0.000200
|
722 |
+
2023-08-08 00:35:44,289 - INFO - [Train] step: 284799, loss: 0.124727, lr: 0.000200
|
723 |
+
2023-08-08 00:37:40,821 - INFO - [Train] step: 285199, loss: 0.114215, lr: 0.000200
|
724 |
+
2023-08-08 00:39:15,164 - INFO - [Train] step: 285599, loss: 0.119803, lr: 0.000200
|
725 |
+
2023-08-08 00:40:50,470 - INFO - [Train] step: 285999, loss: 0.139484, lr: 0.000200
|
726 |
+
2023-08-08 00:42:26,091 - INFO - [Train] step: 286399, loss: 0.129269, lr: 0.000200
|
727 |
+
2023-08-08 00:43:56,441 - INFO - [Train] step: 286799, loss: 0.125098, lr: 0.000200
|
728 |
+
2023-08-08 00:45:31,844 - INFO - [Train] step: 287199, loss: 0.106770, lr: 0.000200
|
729 |
+
2023-08-08 00:47:07,967 - INFO - [Train] step: 287599, loss: 0.121613, lr: 0.000200
|
730 |
+
2023-08-08 00:48:33,806 - INFO - [Train] step: 287999, loss: 0.120980, lr: 0.000200
|
731 |
+
2023-08-08 00:50:09,070 - INFO - [Train] step: 288399, loss: 0.134276, lr: 0.000200
|
732 |
+
2023-08-08 00:51:45,046 - INFO - [Train] step: 288799, loss: 0.082459, lr: 0.000200
|
733 |
+
2023-08-08 00:53:14,335 - INFO - [Train] step: 289199, loss: 0.091214, lr: 0.000200
|
734 |
+
2023-08-08 00:54:49,562 - INFO - [Train] step: 289599, loss: 0.106117, lr: 0.000200
|
735 |
+
2023-08-08 00:56:25,336 - INFO - [Train] step: 289999, loss: 0.102086, lr: 0.000200
|
736 |
+
2023-08-08 00:58:20,444 - INFO - [Train] step: 290399, loss: 0.115763, lr: 0.000200
|
737 |
+
2023-08-08 00:59:54,729 - INFO - [Train] step: 290799, loss: 0.074310, lr: 0.000200
|
738 |
+
2023-08-08 01:01:30,845 - INFO - [Train] step: 291199, loss: 0.070340, lr: 0.000200
|
739 |
+
2023-08-08 01:03:06,663 - INFO - [Train] step: 291599, loss: 0.105800, lr: 0.000200
|
740 |
+
2023-08-08 01:04:39,021 - INFO - [Train] step: 291999, loss: 0.094702, lr: 0.000200
|
741 |
+
2023-08-08 01:06:14,457 - INFO - [Train] step: 292399, loss: 0.106377, lr: 0.000200
|
742 |
+
2023-08-08 01:07:50,208 - INFO - [Train] step: 292799, loss: 0.115168, lr: 0.000200
|
743 |
+
2023-08-08 01:09:23,197 - INFO - [Train] step: 293199, loss: 0.137461, lr: 0.000200
|
744 |
+
2023-08-08 01:10:55,565 - INFO - [Train] step: 293599, loss: 0.140380, lr: 0.000200
|
745 |
+
2023-08-08 01:12:31,067 - INFO - [Train] step: 293999, loss: 0.099768, lr: 0.000200
|
746 |
+
2023-08-08 01:14:05,136 - INFO - [Train] step: 294399, loss: 0.110645, lr: 0.000200
|
747 |
+
2023-08-08 01:15:32,927 - INFO - [Train] step: 294799, loss: 0.119441, lr: 0.000200
|
748 |
+
2023-08-08 01:17:28,985 - INFO - [Train] step: 295199, loss: 0.109740, lr: 0.000200
|
749 |
+
2023-08-08 01:19:05,154 - INFO - [Train] step: 295599, loss: 0.109121, lr: 0.000200
|
750 |
+
2023-08-08 01:20:34,042 - INFO - [Train] step: 295999, loss: 0.153376, lr: 0.000200
|
751 |
+
2023-08-08 01:22:09,280 - INFO - [Train] step: 296399, loss: 0.148037, lr: 0.000200
|
752 |
+
2023-08-08 01:23:45,818 - INFO - [Train] step: 296799, loss: 0.101609, lr: 0.000200
|
753 |
+
2023-08-08 01:25:16,438 - INFO - [Train] step: 297199, loss: 0.083564, lr: 0.000200
|
754 |
+
2023-08-08 01:26:49,770 - INFO - [Train] step: 297599, loss: 0.089507, lr: 0.000200
|
755 |
+
2023-08-08 01:28:25,930 - INFO - [Train] step: 297999, loss: 0.103438, lr: 0.000200
|
756 |
+
2023-08-08 01:30:01,904 - INFO - [Train] step: 298399, loss: 0.120304, lr: 0.000200
|
757 |
+
2023-08-08 01:31:31,271 - INFO - [Train] step: 298799, loss: 0.101671, lr: 0.000200
|
758 |
+
2023-08-08 01:33:06,951 - INFO - [Train] step: 299199, loss: 0.090928, lr: 0.000200
|
759 |
+
2023-08-08 01:34:42,893 - INFO - [Train] step: 299599, loss: 0.096609, lr: 0.000200
|
760 |
+
2023-08-08 01:36:13,167 - INFO - [Train] step: 299999, loss: 0.105133, lr: 0.000200
|
761 |
+
2023-08-08 01:38:10,193 - INFO - [Train] step: 300399, loss: 0.153599, lr: 0.000200
|
762 |
+
2023-08-08 01:39:45,934 - INFO - [Train] step: 300799, loss: 0.099348, lr: 0.000200
|
763 |
+
2023-08-08 01:41:18,613 - INFO - [Train] step: 301199, loss: 0.131756, lr: 0.000200
|
764 |
+
2023-08-08 01:42:52,537 - INFO - [Train] step: 301599, loss: 0.103970, lr: 0.000200
|
765 |
+
2023-08-08 01:44:28,877 - INFO - [Train] step: 301999, loss: 0.101250, lr: 0.000200
|
766 |
+
2023-08-08 01:46:04,245 - INFO - [Train] step: 302399, loss: 0.118983, lr: 0.000200
|
767 |
+
2023-08-08 01:47:37,308 - INFO - [Train] step: 302799, loss: 0.105215, lr: 0.000200
|
768 |
+
2023-08-08 01:49:12,548 - INFO - [Train] step: 303199, loss: 0.086182, lr: 0.000200
|
769 |
+
2023-08-08 01:50:47,913 - INFO - [Train] step: 303599, loss: 0.119117, lr: 0.000200
|
770 |
+
2023-08-08 01:52:22,498 - INFO - [Train] step: 303999, loss: 0.104346, lr: 0.000200
|
771 |
+
2023-08-08 01:53:54,134 - INFO - [Train] step: 304399, loss: 0.112351, lr: 0.000200
|
772 |
+
2023-08-08 01:55:29,212 - INFO - [Train] step: 304799, loss: 0.100717, lr: 0.000200
|
773 |
+
2023-08-08 01:57:25,982 - INFO - [Train] step: 305199, loss: 0.102997, lr: 0.000200
|
774 |
+
2023-08-08 01:58:52,766 - INFO - [Train] step: 305599, loss: 0.091558, lr: 0.000200
|
775 |
+
2023-08-08 02:00:27,824 - INFO - [Train] step: 305999, loss: 0.124631, lr: 0.000200
|
776 |
+
2023-08-08 02:02:04,243 - INFO - [Train] step: 306399, loss: 0.100726, lr: 0.000200
|
777 |
+
2023-08-08 02:03:32,469 - INFO - [Train] step: 306799, loss: 0.119213, lr: 0.000200
|
778 |
+
2023-08-08 02:05:07,745 - INFO - [Train] step: 307199, loss: 0.090630, lr: 0.000200
|
779 |
+
2023-08-08 02:06:44,002 - INFO - [Train] step: 307599, loss: 0.120802, lr: 0.000200
|
780 |
+
2023-08-08 02:08:16,554 - INFO - [Train] step: 307999, loss: 0.124660, lr: 0.000200
|
781 |
+
2023-08-08 02:09:50,417 - INFO - [Train] step: 308399, loss: 0.105882, lr: 0.000200
|
782 |
+
2023-08-08 02:11:26,544 - INFO - [Train] step: 308799, loss: 0.130760, lr: 0.000200
|
783 |
+
2023-08-08 02:13:02,429 - INFO - [Train] step: 309199, loss: 0.117145, lr: 0.000200
|
784 |
+
2023-08-08 02:14:33,895 - INFO - [Train] step: 309599, loss: 0.101169, lr: 0.000200
|
785 |
+
2023-08-08 02:16:09,586 - INFO - [Train] step: 309999, loss: 0.099388, lr: 0.000200
|
786 |
+
2023-08-08 02:18:08,655 - INFO - [Train] step: 310399, loss: 0.126765, lr: 0.000200
|
787 |
+
2023-08-08 02:19:39,624 - INFO - [Train] step: 310799, loss: 0.111021, lr: 0.000200
|
788 |
+
2023-08-08 02:21:11,973 - INFO - [Train] step: 311199, loss: 0.088242, lr: 0.000200
|
789 |
+
2023-08-08 02:22:48,098 - INFO - [Train] step: 311599, loss: 0.123908, lr: 0.000200
|
790 |
+
2023-08-08 02:24:21,602 - INFO - [Train] step: 311999, loss: 0.102665, lr: 0.000200
|
791 |
+
2023-08-08 02:25:52,741 - INFO - [Train] step: 312399, loss: 0.102778, lr: 0.000200
|
792 |
+
2023-08-08 02:27:28,798 - INFO - [Train] step: 312799, loss: 0.149749, lr: 0.000200
|
793 |
+
2023-08-08 02:29:04,047 - INFO - [Train] step: 313199, loss: 0.132170, lr: 0.000200
|
794 |
+
2023-08-08 02:30:35,359 - INFO - [Train] step: 313599, loss: 0.115316, lr: 0.000200
|
795 |
+
2023-08-08 02:32:11,247 - INFO - [Train] step: 313999, loss: 0.149668, lr: 0.000200
|
796 |
+
2023-08-08 02:33:47,158 - INFO - [Train] step: 314399, loss: 0.194232, lr: 0.000200
|
797 |
+
2023-08-08 02:35:22,470 - INFO - [Train] step: 314799, loss: 0.107666, lr: 0.000200
|
798 |
+
2023-08-08 02:37:18,296 - INFO - [Train] step: 315199, loss: 0.127770, lr: 0.000200
|
799 |
+
2023-08-08 02:38:53,857 - INFO - [Train] step: 315599, loss: 0.111345, lr: 0.000200
|
800 |
+
2023-08-08 02:40:29,955 - INFO - [Train] step: 315999, loss: 0.101629, lr: 0.000200
|
801 |
+
2023-08-08 02:42:00,133 - INFO - [Train] step: 316399, loss: 0.113506, lr: 0.000200
|
802 |
+
2023-08-08 02:43:35,836 - INFO - [Train] step: 316799, loss: 0.123914, lr: 0.000200
|
803 |
+
2023-08-08 02:45:11,927 - INFO - [Train] step: 317199, loss: 0.099489, lr: 0.000200
|
804 |
+
2023-08-08 02:46:39,574 - INFO - [Train] step: 317599, loss: 0.145891, lr: 0.000200
|
805 |
+
2023-08-08 02:48:14,756 - INFO - [Train] step: 317999, loss: 0.104523, lr: 0.000200
|
806 |
+
2023-08-08 02:49:50,460 - INFO - [Train] step: 318399, loss: 0.085938, lr: 0.000200
|
807 |
+
2023-08-08 02:51:20,467 - INFO - [Train] step: 318799, loss: 0.133758, lr: 0.000200
|
808 |
+
2023-08-08 02:52:53,229 - INFO - [Train] step: 319199, loss: 0.104609, lr: 0.000200
|
809 |
+
2023-08-08 02:54:29,308 - INFO - [Train] step: 319599, loss: 0.105176, lr: 0.000200
|
810 |
+
2023-08-08 02:56:04,610 - INFO - [Train] step: 319999, loss: 0.126959, lr: 0.000200
|
811 |
+
2023-08-08 02:58:02,211 - INFO - [Train] step: 320399, loss: 0.145715, lr: 0.000200
|
812 |
+
2023-08-08 02:59:38,056 - INFO - [Train] step: 320799, loss: 0.091949, lr: 0.000200
|
813 |
+
2023-08-08 03:01:13,267 - INFO - [Train] step: 321199, loss: 0.100993, lr: 0.000200
|
814 |
+
2023-08-08 03:02:46,036 - INFO - [Train] step: 321599, loss: 0.137251, lr: 0.000200
|
815 |
+
2023-08-08 03:04:19,839 - INFO - [Train] step: 321999, loss: 0.129178, lr: 0.000200
|
816 |
+
2023-08-08 03:05:55,137 - INFO - [Train] step: 322399, loss: 0.153326, lr: 0.000200
|
817 |
+
2023-08-08 03:07:30,612 - INFO - [Train] step: 322799, loss: 0.091629, lr: 0.000200
|
818 |
+
2023-08-08 03:09:01,173 - INFO - [Train] step: 323199, loss: 0.108642, lr: 0.000200
|
819 |
+
2023-08-08 03:10:36,595 - INFO - [Train] step: 323599, loss: 0.114612, lr: 0.000200
|
820 |
+
2023-08-08 03:12:12,520 - INFO - [Train] step: 323999, loss: 0.121940, lr: 0.000200
|
821 |
+
2023-08-08 03:13:40,234 - INFO - [Train] step: 324399, loss: 0.092551, lr: 0.000200
|
822 |
+
2023-08-08 03:15:15,922 - INFO - [Train] step: 324799, loss: 0.127496, lr: 0.000200
|
823 |
+
2023-08-08 03:17:14,156 - INFO - [Train] step: 325199, loss: 0.104814, lr: 0.000200
|
824 |
+
2023-08-08 03:18:43,526 - INFO - [Train] step: 325599, loss: 0.101345, lr: 0.000200
|
825 |
+
2023-08-08 03:20:18,290 - INFO - [Train] step: 325999, loss: 0.105136, lr: 0.000200
|
826 |
+
2023-08-08 03:21:54,474 - INFO - [Train] step: 326399, loss: 0.152865, lr: 0.000200
|
827 |
+
2023-08-08 03:23:30,489 - INFO - [Train] step: 326799, loss: 0.107397, lr: 0.000200
|
828 |
+
2023-08-08 03:25:02,374 - INFO - [Train] step: 327199, loss: 0.129319, lr: 0.000200
|
829 |
+
2023-08-08 03:26:38,370 - INFO - [Train] step: 327599, loss: 0.118777, lr: 0.000200
|
830 |
+
2023-08-08 03:28:14,374 - INFO - [Train] step: 327999, loss: 0.134705, lr: 0.000200
|
831 |
+
2023-08-08 03:29:46,274 - INFO - [Train] step: 328399, loss: 0.106616, lr: 0.000200
|
832 |
+
2023-08-08 03:31:19,521 - INFO - [Train] step: 328799, loss: 0.093173, lr: 0.000200
|
833 |
+
2023-08-08 03:32:55,013 - INFO - [Train] step: 329199, loss: 0.070722, lr: 0.000200
|
834 |
+
2023-08-08 03:34:29,704 - INFO - [Train] step: 329599, loss: 0.104294, lr: 0.000200
|
835 |
+
2023-08-08 03:35:58,359 - INFO - [Train] step: 329999, loss: 0.113696, lr: 0.000200
|
836 |
+
2023-08-08 03:37:56,870 - INFO - [Train] step: 330399, loss: 0.107551, lr: 0.000200
|
837 |
+
2023-08-08 03:39:32,632 - INFO - [Train] step: 330799, loss: 0.108850, lr: 0.000200
|
838 |
+
2023-08-08 03:41:01,181 - INFO - [Train] step: 331199, loss: 0.107326, lr: 0.000200
|
839 |
+
2023-08-08 03:42:36,356 - INFO - [Train] step: 331599, loss: 0.126462, lr: 0.000200
|
840 |
+
2023-08-08 03:44:12,217 - INFO - [Train] step: 331999, loss: 0.096204, lr: 0.000200
|
841 |
+
2023-08-08 03:45:43,514 - INFO - [Train] step: 332399, loss: 0.120802, lr: 0.000200
|
842 |
+
2023-08-08 03:47:18,949 - INFO - [Train] step: 332799, loss: 0.110875, lr: 0.000200
|
843 |
+
2023-08-08 03:48:54,829 - INFO - [Train] step: 333199, loss: 0.106790, lr: 0.000200
|
844 |
+
2023-08-08 03:50:29,275 - INFO - [Train] step: 333599, loss: 0.113384, lr: 0.000200
|
845 |
+
2023-08-08 03:52:02,707 - INFO - [Train] step: 333999, loss: 0.109059, lr: 0.000200
|
846 |
+
2023-08-08 03:53:38,815 - INFO - [Train] step: 334399, loss: 0.132590, lr: 0.000200
|
847 |
+
2023-08-08 03:55:15,013 - INFO - [Train] step: 334799, loss: 0.106684, lr: 0.000200
|
ddpm_pred_x0_cifar10/output-2023-08-09-23-34-33.log
ADDED
The diff for this file is too large to render.
See raw diff
|
|
ddpm_pred_x0_cifar10/samples.zip
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:837da269f961683da6c52f5764b004fb327017fdcd659015188d73cf9916ec62
|
3 |
+
size 22176790
|
ddpm_pred_x0_cifar10/tensorboard/events.out.tfevents.1691408944.aa-ESC4000-G4.24711.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a50c8670058c4d5de173835ee0b65f1edcba92c9f6fe4289f740a027d5a44276
|
3 |
+
size 32800768
|
ddpm_pred_x0_cifar10/tensorboard/events.out.tfevents.1691638477.aa-ESC4000-G4.12367.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:989546dece66e89f2f1c8dd167e876af39d79ab34309a77a419829ad8ca53eb6
|
3 |
+
size 47040088
|