ayushtues commited on
Commit
52edbd6
1 Parent(s): 9b016ad

End of training

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitignore +2 -0
  2. README.md +32 -0
  3. checkpoint-500/unet/config.json +0 -0
  4. checkpoint-500/unet/diffusion_pytorch_model.bin +3 -0
  5. checkpoint-500/unet_ema/config.json +43 -0
  6. checkpoint-500/unet_ema/diffusion_pytorch_model.bin +3 -0
  7. image_0.png +0 -0
  8. image_1.png +0 -0
  9. image_10.png +0 -0
  10. image_11.png +0 -0
  11. image_12.png +0 -0
  12. image_13.png +0 -0
  13. image_14.png +0 -0
  14. image_15.png +0 -0
  15. image_2.png +0 -0
  16. image_3.png +0 -0
  17. image_4.png +0 -0
  18. image_5.png +0 -0
  19. image_6.png +0 -0
  20. image_7.png +0 -0
  21. image_8.png +0 -0
  22. image_9.png +0 -0
  23. logs/consistency-distillation/1689074974.497119/events.out.tfevents.1689074974.DESKTOP-5RK1CSO.27200.1 +3 -0
  24. logs/consistency-distillation/1689074974.5072012/hparams.yml +38 -0
  25. logs/consistency-distillation/1689145670.3773637/events.out.tfevents.1689145670.DESKTOP-5RK1CSO.6260.1 +3 -0
  26. logs/consistency-distillation/1689145670.3856902/hparams.yml +39 -0
  27. logs/consistency-distillation/1689154469.4658377/events.out.tfevents.1689154469.DESKTOP-5RK1CSO.28376.1 +3 -0
  28. logs/consistency-distillation/1689154469.4759262/hparams.yml +39 -0
  29. logs/consistency-distillation/1689154630.6460533/events.out.tfevents.1689154630.DESKTOP-5RK1CSO.900.1 +3 -0
  30. logs/consistency-distillation/1689154630.6750646/hparams.yml +39 -0
  31. logs/consistency-distillation/1689158052.0403893/events.out.tfevents.1689158052.DESKTOP-5RK1CSO.35884.1 +3 -0
  32. logs/consistency-distillation/1689158052.049507/hparams.yml +39 -0
  33. logs/consistency-distillation/1689158287.0088487/events.out.tfevents.1689158287.DESKTOP-5RK1CSO.40660.1 +3 -0
  34. logs/consistency-distillation/1689158287.016374/hparams.yml +39 -0
  35. logs/consistency-distillation/1689158979.1587937/events.out.tfevents.1689158979.DESKTOP-5RK1CSO.44652.1 +3 -0
  36. logs/consistency-distillation/1689158979.1683388/hparams.yml +39 -0
  37. logs/consistency-distillation/1689159112.2933784/events.out.tfevents.1689159112.DESKTOP-5RK1CSO.41484.1 +3 -0
  38. logs/consistency-distillation/1689159112.3039129/hparams.yml +39 -0
  39. logs/consistency-distillation/1689159276.2743661/events.out.tfevents.1689159276.DESKTOP-5RK1CSO.616.1 +3 -0
  40. logs/consistency-distillation/1689159276.2844224/hparams.yml +39 -0
  41. logs/consistency-distillation/1689159506.8864691/events.out.tfevents.1689159506.DESKTOP-5RK1CSO.17680.1 +3 -0
  42. logs/consistency-distillation/1689159506.8971152/hparams.yml +39 -0
  43. logs/consistency-distillation/1689159693.300266/events.out.tfevents.1689159693.DESKTOP-5RK1CSO.3868.1 +3 -0
  44. logs/consistency-distillation/1689159693.3108633/hparams.yml +39 -0
  45. logs/consistency-distillation/events.out.tfevents.1689074974.DESKTOP-5RK1CSO.27200.0 +3 -0
  46. logs/consistency-distillation/events.out.tfevents.1689145670.DESKTOP-5RK1CSO.6260.0 +3 -0
  47. logs/consistency-distillation/events.out.tfevents.1689154469.DESKTOP-5RK1CSO.28376.0 +3 -0
  48. logs/consistency-distillation/events.out.tfevents.1689154630.DESKTOP-5RK1CSO.900.0 +3 -0
  49. logs/consistency-distillation/events.out.tfevents.1689158052.DESKTOP-5RK1CSO.35884.0 +3 -0
  50. logs/consistency-distillation/events.out.tfevents.1689158286.DESKTOP-5RK1CSO.40660.0 +3 -0
.gitignore ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ step_*
2
+ epoch_*
README.md ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+ license: creativeml-openrail-m
4
+ base_model: google/ddpm-cifar10-32
5
+ tags:
6
+ - consistency models
7
+ - diffusers
8
+ inference: true
9
+ ---
10
+
11
+ # Consistency Model - ayushtues/cifar10-32
12
+
13
+ This is a consistency model distilled from google/ddpm-cifar10-32.
14
+ You can find some example images in the following.
15
+
16
+ ![img_0](./image_0.png)
17
+ ![img_1](./image_1.png)
18
+ ![img_2](./image_2.png)
19
+ ![img_3](./image_3.png)
20
+ ![img_4](./image_4.png)
21
+ ![img_5](./image_5.png)
22
+ ![img_6](./image_6.png)
23
+ ![img_7](./image_7.png)
24
+ ![img_8](./image_8.png)
25
+ ![img_9](./image_9.png)
26
+ ![img_10](./image_10.png)
27
+ ![img_11](./image_11.png)
28
+ ![img_12](./image_12.png)
29
+ ![img_13](./image_13.png)
30
+ ![img_14](./image_14.png)
31
+ ![img_15](./image_15.png)
32
+
checkpoint-500/unet/config.json ADDED
File without changes
checkpoint-500/unet/diffusion_pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ca696a137a79a4e6170a3666b1cc4fcf341f35eb4ea80507569f80a197e9d05
3
+ size 5208083
checkpoint-500/unet_ema/config.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "UNet2DModel",
3
+ "_diffusers_version": "0.18.0.dev0",
4
+ "act_fn": "silu",
5
+ "add_attention": true,
6
+ "attention_head_dim": 8,
7
+ "block_out_channels": [
8
+ 32,
9
+ 64
10
+ ],
11
+ "center_input_sample": false,
12
+ "class_embed_type": null,
13
+ "decay": 0.9999,
14
+ "down_block_types": [
15
+ "ResnetDownsampleBlock2D",
16
+ "AttnDownBlock2D"
17
+ ],
18
+ "downsample_padding": 1,
19
+ "downsample_type": "resnet",
20
+ "flip_sin_to_cos": true,
21
+ "freq_shift": 0,
22
+ "in_channels": 3,
23
+ "inv_gamma": 1.0,
24
+ "layers_per_block": 2,
25
+ "mid_block_scale_factor": 1,
26
+ "min_decay": 0.0,
27
+ "norm_eps": 1e-05,
28
+ "norm_num_groups": 32,
29
+ "num_class_embeds": 1000,
30
+ "optimization_step": 500,
31
+ "out_channels": 3,
32
+ "power": 0.75,
33
+ "resnet_time_scale_shift": "scale_shift",
34
+ "sample_size": 32,
35
+ "time_embedding_type": "positional",
36
+ "up_block_types": [
37
+ "AttnUpBlock2D",
38
+ "ResnetUpsampleBlock2D"
39
+ ],
40
+ "update_after_step": 0,
41
+ "upsample_type": "resnet",
42
+ "use_ema_warmup": true
43
+ }
checkpoint-500/unet_ema/diffusion_pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7df2d5e0b5444c0a82f1d4a0d6aa470ae5cd9e263aef9beb864fd8141b119145
3
+ size 5204947
image_0.png ADDED
image_1.png ADDED
image_10.png ADDED
image_11.png ADDED
image_12.png ADDED
image_13.png ADDED
image_14.png ADDED
image_15.png ADDED
image_2.png ADDED
image_3.png ADDED
image_4.png ADDED
image_5.png ADDED
image_6.png ADDED
image_7.png ADDED
image_8.png ADDED
image_9.png ADDED
logs/consistency-distillation/1689074974.497119/events.out.tfevents.1689074974.DESKTOP-5RK1CSO.27200.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:983da251a97f09de4374a9329f8d4048103d5593b470143ad251d6c9af36b415
3
+ size 1770
logs/consistency-distillation/1689074974.5072012/hparams.yml ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.95
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 1.0e-06
5
+ cache_dir: null
6
+ center_crop: true
7
+ checkpointing_steps: 500
8
+ checkpoints_total_limit: null
9
+ dataloader_num_workers: 0
10
+ dataset_config_name: null
11
+ dataset_name: cifar10
12
+ ema_inv_gamma: 1.0
13
+ ema_max_decay: 0.9999
14
+ ema_power: 0.75
15
+ eval_batch_size: 16
16
+ gradient_accumulation_steps: 1
17
+ hub_model_id: null
18
+ hub_private_repo: false
19
+ hub_token: null
20
+ learning_rate: 0.0001
21
+ local_rank: -1
22
+ logger: tensorboard
23
+ logging_dir: logs
24
+ lr_scheduler: cosine
25
+ lr_warmup_steps: 500
26
+ mixed_precision: 'no'
27
+ model_config_name_or_path: null
28
+ num_epochs: 100
29
+ output_dir: cifar10-32
30
+ overwrite_output_dir: false
31
+ push_to_hub: true
32
+ random_flip: true
33
+ resolution: 32
34
+ resume_from_checkpoint: null
35
+ save_images_epochs: 10
36
+ save_model_epochs: 10
37
+ train_batch_size: 16
38
+ train_data_dir: null
logs/consistency-distillation/1689145670.3773637/events.out.tfevents.1689145670.DESKTOP-5RK1CSO.6260.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c74cd81983b34e539153100628516238c846b8d3def61972ab17c12d43d77f5b
3
+ size 1880
logs/consistency-distillation/1689145670.3856902/hparams.yml ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.95
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 1.0e-06
5
+ cache_dir: null
6
+ center_crop: true
7
+ checkpointing_steps: 500
8
+ checkpoints_total_limit: null
9
+ dataloader_num_workers: 0
10
+ dataset_config_name: null
11
+ dataset_name: cifar10
12
+ ema_inv_gamma: 1.0
13
+ ema_max_decay: 0.9999
14
+ ema_power: 0.75
15
+ eval_batch_size: 16
16
+ gradient_accumulation_steps: 1
17
+ hub_model_id: null
18
+ hub_private_repo: false
19
+ hub_token: null
20
+ learning_rate: 0.0001
21
+ local_rank: -1
22
+ logger: tensorboard
23
+ logging_dir: logs
24
+ lr_scheduler: cosine
25
+ lr_warmup_steps: 500
26
+ mixed_precision: 'no'
27
+ model_config_name_or_path: null
28
+ num_epochs: 100
29
+ output_dir: cifar10-32
30
+ overwrite_output_dir: false
31
+ pretrained_teacher_model_name_or_path: google/ddpm-cifar10-32
32
+ push_to_hub: true
33
+ random_flip: true
34
+ resolution: 32
35
+ resume_from_checkpoint: null
36
+ save_images_epochs: 10
37
+ save_model_epochs: 10
38
+ train_batch_size: 16
39
+ train_data_dir: null
logs/consistency-distillation/1689154469.4658377/events.out.tfevents.1689154469.DESKTOP-5RK1CSO.28376.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49bfbc47cdbbe8f99ea2d7709abe681ead121373421ec41d2606e3eee765f765
3
+ size 1880
logs/consistency-distillation/1689154469.4759262/hparams.yml ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.95
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 1.0e-06
5
+ cache_dir: null
6
+ center_crop: true
7
+ checkpointing_steps: 500
8
+ checkpoints_total_limit: null
9
+ dataloader_num_workers: 0
10
+ dataset_config_name: null
11
+ dataset_name: cifar10
12
+ ema_inv_gamma: 1.0
13
+ ema_max_decay: 0.9999
14
+ ema_power: 0.75
15
+ eval_batch_size: 16
16
+ gradient_accumulation_steps: 1
17
+ hub_model_id: null
18
+ hub_private_repo: false
19
+ hub_token: null
20
+ learning_rate: 0.0001
21
+ local_rank: -1
22
+ logger: tensorboard
23
+ logging_dir: logs
24
+ lr_scheduler: cosine
25
+ lr_warmup_steps: 500
26
+ mixed_precision: 'no'
27
+ model_config_name_or_path: null
28
+ num_epochs: 100
29
+ output_dir: cifar10-32
30
+ overwrite_output_dir: false
31
+ pretrained_teacher_model_name_or_path: google/ddpm-cifar10-32
32
+ push_to_hub: true
33
+ random_flip: true
34
+ resolution: 32
35
+ resume_from_checkpoint: null
36
+ save_images_epochs: 10
37
+ save_model_epochs: 10
38
+ train_batch_size: 16
39
+ train_data_dir: null
logs/consistency-distillation/1689154630.6460533/events.out.tfevents.1689154630.DESKTOP-5RK1CSO.900.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9472700e913de8b3ef88cc068c42f76472430e5e85a35f343ca2dbc9984852e6
3
+ size 1880
logs/consistency-distillation/1689154630.6750646/hparams.yml ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.95
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 1.0e-06
5
+ cache_dir: null
6
+ center_crop: true
7
+ checkpointing_steps: 500
8
+ checkpoints_total_limit: null
9
+ dataloader_num_workers: 0
10
+ dataset_config_name: null
11
+ dataset_name: cifar10
12
+ ema_inv_gamma: 1.0
13
+ ema_max_decay: 0.9999
14
+ ema_power: 0.75
15
+ eval_batch_size: 16
16
+ gradient_accumulation_steps: 1
17
+ hub_model_id: null
18
+ hub_private_repo: false
19
+ hub_token: null
20
+ learning_rate: 0.0001
21
+ local_rank: -1
22
+ logger: tensorboard
23
+ logging_dir: logs
24
+ lr_scheduler: cosine
25
+ lr_warmup_steps: 500
26
+ mixed_precision: 'no'
27
+ model_config_name_or_path: null
28
+ num_epochs: 100
29
+ output_dir: cifar10-32
30
+ overwrite_output_dir: false
31
+ pretrained_teacher_model_name_or_path: google/ddpm-cifar10-32
32
+ push_to_hub: true
33
+ random_flip: true
34
+ resolution: 32
35
+ resume_from_checkpoint: null
36
+ save_images_epochs: 10
37
+ save_model_epochs: 10
38
+ train_batch_size: 16
39
+ train_data_dir: null
logs/consistency-distillation/1689158052.0403893/events.out.tfevents.1689158052.DESKTOP-5RK1CSO.35884.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af039f9cc4085a888a0ed43f6bdb916eacc873ed57e554f2e4c110ad3cc3b19f
3
+ size 1880
logs/consistency-distillation/1689158052.049507/hparams.yml ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.95
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 1.0e-06
5
+ cache_dir: null
6
+ center_crop: true
7
+ checkpointing_steps: 500
8
+ checkpoints_total_limit: null
9
+ dataloader_num_workers: 0
10
+ dataset_config_name: null
11
+ dataset_name: cifar10
12
+ ema_inv_gamma: 1.0
13
+ ema_max_decay: 0.9999
14
+ ema_power: 0.75
15
+ eval_batch_size: 16
16
+ gradient_accumulation_steps: 1
17
+ hub_model_id: null
18
+ hub_private_repo: false
19
+ hub_token: null
20
+ learning_rate: 0.0001
21
+ local_rank: -1
22
+ logger: tensorboard
23
+ logging_dir: logs
24
+ lr_scheduler: cosine
25
+ lr_warmup_steps: 500
26
+ mixed_precision: 'no'
27
+ model_config_name_or_path: null
28
+ num_epochs: 100
29
+ output_dir: cifar10-32
30
+ overwrite_output_dir: false
31
+ pretrained_teacher_model_name_or_path: google/ddpm-cifar10-32
32
+ push_to_hub: true
33
+ random_flip: true
34
+ resolution: 32
35
+ resume_from_checkpoint: null
36
+ save_images_epochs: 10
37
+ save_model_epochs: 10
38
+ train_batch_size: 16
39
+ train_data_dir: null
logs/consistency-distillation/1689158287.0088487/events.out.tfevents.1689158287.DESKTOP-5RK1CSO.40660.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91251e5c5b858dff83f8fd26e6ee8b9dd818443ad44a5b1c55b9f51967010a76
3
+ size 1880
logs/consistency-distillation/1689158287.016374/hparams.yml ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.95
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 1.0e-06
5
+ cache_dir: null
6
+ center_crop: true
7
+ checkpointing_steps: 500
8
+ checkpoints_total_limit: null
9
+ dataloader_num_workers: 0
10
+ dataset_config_name: null
11
+ dataset_name: cifar10
12
+ ema_inv_gamma: 1.0
13
+ ema_max_decay: 0.9999
14
+ ema_power: 0.75
15
+ eval_batch_size: 16
16
+ gradient_accumulation_steps: 1
17
+ hub_model_id: null
18
+ hub_private_repo: false
19
+ hub_token: null
20
+ learning_rate: 0.0001
21
+ local_rank: -1
22
+ logger: tensorboard
23
+ logging_dir: logs
24
+ lr_scheduler: cosine
25
+ lr_warmup_steps: 500
26
+ mixed_precision: 'no'
27
+ model_config_name_or_path: null
28
+ num_epochs: 1
29
+ output_dir: cifar10-32
30
+ overwrite_output_dir: false
31
+ pretrained_teacher_model_name_or_path: google/ddpm-cifar10-32
32
+ push_to_hub: true
33
+ random_flip: true
34
+ resolution: 32
35
+ resume_from_checkpoint: null
36
+ save_images_epochs: 10
37
+ save_model_epochs: 10
38
+ train_batch_size: 2
39
+ train_data_dir: null
logs/consistency-distillation/1689158979.1587937/events.out.tfevents.1689158979.DESKTOP-5RK1CSO.44652.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0212bbecd76b84e90eac71093c7536368085d1199655b081e1517f3535a3301
3
+ size 1880
logs/consistency-distillation/1689158979.1683388/hparams.yml ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.95
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 1.0e-06
5
+ cache_dir: null
6
+ center_crop: true
7
+ checkpointing_steps: 500
8
+ checkpoints_total_limit: null
9
+ dataloader_num_workers: 0
10
+ dataset_config_name: null
11
+ dataset_name: cifar10
12
+ ema_inv_gamma: 1.0
13
+ ema_max_decay: 0.9999
14
+ ema_power: 0.75
15
+ eval_batch_size: 16
16
+ gradient_accumulation_steps: 1
17
+ hub_model_id: null
18
+ hub_private_repo: false
19
+ hub_token: null
20
+ learning_rate: 0.0001
21
+ local_rank: -1
22
+ logger: tensorboard
23
+ logging_dir: logs
24
+ lr_scheduler: cosine
25
+ lr_warmup_steps: 500
26
+ mixed_precision: 'no'
27
+ model_config_name_or_path: null
28
+ num_epochs: 1
29
+ output_dir: cifar10-32
30
+ overwrite_output_dir: false
31
+ pretrained_teacher_model_name_or_path: google/ddpm-cifar10-32
32
+ push_to_hub: true
33
+ random_flip: true
34
+ resolution: 32
35
+ resume_from_checkpoint: null
36
+ save_images_epochs: 10
37
+ save_model_epochs: 10
38
+ train_batch_size: 2
39
+ train_data_dir: null
logs/consistency-distillation/1689159112.2933784/events.out.tfevents.1689159112.DESKTOP-5RK1CSO.41484.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9086aaf81f4d5e5dd068b6787f423ffe0ea444243b45aac90c7b075b4d05ca0
3
+ size 1880
logs/consistency-distillation/1689159112.3039129/hparams.yml ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.95
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 1.0e-06
5
+ cache_dir: null
6
+ center_crop: true
7
+ checkpointing_steps: 500
8
+ checkpoints_total_limit: null
9
+ dataloader_num_workers: 0
10
+ dataset_config_name: null
11
+ dataset_name: cifar10
12
+ ema_inv_gamma: 1.0
13
+ ema_max_decay: 0.9999
14
+ ema_power: 0.75
15
+ eval_batch_size: 16
16
+ gradient_accumulation_steps: 1
17
+ hub_model_id: null
18
+ hub_private_repo: false
19
+ hub_token: null
20
+ learning_rate: 0.0001
21
+ local_rank: -1
22
+ logger: tensorboard
23
+ logging_dir: logs
24
+ lr_scheduler: cosine
25
+ lr_warmup_steps: 500
26
+ mixed_precision: 'no'
27
+ model_config_name_or_path: null
28
+ num_epochs: 1
29
+ output_dir: cifar10-32
30
+ overwrite_output_dir: false
31
+ pretrained_teacher_model_name_or_path: google/ddpm-cifar10-32
32
+ push_to_hub: true
33
+ random_flip: true
34
+ resolution: 32
35
+ resume_from_checkpoint: null
36
+ save_images_epochs: 10
37
+ save_model_epochs: 10
38
+ train_batch_size: 2
39
+ train_data_dir: null
logs/consistency-distillation/1689159276.2743661/events.out.tfevents.1689159276.DESKTOP-5RK1CSO.616.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0ef9650b6fa301f372c118d3799879d4aa1c73827e558c723aa0fc5bcfcdc51
3
+ size 1880
logs/consistency-distillation/1689159276.2844224/hparams.yml ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.95
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 1.0e-06
5
+ cache_dir: null
6
+ center_crop: true
7
+ checkpointing_steps: 500
8
+ checkpoints_total_limit: null
9
+ dataloader_num_workers: 0
10
+ dataset_config_name: null
11
+ dataset_name: cifar10
12
+ ema_inv_gamma: 1.0
13
+ ema_max_decay: 0.9999
14
+ ema_power: 0.75
15
+ eval_batch_size: 16
16
+ gradient_accumulation_steps: 1
17
+ hub_model_id: null
18
+ hub_private_repo: false
19
+ hub_token: null
20
+ learning_rate: 0.0001
21
+ local_rank: -1
22
+ logger: tensorboard
23
+ logging_dir: logs
24
+ lr_scheduler: cosine
25
+ lr_warmup_steps: 500
26
+ mixed_precision: 'no'
27
+ model_config_name_or_path: null
28
+ num_epochs: 1
29
+ output_dir: cifar10-32
30
+ overwrite_output_dir: false
31
+ pretrained_teacher_model_name_or_path: google/ddpm-cifar10-32
32
+ push_to_hub: true
33
+ random_flip: true
34
+ resolution: 32
35
+ resume_from_checkpoint: null
36
+ save_images_epochs: 10
37
+ save_model_epochs: 10
38
+ train_batch_size: 2
39
+ train_data_dir: null
logs/consistency-distillation/1689159506.8864691/events.out.tfevents.1689159506.DESKTOP-5RK1CSO.17680.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:216ea4137ff76ed2176b3487cb36254fd5a89d20a7f68d9d7a75e3c3599d16f3
3
+ size 1880
logs/consistency-distillation/1689159506.8971152/hparams.yml ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.95
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 1.0e-06
5
+ cache_dir: null
6
+ center_crop: true
7
+ checkpointing_steps: 500
8
+ checkpoints_total_limit: null
9
+ dataloader_num_workers: 0
10
+ dataset_config_name: null
11
+ dataset_name: cifar10
12
+ ema_inv_gamma: 1.0
13
+ ema_max_decay: 0.9999
14
+ ema_power: 0.75
15
+ eval_batch_size: 16
16
+ gradient_accumulation_steps: 1
17
+ hub_model_id: null
18
+ hub_private_repo: false
19
+ hub_token: null
20
+ learning_rate: 0.0001
21
+ local_rank: -1
22
+ logger: tensorboard
23
+ logging_dir: logs
24
+ lr_scheduler: cosine
25
+ lr_warmup_steps: 500
26
+ mixed_precision: 'no'
27
+ model_config_name_or_path: null
28
+ num_epochs: 1
29
+ output_dir: cifar10-32
30
+ overwrite_output_dir: false
31
+ pretrained_teacher_model_name_or_path: google/ddpm-cifar10-32
32
+ push_to_hub: true
33
+ random_flip: true
34
+ resolution: 32
35
+ resume_from_checkpoint: null
36
+ save_images_epochs: 10
37
+ save_model_epochs: 10
38
+ train_batch_size: 2
39
+ train_data_dir: null
logs/consistency-distillation/1689159693.300266/events.out.tfevents.1689159693.DESKTOP-5RK1CSO.3868.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d6da6cbb1cf6ef9f91d97448c6cd43bc246bdcfee0338c8b61d0d1b613e3e4b
3
+ size 1880
logs/consistency-distillation/1689159693.3108633/hparams.yml ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.95
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 1.0e-06
5
+ cache_dir: null
6
+ center_crop: true
7
+ checkpointing_steps: 500
8
+ checkpoints_total_limit: null
9
+ dataloader_num_workers: 0
10
+ dataset_config_name: null
11
+ dataset_name: cifar10
12
+ ema_inv_gamma: 1.0
13
+ ema_max_decay: 0.9999
14
+ ema_power: 0.75
15
+ eval_batch_size: 16
16
+ gradient_accumulation_steps: 1
17
+ hub_model_id: null
18
+ hub_private_repo: false
19
+ hub_token: null
20
+ learning_rate: 0.0001
21
+ local_rank: -1
22
+ logger: tensorboard
23
+ logging_dir: logs
24
+ lr_scheduler: cosine
25
+ lr_warmup_steps: 500
26
+ mixed_precision: 'no'
27
+ model_config_name_or_path: null
28
+ num_epochs: 1
29
+ output_dir: cifar10-32
30
+ overwrite_output_dir: false
31
+ pretrained_teacher_model_name_or_path: google/ddpm-cifar10-32
32
+ push_to_hub: true
33
+ random_flip: true
34
+ resolution: 32
35
+ resume_from_checkpoint: null
36
+ save_images_epochs: 10
37
+ save_model_epochs: 10
38
+ train_batch_size: 2
39
+ train_data_dir: null
logs/consistency-distillation/events.out.tfevents.1689074974.DESKTOP-5RK1CSO.27200.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:489d27f919ceb85d0460c0e28c2f4f021c43039716f10c7055e14a19a1ce8ea2
3
+ size 86905
logs/consistency-distillation/events.out.tfevents.1689145670.DESKTOP-5RK1CSO.6260.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df1e583cbdd3dfa8c7819c48372fe33aa11b5cad9c6d0e64f42af0c8c15a1f66
3
+ size 88
logs/consistency-distillation/events.out.tfevents.1689154469.DESKTOP-5RK1CSO.28376.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30dd86d1d5745c5af0dee2f8d121f8d473af00babd0a5aecf95d27eb1c5823ce
3
+ size 88
logs/consistency-distillation/events.out.tfevents.1689154630.DESKTOP-5RK1CSO.900.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:924e1916f3331ca0012115dff71912c293e1056bc1e37f0cfbaa1ba43244f59e
3
+ size 88
logs/consistency-distillation/events.out.tfevents.1689158052.DESKTOP-5RK1CSO.35884.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:371277f92dde907a0502190abb9e7d1e0f8a1f5676c2fee5acbebb499781688a
3
+ size 1969
logs/consistency-distillation/events.out.tfevents.1689158286.DESKTOP-5RK1CSO.40660.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:40ef3e70d08ba06aea51172a58a2659bb303240bdde57a3a49a12a13619b77da
3
+ size 91455