jrd971000 commited on
Commit
cd27c71
1 Parent(s): cc4382f

End of training

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. checkpoint-10000/optimizer.bin +3 -0
  2. checkpoint-10000/pytorch_lora_weights.safetensors +3 -0
  3. checkpoint-10000/random_states_0.pkl +3 -0
  4. checkpoint-10000/scaler.pt +3 -0
  5. checkpoint-10000/scheduler.bin +3 -0
  6. checkpoint-2000/optimizer.bin +3 -0
  7. checkpoint-2000/pytorch_lora_weights.safetensors +3 -0
  8. checkpoint-2000/random_states_0.pkl +3 -0
  9. checkpoint-2000/scaler.pt +3 -0
  10. checkpoint-2000/scheduler.bin +3 -0
  11. checkpoint-4000/optimizer.bin +3 -0
  12. checkpoint-4000/pytorch_lora_weights.safetensors +3 -0
  13. checkpoint-4000/random_states_0.pkl +3 -0
  14. checkpoint-4000/scaler.pt +3 -0
  15. checkpoint-4000/scheduler.bin +3 -0
  16. checkpoint-6000/optimizer.bin +3 -0
  17. checkpoint-6000/pytorch_lora_weights.safetensors +3 -0
  18. checkpoint-6000/random_states_0.pkl +3 -0
  19. checkpoint-6000/scaler.pt +3 -0
  20. checkpoint-6000/scheduler.bin +3 -0
  21. checkpoint-8000/optimizer.bin +3 -0
  22. checkpoint-8000/pytorch_lora_weights.safetensors +3 -0
  23. checkpoint-8000/random_states_0.pkl +3 -0
  24. checkpoint-8000/scaler.pt +3 -0
  25. checkpoint-8000/scheduler.bin +3 -0
  26. logs/diffusion-dpo-lora/1708613661.7299602/events.out.tfevents.1708613661.4f9809b-lcedt.1355832.1 +3 -0
  27. logs/diffusion-dpo-lora/1708613661.7326505/hparams.yml +50 -0
  28. logs/diffusion-dpo-lora/1708613796.3902838/events.out.tfevents.1708613796.4f9809b-lcedt.1357248.1 +3 -0
  29. logs/diffusion-dpo-lora/1708613796.3919551/hparams.yml +50 -0
  30. logs/diffusion-dpo-lora/1708613879.877602/events.out.tfevents.1708613879.4f9809b-lcedt.1358274.1 +3 -0
  31. logs/diffusion-dpo-lora/1708613879.880194/hparams.yml +50 -0
  32. logs/diffusion-dpo-lora/1708614245.0886717/events.out.tfevents.1708614245.4f9809b-lcedt.1362678.1 +3 -0
  33. logs/diffusion-dpo-lora/1708614245.091276/hparams.yml +50 -0
  34. logs/diffusion-dpo-lora/1708700486.347142/events.out.tfevents.1708700486.4f9809b-lcedt.2210234.1 +3 -0
  35. logs/diffusion-dpo-lora/1708700486.3496387/hparams.yml +50 -0
  36. logs/diffusion-dpo-lora/1708700603.4479868/events.out.tfevents.1708700603.4f9809b-lcedt.2211491.1 +3 -0
  37. logs/diffusion-dpo-lora/1708700603.4505715/hparams.yml +50 -0
  38. logs/diffusion-dpo-lora/1708700969.3730564/events.out.tfevents.1708700969.4f9809b-lcedt.2215537.1 +3 -0
  39. logs/diffusion-dpo-lora/1708700969.375287/hparams.yml +50 -0
  40. logs/diffusion-dpo-lora/1708701303.0831516/events.out.tfevents.1708701303.4f9809b-lcedt.2219311.1 +3 -0
  41. logs/diffusion-dpo-lora/1708701303.0857437/hparams.yml +50 -0
  42. logs/diffusion-dpo-lora/1708701508.6600106/events.out.tfevents.1708701508.4f9809b-lcedt.2221662.1 +3 -0
  43. logs/diffusion-dpo-lora/1708701508.6627216/hparams.yml +50 -0
  44. logs/diffusion-dpo-lora/1708701834.8009768/events.out.tfevents.1708701834.4f9809b-lcedt.2225371.1 +3 -0
  45. logs/diffusion-dpo-lora/1708701834.8036294/hparams.yml +50 -0
  46. logs/diffusion-dpo-lora/1708701933.5214443/events.out.tfevents.1708701933.4f9809b-lcedt.2226604.1 +3 -0
  47. logs/diffusion-dpo-lora/1708701933.524116/hparams.yml +50 -0
  48. logs/diffusion-dpo-lora/1708702436.6012197/events.out.tfevents.1708702436.4f9809b-lcedt.2232778.1 +3 -0
  49. logs/diffusion-dpo-lora/1708702436.603921/hparams.yml +50 -0
  50. logs/diffusion-dpo-lora/1708702585.2958512/events.out.tfevents.1708702585.4f9809b-lcedt.2235338.1 +3 -0
checkpoint-10000/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69f0f427ce150a27ab0cd5b836eee2a8197bb3f11d5db3a304ea2590c9db09b4
3
+ size 4523514
checkpoint-10000/pytorch_lora_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eedd459a5daa854173e521df77adefe943a29f0b3482c69c93a47bacec8bee4e
3
+ size 6414992
checkpoint-10000/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2416eea7977d9c3d5a663d1b3ff143c6be91a742a56d552fbf8a585ad32f53c7
3
+ size 14344
checkpoint-10000/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b01eaf7507960eda74c6413295234d9ccc7b89831ddf270ef7b99974ac1b36e
3
+ size 988
checkpoint-10000/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d40855e5eb64eb0e33ad44631d2fa5da316a810cc7808ce708a1a5f7ebcd3dd
3
+ size 1000
checkpoint-2000/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a691313d68bf632c3615611dffdeeb598430aaf305fd8d2cdcd78a0357dcbff
3
+ size 4523514
checkpoint-2000/pytorch_lora_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:00b4a6f82e15356fb9c45a5931d2e45c05b9233097430c0ded9c001882f82be7
3
+ size 6414992
checkpoint-2000/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86659091c1c0a4bac35c38b105e41bb5c54d4f9b46ef26732f85f8a5a1287ded
3
+ size 14344
checkpoint-2000/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:595f2f1c1914bd5f99bf5010f299c5c4c350963b179812c407ba4c5971635220
3
+ size 988
checkpoint-2000/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aeff6094d449c4fe8121495138094c21c9431b090806eb76e0a9784bd1769140
3
+ size 1000
checkpoint-4000/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e3f827d4a065ccfdebdf09fbc616afa0d59f33f553a1fa45bab8946b84f08bb
3
+ size 4523514
checkpoint-4000/pytorch_lora_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4fbf41e4b38aa1c72763f1be93f3450c0837a0dae0453e9ca5a19de0752b708
3
+ size 6414992
checkpoint-4000/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf57656d5565749f061b99fcec0e5ac8652a138594c1824d43bf0e78517e5ece
3
+ size 14344
checkpoint-4000/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf67297240940deef61f269552b107ae5159a949a2318f4b9e6133041d110c21
3
+ size 988
checkpoint-4000/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bfae37a959efdf5f32dbcc7097c80840286251f5f0021eddaf1a92e26aa851d1
3
+ size 1000
checkpoint-6000/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63c21d027dbbf68c5be64e9b209e1ab8549d24cebef6d1a749ed23085d2782e6
3
+ size 4523514
checkpoint-6000/pytorch_lora_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5479aac54a1aff54b9d0ef4e0b35a3becbd29ecbf0c2c4c4f8c247ec0e6e13ef
3
+ size 6414992
checkpoint-6000/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f7aa108c72ad120394b1d7ae934879db8afb132651dfe8922f494ff5295c610f
3
+ size 14344
checkpoint-6000/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:53a5d8e98a00ad2deffedfcc0c989554a76bf030efa7b2e008d1f8bf8a1231ce
3
+ size 988
checkpoint-6000/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:faa3a01db4829b75094498dae75c9bf70bec8f4301d3256efc902ad39d0cd558
3
+ size 1000
checkpoint-8000/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:57f38abcfda5f67165df25fd30e41bc1d4b0c31a6d2fdbe6389beb63ccf59f80
3
+ size 4523514
checkpoint-8000/pytorch_lora_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1b3839804e049028db59952193443d93df728a2bd9db9788a3ca160f0ad925d
3
+ size 6414992
checkpoint-8000/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49f80bf53c6a19735ebc0701e0da4ff831d6e8863d32578682d92757be7483d3
3
+ size 14344
checkpoint-8000/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa8545d6c20b8d52fd9b5de3e4e73436bd9878fc77deab133d366fea72e94474
3
+ size 988
checkpoint-8000/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:115465021f9dd7f9fdce29801d88283005fc2a008320f27af32866259e560428
3
+ size 1000
logs/diffusion-dpo-lora/1708613661.7299602/events.out.tfevents.1708613661.4f9809b-lcedt.1355832.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d788304c57549c355b60d15fccd49e416e8d157d5f1ffcac03c7e228beed205
3
+ size 2383
logs/diffusion-dpo-lora/1708613661.7326505/hparams.yml ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 0.01
5
+ allow_tf32: false
6
+ beta_dpo: 2500
7
+ cache_dir: null
8
+ checkpointing_steps: 2000
9
+ checkpoints_total_limit: null
10
+ dataloader_num_workers: 0
11
+ dataset_name: kashif/pickascore
12
+ dataset_split_name: validation
13
+ enable_xformers_memory_efficient_attention: false
14
+ gradient_accumulation_steps: 1
15
+ gradient_checkpointing: true
16
+ hub_model_id: null
17
+ hub_token: null
18
+ learning_rate: 1.0e-05
19
+ local_rank: -1
20
+ logging_dir: logs
21
+ loss_type: sigmoid
22
+ lr_num_cycles: 1
23
+ lr_power: 1.0
24
+ lr_scheduler: constant
25
+ lr_warmup_steps: 0
26
+ max_grad_norm: 1.0
27
+ max_train_samples: null
28
+ max_train_steps: 10000
29
+ mixed_precision: fp16
30
+ no_hflip: false
31
+ num_train_epochs: 20
32
+ output_dir: diffusion-dpo
33
+ pretrained_model_name_or_path: runwayml/stable-diffusion-v1-5
34
+ prior_generation_precision: null
35
+ push_to_hub: true
36
+ random_crop: false
37
+ rank: 8
38
+ report_to: tensorboard
39
+ resolution: 512
40
+ resume_from_checkpoint: null
41
+ revision: null
42
+ run_validation: true
43
+ scale_lr: false
44
+ seed: 0
45
+ tracker_name: diffusion-dpo-lora
46
+ train_batch_size: 1
47
+ use_8bit_adam: true
48
+ vae_encode_batch_size: 8
49
+ validation_steps: 200
50
+ variant: null
logs/diffusion-dpo-lora/1708613796.3902838/events.out.tfevents.1708613796.4f9809b-lcedt.1357248.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3068a36277bf300c28a7e81a3f112fbc72963263aab5a8d188f84f8bc416a88a
3
+ size 2383
logs/diffusion-dpo-lora/1708613796.3919551/hparams.yml ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 0.01
5
+ allow_tf32: false
6
+ beta_dpo: 2500
7
+ cache_dir: null
8
+ checkpointing_steps: 2000
9
+ checkpoints_total_limit: null
10
+ dataloader_num_workers: 0
11
+ dataset_name: kashif/pickascore
12
+ dataset_split_name: validation
13
+ enable_xformers_memory_efficient_attention: false
14
+ gradient_accumulation_steps: 1
15
+ gradient_checkpointing: true
16
+ hub_model_id: null
17
+ hub_token: null
18
+ learning_rate: 1.0e-05
19
+ local_rank: -1
20
+ logging_dir: logs
21
+ loss_type: sigmoid
22
+ lr_num_cycles: 1
23
+ lr_power: 1.0
24
+ lr_scheduler: constant
25
+ lr_warmup_steps: 0
26
+ max_grad_norm: 1.0
27
+ max_train_samples: null
28
+ max_train_steps: 10000
29
+ mixed_precision: fp16
30
+ no_hflip: false
31
+ num_train_epochs: 20
32
+ output_dir: diffusion-dpo
33
+ pretrained_model_name_or_path: runwayml/stable-diffusion-v1-5
34
+ prior_generation_precision: null
35
+ push_to_hub: true
36
+ random_crop: false
37
+ rank: 8
38
+ report_to: tensorboard
39
+ resolution: 512
40
+ resume_from_checkpoint: null
41
+ revision: null
42
+ run_validation: true
43
+ scale_lr: false
44
+ seed: 0
45
+ tracker_name: diffusion-dpo-lora
46
+ train_batch_size: 1
47
+ use_8bit_adam: true
48
+ vae_encode_batch_size: 8
49
+ validation_steps: 200
50
+ variant: null
logs/diffusion-dpo-lora/1708613879.877602/events.out.tfevents.1708613879.4f9809b-lcedt.1358274.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bee292e649cb96f47e14918f3b3e4110440d00c3b35982be829fd42e791339d0
3
+ size 2383
logs/diffusion-dpo-lora/1708613879.880194/hparams.yml ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 0.01
5
+ allow_tf32: false
6
+ beta_dpo: 2500
7
+ cache_dir: null
8
+ checkpointing_steps: 2000
9
+ checkpoints_total_limit: null
10
+ dataloader_num_workers: 0
11
+ dataset_name: kashif/pickascore
12
+ dataset_split_name: validation
13
+ enable_xformers_memory_efficient_attention: false
14
+ gradient_accumulation_steps: 1
15
+ gradient_checkpointing: true
16
+ hub_model_id: null
17
+ hub_token: null
18
+ learning_rate: 1.0e-05
19
+ local_rank: -1
20
+ logging_dir: logs
21
+ loss_type: sigmoid
22
+ lr_num_cycles: 1
23
+ lr_power: 1.0
24
+ lr_scheduler: constant
25
+ lr_warmup_steps: 0
26
+ max_grad_norm: 1.0
27
+ max_train_samples: null
28
+ max_train_steps: 10000
29
+ mixed_precision: fp16
30
+ no_hflip: false
31
+ num_train_epochs: 20
32
+ output_dir: diffusion-dpo
33
+ pretrained_model_name_or_path: runwayml/stable-diffusion-v1-5
34
+ prior_generation_precision: null
35
+ push_to_hub: true
36
+ random_crop: false
37
+ rank: 8
38
+ report_to: tensorboard
39
+ resolution: 512
40
+ resume_from_checkpoint: null
41
+ revision: null
42
+ run_validation: true
43
+ scale_lr: false
44
+ seed: 0
45
+ tracker_name: diffusion-dpo-lora
46
+ train_batch_size: 1
47
+ use_8bit_adam: true
48
+ vae_encode_batch_size: 8
49
+ validation_steps: 200
50
+ variant: null
logs/diffusion-dpo-lora/1708614245.0886717/events.out.tfevents.1708614245.4f9809b-lcedt.1362678.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01f8162b83c50108a311b5eb1ab43abb858082408fd01a659593ffcc9e67dca9
3
+ size 2383
logs/diffusion-dpo-lora/1708614245.091276/hparams.yml ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 0.01
5
+ allow_tf32: false
6
+ beta_dpo: 2500
7
+ cache_dir: null
8
+ checkpointing_steps: 2000
9
+ checkpoints_total_limit: null
10
+ dataloader_num_workers: 0
11
+ dataset_name: kashif/pickascore
12
+ dataset_split_name: validation
13
+ enable_xformers_memory_efficient_attention: false
14
+ gradient_accumulation_steps: 1
15
+ gradient_checkpointing: true
16
+ hub_model_id: null
17
+ hub_token: null
18
+ learning_rate: 1.0e-05
19
+ local_rank: -1
20
+ logging_dir: logs
21
+ loss_type: sigmoid
22
+ lr_num_cycles: 1
23
+ lr_power: 1.0
24
+ lr_scheduler: constant
25
+ lr_warmup_steps: 0
26
+ max_grad_norm: 1.0
27
+ max_train_samples: null
28
+ max_train_steps: 10000
29
+ mixed_precision: fp16
30
+ no_hflip: false
31
+ num_train_epochs: 20
32
+ output_dir: diffusion-dpo
33
+ pretrained_model_name_or_path: runwayml/stable-diffusion-v1-5
34
+ prior_generation_precision: null
35
+ push_to_hub: true
36
+ random_crop: false
37
+ rank: 8
38
+ report_to: tensorboard
39
+ resolution: 512
40
+ resume_from_checkpoint: null
41
+ revision: null
42
+ run_validation: true
43
+ scale_lr: false
44
+ seed: 0
45
+ tracker_name: diffusion-dpo-lora
46
+ train_batch_size: 1
47
+ use_8bit_adam: true
48
+ vae_encode_batch_size: 8
49
+ validation_steps: 200
50
+ variant: null
logs/diffusion-dpo-lora/1708700486.347142/events.out.tfevents.1708700486.4f9809b-lcedt.2210234.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef61e740bbff3de4c0b1ec4e0a26d53a09ba3afcf13b8e8e5e7836da07c0797d
3
+ size 2383
logs/diffusion-dpo-lora/1708700486.3496387/hparams.yml ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 0.01
5
+ allow_tf32: false
6
+ beta_dpo: 2500
7
+ cache_dir: null
8
+ checkpointing_steps: 2000
9
+ checkpoints_total_limit: null
10
+ dataloader_num_workers: 0
11
+ dataset_name: kashif/pickascore
12
+ dataset_split_name: validation
13
+ enable_xformers_memory_efficient_attention: false
14
+ gradient_accumulation_steps: 1
15
+ gradient_checkpointing: true
16
+ hub_model_id: null
17
+ hub_token: null
18
+ learning_rate: 1.0e-05
19
+ local_rank: -1
20
+ logging_dir: logs
21
+ loss_type: sigmoid
22
+ lr_num_cycles: 1
23
+ lr_power: 1.0
24
+ lr_scheduler: constant
25
+ lr_warmup_steps: 0
26
+ max_grad_norm: 1.0
27
+ max_train_samples: null
28
+ max_train_steps: 10000
29
+ mixed_precision: fp16
30
+ no_hflip: false
31
+ num_train_epochs: 20
32
+ output_dir: diffusion-dpo
33
+ pretrained_model_name_or_path: runwayml/stable-diffusion-v1-5
34
+ prior_generation_precision: null
35
+ push_to_hub: true
36
+ random_crop: false
37
+ rank: 8
38
+ report_to: tensorboard
39
+ resolution: 512
40
+ resume_from_checkpoint: null
41
+ revision: null
42
+ run_validation: true
43
+ scale_lr: false
44
+ seed: 0
45
+ tracker_name: diffusion-dpo-lora
46
+ train_batch_size: 1
47
+ use_8bit_adam: true
48
+ vae_encode_batch_size: 8
49
+ validation_steps: 200
50
+ variant: null
logs/diffusion-dpo-lora/1708700603.4479868/events.out.tfevents.1708700603.4f9809b-lcedt.2211491.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fbbfb7af07c245b27445d2e34f9ccb5ff3a03d8a7676c8ad7c1f5b5d5309ba5f
3
+ size 2383
logs/diffusion-dpo-lora/1708700603.4505715/hparams.yml ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 0.01
5
+ allow_tf32: false
6
+ beta_dpo: 2500
7
+ cache_dir: null
8
+ checkpointing_steps: 2000
9
+ checkpoints_total_limit: null
10
+ dataloader_num_workers: 0
11
+ dataset_name: kashif/pickascore
12
+ dataset_split_name: validation
13
+ enable_xformers_memory_efficient_attention: false
14
+ gradient_accumulation_steps: 1
15
+ gradient_checkpointing: true
16
+ hub_model_id: null
17
+ hub_token: null
18
+ learning_rate: 1.0e-05
19
+ local_rank: -1
20
+ logging_dir: logs
21
+ loss_type: sigmoid
22
+ lr_num_cycles: 1
23
+ lr_power: 1.0
24
+ lr_scheduler: constant
25
+ lr_warmup_steps: 0
26
+ max_grad_norm: 1.0
27
+ max_train_samples: null
28
+ max_train_steps: 10000
29
+ mixed_precision: fp16
30
+ no_hflip: false
31
+ num_train_epochs: 20
32
+ output_dir: diffusion-dpo
33
+ pretrained_model_name_or_path: runwayml/stable-diffusion-v1-5
34
+ prior_generation_precision: null
35
+ push_to_hub: true
36
+ random_crop: false
37
+ rank: 8
38
+ report_to: tensorboard
39
+ resolution: 512
40
+ resume_from_checkpoint: null
41
+ revision: null
42
+ run_validation: true
43
+ scale_lr: false
44
+ seed: 0
45
+ tracker_name: diffusion-dpo-lora
46
+ train_batch_size: 1
47
+ use_8bit_adam: true
48
+ vae_encode_batch_size: 8
49
+ validation_steps: 200
50
+ variant: null
logs/diffusion-dpo-lora/1708700969.3730564/events.out.tfevents.1708700969.4f9809b-lcedt.2215537.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c90b297bb3974aadd7ffa6a469c614c8faf1d6795cd9354da685c07e7bf03132
3
+ size 2383
logs/diffusion-dpo-lora/1708700969.375287/hparams.yml ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 0.01
5
+ allow_tf32: false
6
+ beta_dpo: 2500
7
+ cache_dir: null
8
+ checkpointing_steps: 2000
9
+ checkpoints_total_limit: null
10
+ dataloader_num_workers: 0
11
+ dataset_name: kashif/pickascore
12
+ dataset_split_name: validation
13
+ enable_xformers_memory_efficient_attention: false
14
+ gradient_accumulation_steps: 1
15
+ gradient_checkpointing: true
16
+ hub_model_id: null
17
+ hub_token: null
18
+ learning_rate: 1.0e-05
19
+ local_rank: -1
20
+ logging_dir: logs
21
+ loss_type: sigmoid
22
+ lr_num_cycles: 1
23
+ lr_power: 1.0
24
+ lr_scheduler: constant
25
+ lr_warmup_steps: 0
26
+ max_grad_norm: 1.0
27
+ max_train_samples: null
28
+ max_train_steps: 10000
29
+ mixed_precision: fp16
30
+ no_hflip: false
31
+ num_train_epochs: 20
32
+ output_dir: diffusion-dpo
33
+ pretrained_model_name_or_path: runwayml/stable-diffusion-v1-5
34
+ prior_generation_precision: null
35
+ push_to_hub: true
36
+ random_crop: false
37
+ rank: 8
38
+ report_to: tensorboard
39
+ resolution: 512
40
+ resume_from_checkpoint: null
41
+ revision: null
42
+ run_validation: true
43
+ scale_lr: false
44
+ seed: 0
45
+ tracker_name: diffusion-dpo-lora
46
+ train_batch_size: 1
47
+ use_8bit_adam: true
48
+ vae_encode_batch_size: 8
49
+ validation_steps: 200
50
+ variant: null
logs/diffusion-dpo-lora/1708701303.0831516/events.out.tfevents.1708701303.4f9809b-lcedt.2219311.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef14ff50c89be98ae6667637b5364d50e4d416762aed6faa1cb04e647f61f444
3
+ size 2383
logs/diffusion-dpo-lora/1708701303.0857437/hparams.yml ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 0.01
5
+ allow_tf32: false
6
+ beta_dpo: 2500
7
+ cache_dir: null
8
+ checkpointing_steps: 2000
9
+ checkpoints_total_limit: null
10
+ dataloader_num_workers: 0
11
+ dataset_name: kashif/pickascore
12
+ dataset_split_name: validation
13
+ enable_xformers_memory_efficient_attention: false
14
+ gradient_accumulation_steps: 1
15
+ gradient_checkpointing: true
16
+ hub_model_id: null
17
+ hub_token: null
18
+ learning_rate: 1.0e-05
19
+ local_rank: -1
20
+ logging_dir: logs
21
+ loss_type: sigmoid
22
+ lr_num_cycles: 1
23
+ lr_power: 1.0
24
+ lr_scheduler: constant
25
+ lr_warmup_steps: 0
26
+ max_grad_norm: 1.0
27
+ max_train_samples: null
28
+ max_train_steps: 10000
29
+ mixed_precision: fp16
30
+ no_hflip: false
31
+ num_train_epochs: 20
32
+ output_dir: diffusion-dpo
33
+ pretrained_model_name_or_path: runwayml/stable-diffusion-v1-5
34
+ prior_generation_precision: null
35
+ push_to_hub: true
36
+ random_crop: false
37
+ rank: 8
38
+ report_to: tensorboard
39
+ resolution: 512
40
+ resume_from_checkpoint: null
41
+ revision: null
42
+ run_validation: true
43
+ scale_lr: false
44
+ seed: 0
45
+ tracker_name: diffusion-dpo-lora
46
+ train_batch_size: 1
47
+ use_8bit_adam: true
48
+ vae_encode_batch_size: 8
49
+ validation_steps: 200
50
+ variant: null
logs/diffusion-dpo-lora/1708701508.6600106/events.out.tfevents.1708701508.4f9809b-lcedt.2221662.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1dc19aa90946d19e0cf0ef796713d00947976bba8b880052d0cb2d092f1beb2c
3
+ size 2383
logs/diffusion-dpo-lora/1708701508.6627216/hparams.yml ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 0.01
5
+ allow_tf32: false
6
+ beta_dpo: 2500
7
+ cache_dir: null
8
+ checkpointing_steps: 2000
9
+ checkpoints_total_limit: null
10
+ dataloader_num_workers: 0
11
+ dataset_name: kashif/pickascore
12
+ dataset_split_name: validation
13
+ enable_xformers_memory_efficient_attention: false
14
+ gradient_accumulation_steps: 1
15
+ gradient_checkpointing: true
16
+ hub_model_id: null
17
+ hub_token: null
18
+ learning_rate: 1.0e-05
19
+ local_rank: -1
20
+ logging_dir: logs
21
+ loss_type: sigmoid
22
+ lr_num_cycles: 1
23
+ lr_power: 1.0
24
+ lr_scheduler: constant
25
+ lr_warmup_steps: 0
26
+ max_grad_norm: 1.0
27
+ max_train_samples: null
28
+ max_train_steps: 10000
29
+ mixed_precision: fp16
30
+ no_hflip: false
31
+ num_train_epochs: 20
32
+ output_dir: diffusion-dpo
33
+ pretrained_model_name_or_path: runwayml/stable-diffusion-v1-5
34
+ prior_generation_precision: null
35
+ push_to_hub: true
36
+ random_crop: false
37
+ rank: 8
38
+ report_to: tensorboard
39
+ resolution: 512
40
+ resume_from_checkpoint: null
41
+ revision: null
42
+ run_validation: true
43
+ scale_lr: false
44
+ seed: 0
45
+ tracker_name: diffusion-dpo-lora
46
+ train_batch_size: 1
47
+ use_8bit_adam: true
48
+ vae_encode_batch_size: 8
49
+ validation_steps: 200
50
+ variant: null
logs/diffusion-dpo-lora/1708701834.8009768/events.out.tfevents.1708701834.4f9809b-lcedt.2225371.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05b35a5f9012ad410631d8e6a705475c4f9eeaedcfaca66c2d0d8ef43280c998
3
+ size 2383
logs/diffusion-dpo-lora/1708701834.8036294/hparams.yml ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 0.01
5
+ allow_tf32: false
6
+ beta_dpo: 2500
7
+ cache_dir: null
8
+ checkpointing_steps: 2000
9
+ checkpoints_total_limit: null
10
+ dataloader_num_workers: 0
11
+ dataset_name: kashif/pickascore
12
+ dataset_split_name: validation
13
+ enable_xformers_memory_efficient_attention: false
14
+ gradient_accumulation_steps: 1
15
+ gradient_checkpointing: true
16
+ hub_model_id: null
17
+ hub_token: null
18
+ learning_rate: 1.0e-05
19
+ local_rank: -1
20
+ logging_dir: logs
21
+ loss_type: sigmoid
22
+ lr_num_cycles: 1
23
+ lr_power: 1.0
24
+ lr_scheduler: constant
25
+ lr_warmup_steps: 0
26
+ max_grad_norm: 1.0
27
+ max_train_samples: null
28
+ max_train_steps: 10000
29
+ mixed_precision: fp16
30
+ no_hflip: false
31
+ num_train_epochs: 20
32
+ output_dir: diffusion-dpo
33
+ pretrained_model_name_or_path: runwayml/stable-diffusion-v1-5
34
+ prior_generation_precision: null
35
+ push_to_hub: true
36
+ random_crop: false
37
+ rank: 8
38
+ report_to: tensorboard
39
+ resolution: 512
40
+ resume_from_checkpoint: null
41
+ revision: null
42
+ run_validation: true
43
+ scale_lr: false
44
+ seed: 0
45
+ tracker_name: diffusion-dpo-lora
46
+ train_batch_size: 1
47
+ use_8bit_adam: true
48
+ vae_encode_batch_size: 8
49
+ validation_steps: 200
50
+ variant: null
logs/diffusion-dpo-lora/1708701933.5214443/events.out.tfevents.1708701933.4f9809b-lcedt.2226604.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e5d5ed7423908b6f0f64a019791ca5d145788d1bdeb149e8dff3233eef41e842
3
+ size 2383
logs/diffusion-dpo-lora/1708701933.524116/hparams.yml ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 0.01
5
+ allow_tf32: false
6
+ beta_dpo: 2500
7
+ cache_dir: null
8
+ checkpointing_steps: 2000
9
+ checkpoints_total_limit: null
10
+ dataloader_num_workers: 0
11
+ dataset_name: kashif/pickascore
12
+ dataset_split_name: validation
13
+ enable_xformers_memory_efficient_attention: false
14
+ gradient_accumulation_steps: 1
15
+ gradient_checkpointing: true
16
+ hub_model_id: null
17
+ hub_token: null
18
+ learning_rate: 1.0e-05
19
+ local_rank: -1
20
+ logging_dir: logs
21
+ loss_type: sigmoid
22
+ lr_num_cycles: 1
23
+ lr_power: 1.0
24
+ lr_scheduler: constant
25
+ lr_warmup_steps: 0
26
+ max_grad_norm: 1.0
27
+ max_train_samples: null
28
+ max_train_steps: 10000
29
+ mixed_precision: fp16
30
+ no_hflip: false
31
+ num_train_epochs: 20
32
+ output_dir: diffusion-dpo
33
+ pretrained_model_name_or_path: runwayml/stable-diffusion-v1-5
34
+ prior_generation_precision: null
35
+ push_to_hub: true
36
+ random_crop: false
37
+ rank: 8
38
+ report_to: tensorboard
39
+ resolution: 512
40
+ resume_from_checkpoint: null
41
+ revision: null
42
+ run_validation: true
43
+ scale_lr: false
44
+ seed: 0
45
+ tracker_name: diffusion-dpo-lora
46
+ train_batch_size: 1
47
+ use_8bit_adam: true
48
+ vae_encode_batch_size: 8
49
+ validation_steps: 200
50
+ variant: null
logs/diffusion-dpo-lora/1708702436.6012197/events.out.tfevents.1708702436.4f9809b-lcedt.2232778.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1cb486d6a75d891cd75c1fb058ab21b9b606615bafff2ac6e7c6f5dd7ca96c1
3
+ size 2383
logs/diffusion-dpo-lora/1708702436.603921/hparams.yml ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 0.01
5
+ allow_tf32: false
6
+ beta_dpo: 2500
7
+ cache_dir: null
8
+ checkpointing_steps: 2000
9
+ checkpoints_total_limit: null
10
+ dataloader_num_workers: 0
11
+ dataset_name: kashif/pickascore
12
+ dataset_split_name: validation
13
+ enable_xformers_memory_efficient_attention: false
14
+ gradient_accumulation_steps: 1
15
+ gradient_checkpointing: true
16
+ hub_model_id: null
17
+ hub_token: null
18
+ learning_rate: 1.0e-05
19
+ local_rank: -1
20
+ logging_dir: logs
21
+ loss_type: sigmoid
22
+ lr_num_cycles: 1
23
+ lr_power: 1.0
24
+ lr_scheduler: constant
25
+ lr_warmup_steps: 0
26
+ max_grad_norm: 1.0
27
+ max_train_samples: null
28
+ max_train_steps: 10000
29
+ mixed_precision: fp16
30
+ no_hflip: false
31
+ num_train_epochs: 20
32
+ output_dir: diffusion-dpo
33
+ pretrained_model_name_or_path: runwayml/stable-diffusion-v1-5
34
+ prior_generation_precision: null
35
+ push_to_hub: true
36
+ random_crop: false
37
+ rank: 8
38
+ report_to: tensorboard
39
+ resolution: 512
40
+ resume_from_checkpoint: null
41
+ revision: null
42
+ run_validation: true
43
+ scale_lr: false
44
+ seed: 0
45
+ tracker_name: diffusion-dpo-lora
46
+ train_batch_size: 1
47
+ use_8bit_adam: true
48
+ vae_encode_batch_size: 8
49
+ validation_steps: 200
50
+ variant: null
logs/diffusion-dpo-lora/1708702585.2958512/events.out.tfevents.1708702585.4f9809b-lcedt.2235338.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6209719cc8a9b59e2eb4aea5b7fb3579272ced8ffcee9c75a971bb99d7caf6aa
3
+ size 2383