damian0815
commited on
Commit
•
f8b4557
1
Parent(s):
01b94ea
Upload folder using huggingface_hub
Browse files- ff7r_4e-5cosine_everything-cont-20-ep50-3e-7_bs1-20241228-170700.log +37 -0
- ff7r_4e-5cosine_everything-cont-20-ep50-3e-7_bs1-20241228-170700_main.json +149 -0
- ff7r_4e-5cosine_everything-cont-20-ep50-3e-7_bs1-20241228-170700_opt.json +32 -0
- model_index.json +38 -0
- scheduler/scheduler_config.json +20 -0
- text_encoder/config.json +25 -0
- text_encoder/model.safetensors +3 -0
- tokenizer/merges.txt +0 -0
- tokenizer/special_tokens_map.json +30 -0
- tokenizer/tokenizer_config.json +31 -0
- tokenizer/vocab.json +0 -0
- unet/config.json +68 -0
- unet/diffusion_pytorch_model.safetensors +3 -0
- vae/config.json +38 -0
- vae/diffusion_pytorch_model.safetensors +3 -0
ff7r_4e-5cosine_everything-cont-20-ep50-3e-7_bs1-20241228-170700.log
ADDED
@@ -0,0 +1,37 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
12/28/2024 05:07:00 PM [93m Disabling AMP, not recommended.[0m
|
2 |
+
12/28/2024 05:07:00 PM Seed: 756498848
|
3 |
+
12/28/2024 05:07:00 PM unet attention_head_dim: 8
|
4 |
+
12/28/2024 05:07:00 PM Inferred yaml: v1-inference.yaml, attn: sd1, prediction_type: epsilon
|
5 |
+
12/28/2024 05:07:01 PM * Using default (DDPM) noise scheduler for training: ddpm
|
6 |
+
12/28/2024 05:07:01 PM * Using SDP attention *
|
7 |
+
12/28/2024 05:07:01 PM * DLMA resolution 512, buckets: [[512, 512], [576, 448], [448, 576], [640, 384], [384, 640], [768, 320], [320, 768], [896, 256], [256, 896], [1024, 256], [256, 1024]]
|
8 |
+
12/28/2024 05:07:01 PM Preloading images...
|
9 |
+
12/28/2024 05:07:06 PM * Loaded 40 validation images for validation set 'val' from ../data/val
|
10 |
+
12/28/2024 05:07:06 PM * DLMA initialized with 40 images.
|
11 |
+
12/28/2024 05:07:06 PM ** Dataset 'val': 40 batches, num_images: 40, batch_size: 1
|
12 |
+
12/28/2024 05:07:06 PM * [91mAspect ratio bucket ('default_batch', 768, 320) has only 1 images[0m. At batch size 1 this makes for an effective multiplier of 2.0, which may cause problems. Consider adding 1 or more images with aspect ratio 12:5, or reducing your batch_size.
|
13 |
+
12/28/2024 05:07:06 PM - Plugin plugins.interruptible.InterruptiblePlugin loaded to <class 'plugins.interruptible.InterruptiblePlugin'>
|
14 |
+
12/28/2024 05:07:06 PM * DLMA initialized with 1913 images.
|
15 |
+
12/28/2024 05:07:06 PM ** Dataset 'train': 1913 batches, num_images: 1913, batch_size: 1
|
16 |
+
12/28/2024 05:07:06 PM [36m * unet optimizer: AdamW (686 parameters) *[0m
|
17 |
+
12/28/2024 05:07:06 PM [36m lr: 3e-07, betas: [0.9, 0.99], epsilon: 1e-08, weight_decay: 0.01 *[0m
|
18 |
+
12/28/2024 05:07:06 PM Grad scaler enabled: False (amp mode)
|
19 |
+
12/28/2024 05:07:06 PM Pretraining GPU Memory: 5102 / 24564 MB
|
20 |
+
12/28/2024 05:07:06 PM saving ckpts every 1000000000.0 minutes
|
21 |
+
12/28/2024 05:07:06 PM saving ckpts every 4 epochs
|
22 |
+
12/28/2024 05:07:06 PM unet device: cuda:0, precision: torch.float32, training: True
|
23 |
+
12/28/2024 05:07:06 PM text_encoder device: cuda:0, precision: torch.float32, training: False
|
24 |
+
12/28/2024 05:07:06 PM vae device: cuda:0, precision: torch.float32, training: False
|
25 |
+
12/28/2024 05:07:06 PM scheduler: <class 'diffusers.schedulers.scheduling_ddpm.DDPMScheduler'>
|
26 |
+
12/28/2024 05:07:06 PM [32mProject name: [0m[92mff7r_4e-5cosine_everything-cont-20-ep50-3e-7_bs1[0m
|
27 |
+
12/28/2024 05:07:06 PM [32mgrad_accum: [0m[92m1[0m
|
28 |
+
12/28/2024 05:07:06 PM [32mbatch_size: [0m[92m1[0m
|
29 |
+
12/28/2024 05:07:06 PM [32mepoch_len: [92m1913[0m
|
30 |
+
12/28/2024 05:45:02 PM * Saving diffusers model to ../logs/ff7r_4e-5cosine_everything-cont-20-ep50-3e-7_bs1-20241228-170700/ckpts/ff7r_4e-5cosine_everything-cont-20-ep50-3e-7_bs1-ep04-gs07652
|
31 |
+
12/28/2024 05:45:04 PM Saving optimizer state to ../logs/ff7r_4e-5cosine_everything-cont-20-ep50-3e-7_bs1-20241228-170700/ckpts/ff7r_4e-5cosine_everything-cont-20-ep50-3e-7_bs1-ep04-gs07652
|
32 |
+
12/28/2024 05:45:08 PM [36mTraining complete[0m
|
33 |
+
12/28/2024 05:45:08 PM Total training time took 38.03 minutes, total steps: 7652
|
34 |
+
12/28/2024 05:45:08 PM Average epoch time: 9.46 minutes
|
35 |
+
12/28/2024 05:45:08 PM [97m ***************************[0m
|
36 |
+
12/28/2024 05:45:08 PM [97m **** Finished training ****[0m
|
37 |
+
12/28/2024 05:45:08 PM [97m ***************************[0m
|
ff7r_4e-5cosine_everything-cont-20-ep50-3e-7_bs1-20241228-170700_main.json
ADDED
@@ -0,0 +1,149 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"config": "ff7r.json",
|
3 |
+
"batch_size": 1,
|
4 |
+
"ckpt_every_n_minutes": 1000000000.0,
|
5 |
+
"clip_grad_norm": 1000000000.0,
|
6 |
+
"clip_skip": 0,
|
7 |
+
"contrastive_learning_batch_ids": [],
|
8 |
+
"contrastive_learning_max_negative_loss": 1.0,
|
9 |
+
"contrastive_learning_negative_loss_scale": 0.1,
|
10 |
+
"everything_contrastive_learning_p": 0.0,
|
11 |
+
"contrastive_learning_delta_loss_method": true,
|
12 |
+
"max_backward_accum": 5,
|
13 |
+
"init_grad_scale": 11500,
|
14 |
+
"auto_decay_steps_multiplier": 1.001,
|
15 |
+
"cond_dropout": 0.02,
|
16 |
+
"data_root": "../data/train",
|
17 |
+
"disable_amp": true,
|
18 |
+
"disable_textenc_training": true,
|
19 |
+
"disable_unet_training": false,
|
20 |
+
"disable_xformers": false,
|
21 |
+
"flip_p": 1.0,
|
22 |
+
"gpuid": 0,
|
23 |
+
"gradient_checkpointing": true,
|
24 |
+
"grad_accum": 1,
|
25 |
+
"logdir": "../logs",
|
26 |
+
"log_step": 25,
|
27 |
+
"lowvram": false,
|
28 |
+
"max_epochs": 4,
|
29 |
+
"notebook": false,
|
30 |
+
"optimizer_config": "optimizer-unet-3e7lin-nowarmup-adamw.json",
|
31 |
+
"plugins": [
|
32 |
+
"plugins.interruptible.InterruptiblePlugin"
|
33 |
+
],
|
34 |
+
"project_name": "ff7r_4e-5cosine_everything-cont-20-ep50-3e-7_bs1",
|
35 |
+
"resolution": 512,
|
36 |
+
"resume_ckpt": "../logs/ff7r_4e-5cosine_everything-cont-20-20241228-105146/ckpts/ff7r_4e-5cosine_everything-cont-20-ep50-gs16100",
|
37 |
+
"run_name": null,
|
38 |
+
"sample_prompts": "sample_prompts.json",
|
39 |
+
"sample_steps": 666,
|
40 |
+
"save_ckpt_dir": null,
|
41 |
+
"save_ckpts_from_n_epochs": 0,
|
42 |
+
"save_every_n_epochs": 4,
|
43 |
+
"save_optimizer": true,
|
44 |
+
"scale_lr": false,
|
45 |
+
"seed": 756498848,
|
46 |
+
"shuffle_tags": false,
|
47 |
+
"validation_config": "validation.json",
|
48 |
+
"wandb": false,
|
49 |
+
"write_schedule": false,
|
50 |
+
"rated_dataset": false,
|
51 |
+
"rated_dataset_target_dropout_percent": 50,
|
52 |
+
"zero_frequency_noise_ratio": 0.0,
|
53 |
+
"enable_zero_terminal_snr": false,
|
54 |
+
"load_settings_every_epoch": false,
|
55 |
+
"min_snr_gamma": 1.5,
|
56 |
+
"loss_type": "mse_huber",
|
57 |
+
"ema_decay_rate": null,
|
58 |
+
"ema_strength_target": null,
|
59 |
+
"ema_update_interval": null,
|
60 |
+
"ema_device": null,
|
61 |
+
"ema_sample_nonema_model": false,
|
62 |
+
"ema_sample_ema_model": false,
|
63 |
+
"ema_resume_model": null,
|
64 |
+
"no_save_ckpt": true,
|
65 |
+
"no_prepend_last": true,
|
66 |
+
"amp": false,
|
67 |
+
"attn_type": "sdp",
|
68 |
+
"embedding_perturbation": 1.0,
|
69 |
+
"latents_perturbation": 0.0,
|
70 |
+
"log_huber": false,
|
71 |
+
"save_full_precision": false,
|
72 |
+
"timestep_start": 0,
|
73 |
+
"timestep_end": 1000,
|
74 |
+
"train_sampler": "ddpm",
|
75 |
+
"keep_tags": 0,
|
76 |
+
"pyramid_noise_discount": null,
|
77 |
+
"lr": null,
|
78 |
+
"lr_decay_steps": 7659,
|
79 |
+
"lr_scheduler": "constant",
|
80 |
+
"lr_warmup_steps": 153,
|
81 |
+
"no_initial_validation": false,
|
82 |
+
"mix_zero_terminal_snr": null,
|
83 |
+
"match_zero_terminal_snr": null,
|
84 |
+
"debug_invert_min_snr_gamma": false,
|
85 |
+
"batch_share_noise": false,
|
86 |
+
"batch_share_timesteps": false,
|
87 |
+
"teacher": null,
|
88 |
+
"teacher_is_sdxl": false,
|
89 |
+
"teacher_loss_scale": 1,
|
90 |
+
"contrastive_learning_use_l1_loss": false,
|
91 |
+
"contrastive_learning_no_average_negatives": false,
|
92 |
+
"contrastive_learning_save_on_cpu": false,
|
93 |
+
"contrastive_learning_delta_timestep_start": 150,
|
94 |
+
"contrastive_learning_dropout_p": 0,
|
95 |
+
"batch_id_dropout_p": 0,
|
96 |
+
"cond_dropout_noise_p": 0,
|
97 |
+
"jacobian_descent": false,
|
98 |
+
"use_masks": false,
|
99 |
+
"lora": false,
|
100 |
+
"lora_resume": null,
|
101 |
+
"lora_rank": 16,
|
102 |
+
"lora_alpha": 8,
|
103 |
+
"aspects": [
|
104 |
+
[
|
105 |
+
512,
|
106 |
+
512
|
107 |
+
],
|
108 |
+
[
|
109 |
+
576,
|
110 |
+
448
|
111 |
+
],
|
112 |
+
[
|
113 |
+
448,
|
114 |
+
576
|
115 |
+
],
|
116 |
+
[
|
117 |
+
640,
|
118 |
+
384
|
119 |
+
],
|
120 |
+
[
|
121 |
+
384,
|
122 |
+
640
|
123 |
+
],
|
124 |
+
[
|
125 |
+
768,
|
126 |
+
320
|
127 |
+
],
|
128 |
+
[
|
129 |
+
320,
|
130 |
+
768
|
131 |
+
],
|
132 |
+
[
|
133 |
+
896,
|
134 |
+
256
|
135 |
+
],
|
136 |
+
[
|
137 |
+
256,
|
138 |
+
896
|
139 |
+
],
|
140 |
+
[
|
141 |
+
1024,
|
142 |
+
256
|
143 |
+
],
|
144 |
+
[
|
145 |
+
256,
|
146 |
+
1024
|
147 |
+
]
|
148 |
+
]
|
149 |
+
}
|
ff7r_4e-5cosine_everything-cont-20-ep50-3e-7_bs1-20241228-170700_opt.json
ADDED
@@ -0,0 +1,32 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"base": {
|
3 |
+
"optimizer": "adamw",
|
4 |
+
"lr": 3e-07,
|
5 |
+
"lr_scheduler": "linear",
|
6 |
+
"lr_decay_steps": 7659,
|
7 |
+
"lr_warmup_steps": 1,
|
8 |
+
"betas": [
|
9 |
+
0.9,
|
10 |
+
0.99
|
11 |
+
],
|
12 |
+
"epsilon": 1e-08,
|
13 |
+
"weight_decay": 0.01
|
14 |
+
},
|
15 |
+
"text_encoder_overrides": {
|
16 |
+
"optimizer": "adamw",
|
17 |
+
"lr": 1e-06,
|
18 |
+
"lr_scheduler": "linear",
|
19 |
+
"lr_decay_steps": 7659,
|
20 |
+
"lr_warmup_steps": 1,
|
21 |
+
"betas": [
|
22 |
+
0.9,
|
23 |
+
0.99
|
24 |
+
],
|
25 |
+
"epsilon": 1e-08,
|
26 |
+
"weight_decay": 0.01
|
27 |
+
},
|
28 |
+
"text_encoder_freezing": {
|
29 |
+
"unfreeze_last_n_layers": 2
|
30 |
+
},
|
31 |
+
"apply_grad_scaler_step_tweaks": false
|
32 |
+
}
|
model_index.json
ADDED
@@ -0,0 +1,38 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_class_name": "StableDiffusionPipeline",
|
3 |
+
"_diffusers_version": "0.31.0",
|
4 |
+
"_name_or_path": "../logs/ff7r_4e-5cosine_everything-cont-20-ep50-3e-7_bs1-20241228-170700/ckpts/ff7r_4e-5cosine_everything-cont-20-ep50-3e-7_bs1-ep04-gs07652",
|
5 |
+
"feature_extractor": [
|
6 |
+
null,
|
7 |
+
null
|
8 |
+
],
|
9 |
+
"image_encoder": [
|
10 |
+
null,
|
11 |
+
null
|
12 |
+
],
|
13 |
+
"requires_safety_checker": null,
|
14 |
+
"safety_checker": [
|
15 |
+
null,
|
16 |
+
null
|
17 |
+
],
|
18 |
+
"scheduler": [
|
19 |
+
"diffusers",
|
20 |
+
"DDIMScheduler"
|
21 |
+
],
|
22 |
+
"text_encoder": [
|
23 |
+
"transformers",
|
24 |
+
"CLIPTextModel"
|
25 |
+
],
|
26 |
+
"tokenizer": [
|
27 |
+
"transformers",
|
28 |
+
"CLIPTokenizer"
|
29 |
+
],
|
30 |
+
"unet": [
|
31 |
+
"diffusers",
|
32 |
+
"UNet2DConditionModel"
|
33 |
+
],
|
34 |
+
"vae": [
|
35 |
+
"diffusers",
|
36 |
+
"AutoencoderKL"
|
37 |
+
]
|
38 |
+
}
|
scheduler/scheduler_config.json
ADDED
@@ -0,0 +1,20 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_class_name": "DDIMScheduler",
|
3 |
+
"_diffusers_version": "0.31.0",
|
4 |
+
"beta_end": 0.012,
|
5 |
+
"beta_schedule": "scaled_linear",
|
6 |
+
"beta_start": 0.00085,
|
7 |
+
"clip_sample": false,
|
8 |
+
"clip_sample_range": 1.0,
|
9 |
+
"dynamic_thresholding_ratio": 0.995,
|
10 |
+
"num_train_timesteps": 1000,
|
11 |
+
"prediction_type": "epsilon",
|
12 |
+
"rescale_betas_zero_snr": false,
|
13 |
+
"sample_max_value": 1.0,
|
14 |
+
"set_alpha_to_one": false,
|
15 |
+
"skip_prk_steps": true,
|
16 |
+
"steps_offset": 1,
|
17 |
+
"thresholding": false,
|
18 |
+
"timestep_spacing": "leading",
|
19 |
+
"trained_betas": null
|
20 |
+
}
|
text_encoder/config.json
ADDED
@@ -0,0 +1,25 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "../logs/ff7r_4e-5cosine_everything-cont-20-ep50-3e-7_bs1-20241228-170700/ckpts/ff7r_4e-5cosine_everything-cont-20-ep50-3e-7_bs1-ep04-gs07652/text_encoder",
|
3 |
+
"architectures": [
|
4 |
+
"CLIPTextModel"
|
5 |
+
],
|
6 |
+
"attention_dropout": 0.0,
|
7 |
+
"bos_token_id": 0,
|
8 |
+
"dropout": 0.0,
|
9 |
+
"eos_token_id": 2,
|
10 |
+
"hidden_act": "quick_gelu",
|
11 |
+
"hidden_size": 768,
|
12 |
+
"initializer_factor": 1.0,
|
13 |
+
"initializer_range": 0.02,
|
14 |
+
"intermediate_size": 3072,
|
15 |
+
"layer_norm_eps": 1e-05,
|
16 |
+
"max_position_embeddings": 77,
|
17 |
+
"model_type": "clip_text_model",
|
18 |
+
"num_attention_heads": 12,
|
19 |
+
"num_hidden_layers": 12,
|
20 |
+
"pad_token_id": 1,
|
21 |
+
"projection_dim": 768,
|
22 |
+
"torch_dtype": "float16",
|
23 |
+
"transformers_version": "4.38.2",
|
24 |
+
"vocab_size": 49408
|
25 |
+
}
|
text_encoder/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:660c6f5b1abae9dc498ac2d21e1347d2abdb0cf6c0c0c8576cd796491d9a6cdd
|
3 |
+
size 246144152
|
tokenizer/merges.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer/special_tokens_map.json
ADDED
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": {
|
3 |
+
"content": "<|startoftext|>",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": true,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"eos_token": {
|
10 |
+
"content": "<|endoftext|>",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": true,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"pad_token": {
|
17 |
+
"content": "<|endoftext|>",
|
18 |
+
"lstrip": false,
|
19 |
+
"normalized": true,
|
20 |
+
"rstrip": false,
|
21 |
+
"single_word": false
|
22 |
+
},
|
23 |
+
"unk_token": {
|
24 |
+
"content": "<|endoftext|>",
|
25 |
+
"lstrip": false,
|
26 |
+
"normalized": true,
|
27 |
+
"rstrip": false,
|
28 |
+
"single_word": false
|
29 |
+
}
|
30 |
+
}
|
tokenizer/tokenizer_config.json
ADDED
@@ -0,0 +1,31 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_prefix_space": false,
|
3 |
+
"added_tokens_decoder": {
|
4 |
+
"49406": {
|
5 |
+
"content": "<|startoftext|>",
|
6 |
+
"lstrip": false,
|
7 |
+
"normalized": true,
|
8 |
+
"rstrip": false,
|
9 |
+
"single_word": false,
|
10 |
+
"special": true
|
11 |
+
},
|
12 |
+
"49407": {
|
13 |
+
"content": "<|endoftext|>",
|
14 |
+
"lstrip": false,
|
15 |
+
"normalized": true,
|
16 |
+
"rstrip": false,
|
17 |
+
"single_word": false,
|
18 |
+
"special": true
|
19 |
+
}
|
20 |
+
},
|
21 |
+
"bos_token": "<|startoftext|>",
|
22 |
+
"clean_up_tokenization_spaces": true,
|
23 |
+
"do_lower_case": true,
|
24 |
+
"eos_token": "<|endoftext|>",
|
25 |
+
"errors": "replace",
|
26 |
+
"model_max_length": 77,
|
27 |
+
"pad_token": "<|endoftext|>",
|
28 |
+
"tokenizer_class": "CLIPTokenizer",
|
29 |
+
"unk_token": "<|endoftext|>",
|
30 |
+
"use_fast": false
|
31 |
+
}
|
tokenizer/vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
unet/config.json
ADDED
@@ -0,0 +1,68 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_class_name": "UNet2DConditionModel",
|
3 |
+
"_diffusers_version": "0.31.0",
|
4 |
+
"_name_or_path": "../logs/ff7r_4e-5cosine_everything-cont-20-ep50-3e-7_bs1-20241228-170700/ckpts/ff7r_4e-5cosine_everything-cont-20-ep50-3e-7_bs1-ep04-gs07652/unet",
|
5 |
+
"act_fn": "silu",
|
6 |
+
"addition_embed_type": null,
|
7 |
+
"addition_embed_type_num_heads": 64,
|
8 |
+
"addition_time_embed_dim": null,
|
9 |
+
"attention_head_dim": 8,
|
10 |
+
"attention_type": "default",
|
11 |
+
"block_out_channels": [
|
12 |
+
320,
|
13 |
+
640,
|
14 |
+
1280,
|
15 |
+
1280
|
16 |
+
],
|
17 |
+
"center_input_sample": false,
|
18 |
+
"class_embed_type": null,
|
19 |
+
"class_embeddings_concat": false,
|
20 |
+
"conv_in_kernel": 3,
|
21 |
+
"conv_out_kernel": 3,
|
22 |
+
"cross_attention_dim": 768,
|
23 |
+
"cross_attention_norm": null,
|
24 |
+
"down_block_types": [
|
25 |
+
"CrossAttnDownBlock2D",
|
26 |
+
"CrossAttnDownBlock2D",
|
27 |
+
"CrossAttnDownBlock2D",
|
28 |
+
"DownBlock2D"
|
29 |
+
],
|
30 |
+
"downsample_padding": 1,
|
31 |
+
"dropout": 0.0,
|
32 |
+
"dual_cross_attention": false,
|
33 |
+
"encoder_hid_dim": null,
|
34 |
+
"encoder_hid_dim_type": null,
|
35 |
+
"flip_sin_to_cos": true,
|
36 |
+
"freq_shift": 0,
|
37 |
+
"in_channels": 4,
|
38 |
+
"layers_per_block": 2,
|
39 |
+
"mid_block_only_cross_attention": null,
|
40 |
+
"mid_block_scale_factor": 1,
|
41 |
+
"mid_block_type": "UNetMidBlock2DCrossAttn",
|
42 |
+
"norm_eps": 1e-05,
|
43 |
+
"norm_num_groups": 32,
|
44 |
+
"num_attention_heads": null,
|
45 |
+
"num_class_embeds": null,
|
46 |
+
"only_cross_attention": false,
|
47 |
+
"out_channels": 4,
|
48 |
+
"projection_class_embeddings_input_dim": null,
|
49 |
+
"resnet_out_scale_factor": 1.0,
|
50 |
+
"resnet_skip_time_act": false,
|
51 |
+
"resnet_time_scale_shift": "default",
|
52 |
+
"reverse_transformer_layers_per_block": null,
|
53 |
+
"sample_size": 64,
|
54 |
+
"time_cond_proj_dim": null,
|
55 |
+
"time_embedding_act_fn": null,
|
56 |
+
"time_embedding_dim": null,
|
57 |
+
"time_embedding_type": "positional",
|
58 |
+
"timestep_post_act": null,
|
59 |
+
"transformer_layers_per_block": 1,
|
60 |
+
"up_block_types": [
|
61 |
+
"UpBlock2D",
|
62 |
+
"CrossAttnUpBlock2D",
|
63 |
+
"CrossAttnUpBlock2D",
|
64 |
+
"CrossAttnUpBlock2D"
|
65 |
+
],
|
66 |
+
"upcast_attention": false,
|
67 |
+
"use_linear_projection": false
|
68 |
+
}
|
unet/diffusion_pytorch_model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a30d8eaa78c527fc9a84265856a8d50a3c6a0b6de670b22b647d2b2d3eb51b6c
|
3 |
+
size 1719125304
|
vae/config.json
ADDED
@@ -0,0 +1,38 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_class_name": "AutoencoderKL",
|
3 |
+
"_diffusers_version": "0.31.0",
|
4 |
+
"_name_or_path": "../logs/ff7r_4e-5cosine_everything-cont-20-ep50-3e-7_bs1-20241228-170700/ckpts/ff7r_4e-5cosine_everything-cont-20-ep50-3e-7_bs1-ep04-gs07652/vae",
|
5 |
+
"act_fn": "silu",
|
6 |
+
"block_out_channels": [
|
7 |
+
128,
|
8 |
+
256,
|
9 |
+
512,
|
10 |
+
512
|
11 |
+
],
|
12 |
+
"down_block_types": [
|
13 |
+
"DownEncoderBlock2D",
|
14 |
+
"DownEncoderBlock2D",
|
15 |
+
"DownEncoderBlock2D",
|
16 |
+
"DownEncoderBlock2D"
|
17 |
+
],
|
18 |
+
"force_upcast": true,
|
19 |
+
"in_channels": 3,
|
20 |
+
"latent_channels": 4,
|
21 |
+
"latents_mean": null,
|
22 |
+
"latents_std": null,
|
23 |
+
"layers_per_block": 2,
|
24 |
+
"mid_block_add_attention": true,
|
25 |
+
"norm_num_groups": 32,
|
26 |
+
"out_channels": 3,
|
27 |
+
"sample_size": 512,
|
28 |
+
"scaling_factor": 0.18215,
|
29 |
+
"shift_factor": null,
|
30 |
+
"up_block_types": [
|
31 |
+
"UpDecoderBlock2D",
|
32 |
+
"UpDecoderBlock2D",
|
33 |
+
"UpDecoderBlock2D",
|
34 |
+
"UpDecoderBlock2D"
|
35 |
+
],
|
36 |
+
"use_post_quant_conv": true,
|
37 |
+
"use_quant_conv": true
|
38 |
+
}
|
vae/diffusion_pytorch_model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3e4c08995484ee61270175e9e7a072b66a6e4eeb5f0c266667fe1f45b90daf9a
|
3 |
+
size 167335342
|