diff --git a/animatediff_models/animate_lcm_sd15_t2v_beta.ckpt b/animatediff_models/animate_lcm_sd15_t2v_beta.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..eb00ded3b83d2481b7f8d07fa2d055fddb695e94 --- /dev/null +++ b/animatediff_models/animate_lcm_sd15_t2v_beta.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b46c3de62e5696af72c4056e3cdcbea12fbc19581c0aad7b6f2b027851148f5f +size 1813041929 diff --git a/animatediff_models/hsxl_temporal_layers.f16.safetensors b/animatediff_models/hsxl_temporal_layers.f16.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c305203d001da3c588fbfb9145d0e7d371315b14 --- /dev/null +++ b/animatediff_models/hsxl_temporal_layers.f16.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b46ce6b4ec297b21e508b8888135f65b5a07aa76d755cc086dc5ee804064bb1c +size 474694696 diff --git a/animatediff_models/v3_sd15_mm.ckpt b/animatediff_models/v3_sd15_mm.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..93094234245e66268d2fa86def02f52817741eef --- /dev/null +++ b/animatediff_models/v3_sd15_mm.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2412711886f61091846f53204aabc38aa6e09356d62a9808abe4daa802168343 +size 1673262583 diff --git a/checkpoints/Juggernaut_RunDiffusionPhoto2_Lightning_4Steps.safetensors b/checkpoints/Juggernaut_RunDiffusionPhoto2_Lightning_4Steps.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c5c3a153de2810aecfcd05716a0c63400c937c60 --- /dev/null +++ b/checkpoints/Juggernaut_RunDiffusionPhoto2_Lightning_4Steps.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c8df560d2992ac04299412be6a36fa53a4e7a1b74f27b94867ad3f84f4b425a5 +size 7105348284 diff --git a/checkpoints/crystalClear2_crystalClear2V1.safetensors b/checkpoints/crystalClear2_crystalClear2V1.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2f680281bf75b929c8e6f043fe1ec3d1802bc06e --- /dev/null +++ b/checkpoints/crystalClear2_crystalClear2V1.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d5a9e7658a76db46e8a56605dc50fa8596cf50da741a2bb999f88a0794d5595a +size 4244098750 diff --git a/checkpoints/epicrealism_naturalSinRC1VAE.safetensors b/checkpoints/epicrealism_naturalSinRC1VAE.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1a1c20a4e478f7caecab4e1619aa9084979acb16 --- /dev/null +++ b/checkpoints/epicrealism_naturalSinRC1VAE.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:84d76a0328ee5a1a2a1d466ba7478ab59183e3fea385441528aa4c5d567fd43e +size 2132625612 diff --git a/checkpoints/photonLCM_v10.safetensors b/checkpoints/photonLCM_v10.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..540858b7523e0f89e2fb5b4a90dde33fd578f66d --- /dev/null +++ b/checkpoints/photonLCM_v10.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f2d45bbd0f7d28b7ff5e33cf5f7dca82ccf7a14fa55a867ec4dc7f701cb2859f +size 2133817176 diff --git a/checkpoints/put_checkpoints_here b/checkpoints/put_checkpoints_here new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/checkpoints/realisticVisionV51_v51VAE.safetensors b/checkpoints/realisticVisionV51_v51VAE.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..be041649019837e489d917b5371c46a6ba343ebf --- /dev/null +++ b/checkpoints/realisticVisionV51_v51VAE.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:15012c538f503ce2ebfc2c8547b268c75ccdaff7a281db55399940ff1d70e21d +size 2132625894 diff --git a/clip/put_clip_or_text_encoder_models_here b/clip/put_clip_or_text_encoder_models_here new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/clip_vision/CLIP-ViT-H-14-laion2B-s32B-b79K.safetensors b/clip_vision/CLIP-ViT-H-14-laion2B-s32B-b79K.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..590c45d8aba49bfa35bc489af660f732aab16568 --- /dev/null +++ b/clip_vision/CLIP-ViT-H-14-laion2B-s32B-b79K.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6ca9667da1ca9e0b0f75e46bb030f7e011f44f86cbfb8d5a36590fcd7507b030 +size 2528373448 diff --git a/clip_vision/ip-adapter_image_encoder_sd15.safetensors b/clip_vision/ip-adapter_image_encoder_sd15.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..590c45d8aba49bfa35bc489af660f732aab16568 --- /dev/null +++ b/clip_vision/ip-adapter_image_encoder_sd15.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6ca9667da1ca9e0b0f75e46bb030f7e011f44f86cbfb8d5a36590fcd7507b030 +size 2528373448 diff --git a/clip_vision/put_clip_vision_models_here b/clip_vision/put_clip_vision_models_here new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/configs/anything_v3.yaml b/configs/anything_v3.yaml new file mode 100644 index 0000000000000000000000000000000000000000..8bcfe584ae73d60e2c7a6f89b3f7befbd487ea34 --- /dev/null +++ b/configs/anything_v3.yaml @@ -0,0 +1,73 @@ +model: + base_learning_rate: 1.0e-04 + target: ldm.models.diffusion.ddpm.LatentDiffusion + params: + linear_start: 0.00085 + linear_end: 0.0120 + num_timesteps_cond: 1 + log_every_t: 200 + timesteps: 1000 + first_stage_key: "jpg" + cond_stage_key: "txt" + image_size: 64 + channels: 4 + cond_stage_trainable: false # Note: different from the one we trained before + conditioning_key: crossattn + monitor: val/loss_simple_ema + scale_factor: 0.18215 + use_ema: False + + scheduler_config: # 10000 warmup steps + target: ldm.lr_scheduler.LambdaLinearScheduler + params: + warm_up_steps: [ 10000 ] + cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases + f_start: [ 1.e-6 ] + f_max: [ 1. ] + f_min: [ 1. ] + + unet_config: + target: ldm.modules.diffusionmodules.openaimodel.UNetModel + params: + image_size: 32 # unused + in_channels: 4 + out_channels: 4 + model_channels: 320 + attention_resolutions: [ 4, 2, 1 ] + num_res_blocks: 2 + channel_mult: [ 1, 2, 4, 4 ] + num_heads: 8 + use_spatial_transformer: True + transformer_depth: 1 + context_dim: 768 + use_checkpoint: True + legacy: False + + first_stage_config: + target: ldm.models.autoencoder.AutoencoderKL + params: + embed_dim: 4 + monitor: val/rec_loss + ddconfig: + double_z: true + z_channels: 4 + resolution: 256 + in_channels: 3 + out_ch: 3 + ch: 128 + ch_mult: + - 1 + - 2 + - 4 + - 4 + num_res_blocks: 2 + attn_resolutions: [] + dropout: 0.0 + lossconfig: + target: torch.nn.Identity + + cond_stage_config: + target: ldm.modules.encoders.modules.FrozenCLIPEmbedder + params: + layer: "hidden" + layer_idx: -2 diff --git a/configs/v1-inference.yaml b/configs/v1-inference.yaml new file mode 100644 index 0000000000000000000000000000000000000000..d4effe569e897369918625f9d8be5603a0e6a0d6 --- /dev/null +++ b/configs/v1-inference.yaml @@ -0,0 +1,70 @@ +model: + base_learning_rate: 1.0e-04 + target: ldm.models.diffusion.ddpm.LatentDiffusion + params: + linear_start: 0.00085 + linear_end: 0.0120 + num_timesteps_cond: 1 + log_every_t: 200 + timesteps: 1000 + first_stage_key: "jpg" + cond_stage_key: "txt" + image_size: 64 + channels: 4 + cond_stage_trainable: false # Note: different from the one we trained before + conditioning_key: crossattn + monitor: val/loss_simple_ema + scale_factor: 0.18215 + use_ema: False + + scheduler_config: # 10000 warmup steps + target: ldm.lr_scheduler.LambdaLinearScheduler + params: + warm_up_steps: [ 10000 ] + cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases + f_start: [ 1.e-6 ] + f_max: [ 1. ] + f_min: [ 1. ] + + unet_config: + target: ldm.modules.diffusionmodules.openaimodel.UNetModel + params: + image_size: 32 # unused + in_channels: 4 + out_channels: 4 + model_channels: 320 + attention_resolutions: [ 4, 2, 1 ] + num_res_blocks: 2 + channel_mult: [ 1, 2, 4, 4 ] + num_heads: 8 + use_spatial_transformer: True + transformer_depth: 1 + context_dim: 768 + use_checkpoint: True + legacy: False + + first_stage_config: + target: ldm.models.autoencoder.AutoencoderKL + params: + embed_dim: 4 + monitor: val/rec_loss + ddconfig: + double_z: true + z_channels: 4 + resolution: 256 + in_channels: 3 + out_ch: 3 + ch: 128 + ch_mult: + - 1 + - 2 + - 4 + - 4 + num_res_blocks: 2 + attn_resolutions: [] + dropout: 0.0 + lossconfig: + target: torch.nn.Identity + + cond_stage_config: + target: ldm.modules.encoders.modules.FrozenCLIPEmbedder diff --git a/configs/v1-inference_clip_skip_2.yaml b/configs/v1-inference_clip_skip_2.yaml new file mode 100644 index 0000000000000000000000000000000000000000..8bcfe584ae73d60e2c7a6f89b3f7befbd487ea34 --- /dev/null +++ b/configs/v1-inference_clip_skip_2.yaml @@ -0,0 +1,73 @@ +model: + base_learning_rate: 1.0e-04 + target: ldm.models.diffusion.ddpm.LatentDiffusion + params: + linear_start: 0.00085 + linear_end: 0.0120 + num_timesteps_cond: 1 + log_every_t: 200 + timesteps: 1000 + first_stage_key: "jpg" + cond_stage_key: "txt" + image_size: 64 + channels: 4 + cond_stage_trainable: false # Note: different from the one we trained before + conditioning_key: crossattn + monitor: val/loss_simple_ema + scale_factor: 0.18215 + use_ema: False + + scheduler_config: # 10000 warmup steps + target: ldm.lr_scheduler.LambdaLinearScheduler + params: + warm_up_steps: [ 10000 ] + cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases + f_start: [ 1.e-6 ] + f_max: [ 1. ] + f_min: [ 1. ] + + unet_config: + target: ldm.modules.diffusionmodules.openaimodel.UNetModel + params: + image_size: 32 # unused + in_channels: 4 + out_channels: 4 + model_channels: 320 + attention_resolutions: [ 4, 2, 1 ] + num_res_blocks: 2 + channel_mult: [ 1, 2, 4, 4 ] + num_heads: 8 + use_spatial_transformer: True + transformer_depth: 1 + context_dim: 768 + use_checkpoint: True + legacy: False + + first_stage_config: + target: ldm.models.autoencoder.AutoencoderKL + params: + embed_dim: 4 + monitor: val/rec_loss + ddconfig: + double_z: true + z_channels: 4 + resolution: 256 + in_channels: 3 + out_ch: 3 + ch: 128 + ch_mult: + - 1 + - 2 + - 4 + - 4 + num_res_blocks: 2 + attn_resolutions: [] + dropout: 0.0 + lossconfig: + target: torch.nn.Identity + + cond_stage_config: + target: ldm.modules.encoders.modules.FrozenCLIPEmbedder + params: + layer: "hidden" + layer_idx: -2 diff --git a/configs/v1-inference_clip_skip_2_fp16.yaml b/configs/v1-inference_clip_skip_2_fp16.yaml new file mode 100644 index 0000000000000000000000000000000000000000..7eca31c7b5e571c2b1348e94ed9d69978ebd2d52 --- /dev/null +++ b/configs/v1-inference_clip_skip_2_fp16.yaml @@ -0,0 +1,74 @@ +model: + base_learning_rate: 1.0e-04 + target: ldm.models.diffusion.ddpm.LatentDiffusion + params: + linear_start: 0.00085 + linear_end: 0.0120 + num_timesteps_cond: 1 + log_every_t: 200 + timesteps: 1000 + first_stage_key: "jpg" + cond_stage_key: "txt" + image_size: 64 + channels: 4 + cond_stage_trainable: false # Note: different from the one we trained before + conditioning_key: crossattn + monitor: val/loss_simple_ema + scale_factor: 0.18215 + use_ema: False + + scheduler_config: # 10000 warmup steps + target: ldm.lr_scheduler.LambdaLinearScheduler + params: + warm_up_steps: [ 10000 ] + cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases + f_start: [ 1.e-6 ] + f_max: [ 1. ] + f_min: [ 1. ] + + unet_config: + target: ldm.modules.diffusionmodules.openaimodel.UNetModel + params: + use_fp16: True + image_size: 32 # unused + in_channels: 4 + out_channels: 4 + model_channels: 320 + attention_resolutions: [ 4, 2, 1 ] + num_res_blocks: 2 + channel_mult: [ 1, 2, 4, 4 ] + num_heads: 8 + use_spatial_transformer: True + transformer_depth: 1 + context_dim: 768 + use_checkpoint: True + legacy: False + + first_stage_config: + target: ldm.models.autoencoder.AutoencoderKL + params: + embed_dim: 4 + monitor: val/rec_loss + ddconfig: + double_z: true + z_channels: 4 + resolution: 256 + in_channels: 3 + out_ch: 3 + ch: 128 + ch_mult: + - 1 + - 2 + - 4 + - 4 + num_res_blocks: 2 + attn_resolutions: [] + dropout: 0.0 + lossconfig: + target: torch.nn.Identity + + cond_stage_config: + target: ldm.modules.encoders.modules.FrozenCLIPEmbedder + params: + layer: "hidden" + layer_idx: -2 diff --git a/configs/v1-inference_fp16.yaml b/configs/v1-inference_fp16.yaml new file mode 100644 index 0000000000000000000000000000000000000000..147f42b17b835cc839338156f99e8f971df5c1aa --- /dev/null +++ b/configs/v1-inference_fp16.yaml @@ -0,0 +1,71 @@ +model: + base_learning_rate: 1.0e-04 + target: ldm.models.diffusion.ddpm.LatentDiffusion + params: + linear_start: 0.00085 + linear_end: 0.0120 + num_timesteps_cond: 1 + log_every_t: 200 + timesteps: 1000 + first_stage_key: "jpg" + cond_stage_key: "txt" + image_size: 64 + channels: 4 + cond_stage_trainable: false # Note: different from the one we trained before + conditioning_key: crossattn + monitor: val/loss_simple_ema + scale_factor: 0.18215 + use_ema: False + + scheduler_config: # 10000 warmup steps + target: ldm.lr_scheduler.LambdaLinearScheduler + params: + warm_up_steps: [ 10000 ] + cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases + f_start: [ 1.e-6 ] + f_max: [ 1. ] + f_min: [ 1. ] + + unet_config: + target: ldm.modules.diffusionmodules.openaimodel.UNetModel + params: + use_fp16: True + image_size: 32 # unused + in_channels: 4 + out_channels: 4 + model_channels: 320 + attention_resolutions: [ 4, 2, 1 ] + num_res_blocks: 2 + channel_mult: [ 1, 2, 4, 4 ] + num_heads: 8 + use_spatial_transformer: True + transformer_depth: 1 + context_dim: 768 + use_checkpoint: True + legacy: False + + first_stage_config: + target: ldm.models.autoencoder.AutoencoderKL + params: + embed_dim: 4 + monitor: val/rec_loss + ddconfig: + double_z: true + z_channels: 4 + resolution: 256 + in_channels: 3 + out_ch: 3 + ch: 128 + ch_mult: + - 1 + - 2 + - 4 + - 4 + num_res_blocks: 2 + attn_resolutions: [] + dropout: 0.0 + lossconfig: + target: torch.nn.Identity + + cond_stage_config: + target: ldm.modules.encoders.modules.FrozenCLIPEmbedder diff --git a/configs/v1-inpainting-inference.yaml b/configs/v1-inpainting-inference.yaml new file mode 100644 index 0000000000000000000000000000000000000000..45f3f82d461cd8c6109f26ec3b1da75366eda0b0 --- /dev/null +++ b/configs/v1-inpainting-inference.yaml @@ -0,0 +1,71 @@ +model: + base_learning_rate: 7.5e-05 + target: ldm.models.diffusion.ddpm.LatentInpaintDiffusion + params: + linear_start: 0.00085 + linear_end: 0.0120 + num_timesteps_cond: 1 + log_every_t: 200 + timesteps: 1000 + first_stage_key: "jpg" + cond_stage_key: "txt" + image_size: 64 + channels: 4 + cond_stage_trainable: false # Note: different from the one we trained before + conditioning_key: hybrid # important + monitor: val/loss_simple_ema + scale_factor: 0.18215 + finetune_keys: null + + scheduler_config: # 10000 warmup steps + target: ldm.lr_scheduler.LambdaLinearScheduler + params: + warm_up_steps: [ 2500 ] # NOTE for resuming. use 10000 if starting from scratch + cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases + f_start: [ 1.e-6 ] + f_max: [ 1. ] + f_min: [ 1. ] + + unet_config: + target: ldm.modules.diffusionmodules.openaimodel.UNetModel + params: + image_size: 32 # unused + in_channels: 9 # 4 data + 4 downscaled image + 1 mask + out_channels: 4 + model_channels: 320 + attention_resolutions: [ 4, 2, 1 ] + num_res_blocks: 2 + channel_mult: [ 1, 2, 4, 4 ] + num_heads: 8 + use_spatial_transformer: True + transformer_depth: 1 + context_dim: 768 + use_checkpoint: True + legacy: False + + first_stage_config: + target: ldm.models.autoencoder.AutoencoderKL + params: + embed_dim: 4 + monitor: val/rec_loss + ddconfig: + double_z: true + z_channels: 4 + resolution: 256 + in_channels: 3 + out_ch: 3 + ch: 128 + ch_mult: + - 1 + - 2 + - 4 + - 4 + num_res_blocks: 2 + attn_resolutions: [] + dropout: 0.0 + lossconfig: + target: torch.nn.Identity + + cond_stage_config: + target: ldm.modules.encoders.modules.FrozenCLIPEmbedder + diff --git a/configs/v2-inference-v.yaml b/configs/v2-inference-v.yaml new file mode 100644 index 0000000000000000000000000000000000000000..8ec8dfbfefe94ae8522c93017668fea78d580acf --- /dev/null +++ b/configs/v2-inference-v.yaml @@ -0,0 +1,68 @@ +model: + base_learning_rate: 1.0e-4 + target: ldm.models.diffusion.ddpm.LatentDiffusion + params: + parameterization: "v" + linear_start: 0.00085 + linear_end: 0.0120 + num_timesteps_cond: 1 + log_every_t: 200 + timesteps: 1000 + first_stage_key: "jpg" + cond_stage_key: "txt" + image_size: 64 + channels: 4 + cond_stage_trainable: false + conditioning_key: crossattn + monitor: val/loss_simple_ema + scale_factor: 0.18215 + use_ema: False # we set this to false because this is an inference only config + + unet_config: + target: ldm.modules.diffusionmodules.openaimodel.UNetModel + params: + use_checkpoint: True + use_fp16: True + image_size: 32 # unused + in_channels: 4 + out_channels: 4 + model_channels: 320 + attention_resolutions: [ 4, 2, 1 ] + num_res_blocks: 2 + channel_mult: [ 1, 2, 4, 4 ] + num_head_channels: 64 # need to fix for flash-attn + use_spatial_transformer: True + use_linear_in_transformer: True + transformer_depth: 1 + context_dim: 1024 + legacy: False + + first_stage_config: + target: ldm.models.autoencoder.AutoencoderKL + params: + embed_dim: 4 + monitor: val/rec_loss + ddconfig: + #attn_type: "vanilla-xformers" + double_z: true + z_channels: 4 + resolution: 256 + in_channels: 3 + out_ch: 3 + ch: 128 + ch_mult: + - 1 + - 2 + - 4 + - 4 + num_res_blocks: 2 + attn_resolutions: [] + dropout: 0.0 + lossconfig: + target: torch.nn.Identity + + cond_stage_config: + target: ldm.modules.encoders.modules.FrozenOpenCLIPEmbedder + params: + freeze: True + layer: "penultimate" diff --git a/configs/v2-inference-v_fp32.yaml b/configs/v2-inference-v_fp32.yaml new file mode 100644 index 0000000000000000000000000000000000000000..d5c9b9cb29ca162ade44a7c922f59e75d7d57813 --- /dev/null +++ b/configs/v2-inference-v_fp32.yaml @@ -0,0 +1,68 @@ +model: + base_learning_rate: 1.0e-4 + target: ldm.models.diffusion.ddpm.LatentDiffusion + params: + parameterization: "v" + linear_start: 0.00085 + linear_end: 0.0120 + num_timesteps_cond: 1 + log_every_t: 200 + timesteps: 1000 + first_stage_key: "jpg" + cond_stage_key: "txt" + image_size: 64 + channels: 4 + cond_stage_trainable: false + conditioning_key: crossattn + monitor: val/loss_simple_ema + scale_factor: 0.18215 + use_ema: False # we set this to false because this is an inference only config + + unet_config: + target: ldm.modules.diffusionmodules.openaimodel.UNetModel + params: + use_checkpoint: True + use_fp16: False + image_size: 32 # unused + in_channels: 4 + out_channels: 4 + model_channels: 320 + attention_resolutions: [ 4, 2, 1 ] + num_res_blocks: 2 + channel_mult: [ 1, 2, 4, 4 ] + num_head_channels: 64 # need to fix for flash-attn + use_spatial_transformer: True + use_linear_in_transformer: True + transformer_depth: 1 + context_dim: 1024 + legacy: False + + first_stage_config: + target: ldm.models.autoencoder.AutoencoderKL + params: + embed_dim: 4 + monitor: val/rec_loss + ddconfig: + #attn_type: "vanilla-xformers" + double_z: true + z_channels: 4 + resolution: 256 + in_channels: 3 + out_ch: 3 + ch: 128 + ch_mult: + - 1 + - 2 + - 4 + - 4 + num_res_blocks: 2 + attn_resolutions: [] + dropout: 0.0 + lossconfig: + target: torch.nn.Identity + + cond_stage_config: + target: ldm.modules.encoders.modules.FrozenOpenCLIPEmbedder + params: + freeze: True + layer: "penultimate" diff --git a/configs/v2-inference.yaml b/configs/v2-inference.yaml new file mode 100644 index 0000000000000000000000000000000000000000..152c4f3c2b36c3b246a9cb10eb8166134b0d2e1c --- /dev/null +++ b/configs/v2-inference.yaml @@ -0,0 +1,67 @@ +model: + base_learning_rate: 1.0e-4 + target: ldm.models.diffusion.ddpm.LatentDiffusion + params: + linear_start: 0.00085 + linear_end: 0.0120 + num_timesteps_cond: 1 + log_every_t: 200 + timesteps: 1000 + first_stage_key: "jpg" + cond_stage_key: "txt" + image_size: 64 + channels: 4 + cond_stage_trainable: false + conditioning_key: crossattn + monitor: val/loss_simple_ema + scale_factor: 0.18215 + use_ema: False # we set this to false because this is an inference only config + + unet_config: + target: ldm.modules.diffusionmodules.openaimodel.UNetModel + params: + use_checkpoint: True + use_fp16: True + image_size: 32 # unused + in_channels: 4 + out_channels: 4 + model_channels: 320 + attention_resolutions: [ 4, 2, 1 ] + num_res_blocks: 2 + channel_mult: [ 1, 2, 4, 4 ] + num_head_channels: 64 # need to fix for flash-attn + use_spatial_transformer: True + use_linear_in_transformer: True + transformer_depth: 1 + context_dim: 1024 + legacy: False + + first_stage_config: + target: ldm.models.autoencoder.AutoencoderKL + params: + embed_dim: 4 + monitor: val/rec_loss + ddconfig: + #attn_type: "vanilla-xformers" + double_z: true + z_channels: 4 + resolution: 256 + in_channels: 3 + out_ch: 3 + ch: 128 + ch_mult: + - 1 + - 2 + - 4 + - 4 + num_res_blocks: 2 + attn_resolutions: [] + dropout: 0.0 + lossconfig: + target: torch.nn.Identity + + cond_stage_config: + target: ldm.modules.encoders.modules.FrozenOpenCLIPEmbedder + params: + freeze: True + layer: "penultimate" diff --git a/configs/v2-inference_fp32.yaml b/configs/v2-inference_fp32.yaml new file mode 100644 index 0000000000000000000000000000000000000000..0d03231f3f2c2e8ef8fbe0d781e5f3d65409ef3a --- /dev/null +++ b/configs/v2-inference_fp32.yaml @@ -0,0 +1,67 @@ +model: + base_learning_rate: 1.0e-4 + target: ldm.models.diffusion.ddpm.LatentDiffusion + params: + linear_start: 0.00085 + linear_end: 0.0120 + num_timesteps_cond: 1 + log_every_t: 200 + timesteps: 1000 + first_stage_key: "jpg" + cond_stage_key: "txt" + image_size: 64 + channels: 4 + cond_stage_trainable: false + conditioning_key: crossattn + monitor: val/loss_simple_ema + scale_factor: 0.18215 + use_ema: False # we set this to false because this is an inference only config + + unet_config: + target: ldm.modules.diffusionmodules.openaimodel.UNetModel + params: + use_checkpoint: True + use_fp16: False + image_size: 32 # unused + in_channels: 4 + out_channels: 4 + model_channels: 320 + attention_resolutions: [ 4, 2, 1 ] + num_res_blocks: 2 + channel_mult: [ 1, 2, 4, 4 ] + num_head_channels: 64 # need to fix for flash-attn + use_spatial_transformer: True + use_linear_in_transformer: True + transformer_depth: 1 + context_dim: 1024 + legacy: False + + first_stage_config: + target: ldm.models.autoencoder.AutoencoderKL + params: + embed_dim: 4 + monitor: val/rec_loss + ddconfig: + #attn_type: "vanilla-xformers" + double_z: true + z_channels: 4 + resolution: 256 + in_channels: 3 + out_ch: 3 + ch: 128 + ch_mult: + - 1 + - 2 + - 4 + - 4 + num_res_blocks: 2 + attn_resolutions: [] + dropout: 0.0 + lossconfig: + target: torch.nn.Identity + + cond_stage_config: + target: ldm.modules.encoders.modules.FrozenOpenCLIPEmbedder + params: + freeze: True + layer: "penultimate" diff --git a/configs/v2-inpainting-inference.yaml b/configs/v2-inpainting-inference.yaml new file mode 100644 index 0000000000000000000000000000000000000000..32a9471d71b828c51bcbbabfe34c5f6c8282c803 --- /dev/null +++ b/configs/v2-inpainting-inference.yaml @@ -0,0 +1,158 @@ +model: + base_learning_rate: 5.0e-05 + target: ldm.models.diffusion.ddpm.LatentInpaintDiffusion + params: + linear_start: 0.00085 + linear_end: 0.0120 + num_timesteps_cond: 1 + log_every_t: 200 + timesteps: 1000 + first_stage_key: "jpg" + cond_stage_key: "txt" + image_size: 64 + channels: 4 + cond_stage_trainable: false + conditioning_key: hybrid + scale_factor: 0.18215 + monitor: val/loss_simple_ema + finetune_keys: null + use_ema: False + + unet_config: + target: ldm.modules.diffusionmodules.openaimodel.UNetModel + params: + use_checkpoint: True + image_size: 32 # unused + in_channels: 9 + out_channels: 4 + model_channels: 320 + attention_resolutions: [ 4, 2, 1 ] + num_res_blocks: 2 + channel_mult: [ 1, 2, 4, 4 ] + num_head_channels: 64 # need to fix for flash-attn + use_spatial_transformer: True + use_linear_in_transformer: True + transformer_depth: 1 + context_dim: 1024 + legacy: False + + first_stage_config: + target: ldm.models.autoencoder.AutoencoderKL + params: + embed_dim: 4 + monitor: val/rec_loss + ddconfig: + #attn_type: "vanilla-xformers" + double_z: true + z_channels: 4 + resolution: 256 + in_channels: 3 + out_ch: 3 + ch: 128 + ch_mult: + - 1 + - 2 + - 4 + - 4 + num_res_blocks: 2 + attn_resolutions: [ ] + dropout: 0.0 + lossconfig: + target: torch.nn.Identity + + cond_stage_config: + target: ldm.modules.encoders.modules.FrozenOpenCLIPEmbedder + params: + freeze: True + layer: "penultimate" + + +data: + target: ldm.data.laion.WebDataModuleFromConfig + params: + tar_base: null # for concat as in LAION-A + p_unsafe_threshold: 0.1 + filter_word_list: "data/filters.yaml" + max_pwatermark: 0.45 + batch_size: 8 + num_workers: 6 + multinode: True + min_size: 512 + train: + shards: + - "pipe:aws s3 cp s3://stability-aws/laion-a-native/part-0/{00000..18699}.tar -" + - "pipe:aws s3 cp s3://stability-aws/laion-a-native/part-1/{00000..18699}.tar -" + - "pipe:aws s3 cp s3://stability-aws/laion-a-native/part-2/{00000..18699}.tar -" + - "pipe:aws s3 cp s3://stability-aws/laion-a-native/part-3/{00000..18699}.tar -" + - "pipe:aws s3 cp s3://stability-aws/laion-a-native/part-4/{00000..18699}.tar -" #{00000-94333}.tar" + shuffle: 10000 + image_key: jpg + image_transforms: + - target: torchvision.transforms.Resize + params: + size: 512 + interpolation: 3 + - target: torchvision.transforms.RandomCrop + params: + size: 512 + postprocess: + target: ldm.data.laion.AddMask + params: + mode: "512train-large" + p_drop: 0.25 + # NOTE use enough shards to avoid empty validation loops in workers + validation: + shards: + - "pipe:aws s3 cp s3://deep-floyd-s3/datasets/laion_cleaned-part5/{93001..94333}.tar - " + shuffle: 0 + image_key: jpg + image_transforms: + - target: torchvision.transforms.Resize + params: + size: 512 + interpolation: 3 + - target: torchvision.transforms.CenterCrop + params: + size: 512 + postprocess: + target: ldm.data.laion.AddMask + params: + mode: "512train-large" + p_drop: 0.25 + +lightning: + find_unused_parameters: True + modelcheckpoint: + params: + every_n_train_steps: 5000 + + callbacks: + metrics_over_trainsteps_checkpoint: + params: + every_n_train_steps: 10000 + + image_logger: + target: main.ImageLogger + params: + enable_autocast: False + disabled: False + batch_frequency: 1000 + max_images: 4 + increase_log_steps: False + log_first_step: False + log_images_kwargs: + use_ema_scope: False + inpaint: False + plot_progressive_rows: False + plot_diffusion_rows: False + N: 4 + unconditional_guidance_scale: 5.0 + unconditional_guidance_label: [""] + ddim_steps: 50 # todo check these out for depth2img, + ddim_eta: 0.0 # todo check these out for depth2img, + + trainer: + benchmark: True + val_check_interval: 5000000 + num_sanity_val_steps: 0 + accumulate_grad_batches: 1 diff --git a/controlnet/SD1.5/animatediff/v3_sd15_sparsectrl_rgb.ckpt b/controlnet/SD1.5/animatediff/v3_sd15_sparsectrl_rgb.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..1d7e8c5a5b21f50a611b6d234154a7b2db6d8105 --- /dev/null +++ b/controlnet/SD1.5/animatediff/v3_sd15_sparsectrl_rgb.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c93f27a3cd15edf99bbddf4522509f2e831f515cc7db9c820d955037ddafbe45 +size 1988040333 diff --git a/controlnet/SD1.5/animatediff/v3_sd15_sparsectrl_scribble.ckpt b/controlnet/SD1.5/animatediff/v3_sd15_sparsectrl_scribble.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..b9d3f45d8ca28bd5db4953dbf4173a64d108e10d --- /dev/null +++ b/controlnet/SD1.5/animatediff/v3_sd15_sparsectrl_scribble.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:148399d3267e104bee1b1171100af0ad7e125c4f99aacada21342d778eaf3896 +size 1992335697 diff --git a/controlnet/TTPLANET_Controlnet_Tile_realistic_v1_fp16.safetensors b/controlnet/TTPLANET_Controlnet_Tile_realistic_v1_fp16.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8dc734abc66214f35d3eff39bc5cf7de5c1a87f9 --- /dev/null +++ b/controlnet/TTPLANET_Controlnet_Tile_realistic_v1_fp16.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fa2a5f2fec814819c835403c778be2c2437d147931921304569fccec2b45173c +size 2502139104 diff --git a/controlnet/bdsqlsz_controlllite_xl_lineart_anime_denoise.safetensors b/controlnet/bdsqlsz_controlllite_xl_lineart_anime_denoise.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ad2ab223822646bf569aa7f38223b4291d11183e --- /dev/null +++ b/controlnet/bdsqlsz_controlllite_xl_lineart_anime_denoise.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:311e142ab5602b172d76c1d84abb7f664b5c19d9cc7e840ee5668096d513e071 +size 223906688 diff --git a/controlnet/control-lora-depth-rank256.safetensors b/controlnet/control-lora-depth-rank256.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4e555b1e2643d07ee68138eb9d383218eb1f85d7 --- /dev/null +++ b/controlnet/control-lora-depth-rank256.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:559d2468951bf254c13bacd9c5d05d01ad67b060f6a73e8131d26ebf459c1c79 +size 774445779 diff --git a/controlnet/control_v11f1e_sd15_tile.pth b/controlnet/control_v11f1e_sd15_tile.pth new file mode 100644 index 0000000000000000000000000000000000000000..f0c50aea7aac578a1bb6f908ff7236ed2f26ee73 --- /dev/null +++ b/controlnet/control_v11f1e_sd15_tile.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8aa69b8d391e72c2fced6a650268137dc0ed594cafe8a2c0f8b994799a21979b +size 1445235023 diff --git a/controlnet/control_v11f1p_sd15_depth.pth b/controlnet/control_v11f1p_sd15_depth.pth new file mode 100644 index 0000000000000000000000000000000000000000..e20ec5a5140ed678f1db1bd0efe02e74a757fd91 --- /dev/null +++ b/controlnet/control_v11f1p_sd15_depth.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:761077ffe369fe8cf16ae353f8226bd4ca29805b161052f82c0170c7b50f1d99 +size 1445235365 diff --git a/controlnet/control_v11p_sd15_canny.pth b/controlnet/control_v11p_sd15_canny.pth new file mode 100644 index 0000000000000000000000000000000000000000..a39717042be81190d25eba173a94f3bf35606090 --- /dev/null +++ b/controlnet/control_v11p_sd15_canny.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f99cfe4c70910e38e3fece9918a4979ed7d3dcf9b81cee293e1755363af5406a +size 1445234681 diff --git a/controlnet/control_v11p_sd15_lineart.pth b/controlnet/control_v11p_sd15_lineart.pth new file mode 100644 index 0000000000000000000000000000000000000000..36362b436291818e2838f04d4c9b8e5e38082c90 --- /dev/null +++ b/controlnet/control_v11p_sd15_lineart.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fb5fb912696830da948e5056b50adfa3fcc68ff760c43e4e7a8499853af43c61 +size 1445235365 diff --git a/controlnet/control_v11p_sd15_openpose.pth b/controlnet/control_v11p_sd15_openpose.pth new file mode 100644 index 0000000000000000000000000000000000000000..56b8d5bad6d1c47d5ff9c7878c3848d1962a29ff --- /dev/null +++ b/controlnet/control_v11p_sd15_openpose.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:db97becd92cd19aff71352a60e93c2508decba3dee64f01f686727b9b406a9dd +size 1445235707 diff --git a/controlnet/control_v11p_sd15_scribble.pth b/controlnet/control_v11p_sd15_scribble.pth new file mode 100644 index 0000000000000000000000000000000000000000..3959453782ccca5b138b4943937447eba140b118 --- /dev/null +++ b/controlnet/control_v11p_sd15_scribble.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:14d33e8bc14ba679595d976789e4eedc36a3d2d21866ab5896355a7e6677749f +size 1445235707 diff --git a/controlnet/motion_controlnet_checkpoint.ckpt b/controlnet/motion_controlnet_checkpoint.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..24009719fb3865d3ed6edff73f76a93e9296bf3d --- /dev/null +++ b/controlnet/motion_controlnet_checkpoint.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0dcba88a7ea5d0f969c87db23eab1522e2a9cfb87a25d131c71ddcccdec34948 +size 1445261885 diff --git a/controlnet/put_controlnets_and_t2i_here b/controlnet/put_controlnets_and_t2i_here new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/diffusers/put_diffusers_models_here b/diffusers/put_diffusers_models_here new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/embeddings/put_embeddings_or_textual_inversion_concepts_here b/embeddings/put_embeddings_or_textual_inversion_concepts_here new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/gligen/put_gligen_models_here b/gligen/put_gligen_models_here new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/hypernetworks/put_hypernetworks_here b/hypernetworks/put_hypernetworks_here new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/ipadapter/ip-adapter-faceid-plus_sd15.bin b/ipadapter/ip-adapter-faceid-plus_sd15.bin new file mode 100644 index 0000000000000000000000000000000000000000..57e1d0254a2f24d060f890d34e233e9feef685d1 --- /dev/null +++ b/ipadapter/ip-adapter-faceid-plus_sd15.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:252fb53e0d018489d9e7f9b9e2001a52ff700e491894011ada7cfb471e0fadf2 +size 156558503 diff --git a/ipadapter/ip-adapter-faceid-plusv2_sd15.bin b/ipadapter/ip-adapter-faceid-plusv2_sd15.bin new file mode 100644 index 0000000000000000000000000000000000000000..2b83b06efc1d0f8bc11c2aecdb1932d7c80faefa --- /dev/null +++ b/ipadapter/ip-adapter-faceid-plusv2_sd15.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:26d0d86a1d60d6cc811d3b8862178b461e1eeb651e6fe2b72ba17aa95411e313 +size 156558509 diff --git a/ipadapter/ip-adapter-faceid-portrait_sd15.bin b/ipadapter/ip-adapter-faceid-portrait_sd15.bin new file mode 100644 index 0000000000000000000000000000000000000000..c815d258a84702d84cf9007d0712ca50a9140976 --- /dev/null +++ b/ipadapter/ip-adapter-faceid-portrait_sd15.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:68570f3c14fe125b00d4abf416af33d8b13bb496f4e55af2eb5d0a6017ee99a4 +size 64586611 diff --git a/ipadapter/ip-adapter-faceid_sd15.bin b/ipadapter/ip-adapter-faceid_sd15.bin new file mode 100644 index 0000000000000000000000000000000000000000..95b32b7455efc1aef93dc4b3d162b6e0b7bc8a8e --- /dev/null +++ b/ipadapter/ip-adapter-faceid_sd15.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:201344e22e6f55849cf07ca7a6e53d8c3b001327c66cb9710d69fd5da48a8da7 +size 96740574 diff --git a/ipadapter/ip-adapter-full-face_sd15.safetensors b/ipadapter/ip-adapter-full-face_sd15.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..53e15c464d79ac549dd16d9de42cdf76bda84e66 --- /dev/null +++ b/ipadapter/ip-adapter-full-face_sd15.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f4a17fb643bf876235a45a0e87a49da2855be6584b28ca04c62a97ab5ff1c6f3 +size 43592352 diff --git a/ipadapter/ip-adapter-plus-face_sd15.safetensors b/ipadapter/ip-adapter-plus-face_sd15.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..767025ddb47fe89baa927b8c2eee8fe99a1f6edb --- /dev/null +++ b/ipadapter/ip-adapter-plus-face_sd15.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1c9edc21af6f737dc1d6e0e734190e976cfacf802d6b024b77aa3be922f7569b +size 98183288 diff --git a/ipadapter/ip-adapter-plus_sd15.bin b/ipadapter/ip-adapter-plus_sd15.bin new file mode 100644 index 0000000000000000000000000000000000000000..9598908b1a315d77bad89c9fe572f79c6afc1bd3 --- /dev/null +++ b/ipadapter/ip-adapter-plus_sd15.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1cb77fc0613369b66be1531cc452b823a4af7d87ee56956000a69fc39e3817ba +size 158033179 diff --git a/ipadapter/ip-adapter-plus_sd15.safetensors b/ipadapter/ip-adapter-plus_sd15.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1ab10ab02149141a0111e5ba67c0653008319121 --- /dev/null +++ b/ipadapter/ip-adapter-plus_sd15.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a1c250be40455cc61a43da1201ec3f1edaea71214865fb47f57927e06cbe4996 +size 98183288 diff --git a/ipadapter/ip-adapter-plus_sdxl_vit-h.bin b/ipadapter/ip-adapter-plus_sdxl_vit-h.bin new file mode 100644 index 0000000000000000000000000000000000000000..fd50c88d8ae4fb54ff3a56980475501d6768fd18 --- /dev/null +++ b/ipadapter/ip-adapter-plus_sdxl_vit-h.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec70edb7cc8e769c9388d94eeaea3e4526352c9fae793a608782d1d8951fde90 +size 1013454427 diff --git a/ipadapter/ip-adapter_sd15.safetensors b/ipadapter/ip-adapter_sd15.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..67788447289e5858c26dbc8fbf0a7ce3fd6c50fb --- /dev/null +++ b/ipadapter/ip-adapter_sd15.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:289b45f16d043d0bf542e45831f971dcdaabe18b656f11e86d9dfba7e9ee3369 +size 44642768 diff --git a/ipadapter/ip-adapter_sd15_light.safetensors b/ipadapter/ip-adapter_sd15_light.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c45a20f52f297aa523bedd4bee907b276aac0c3b --- /dev/null +++ b/ipadapter/ip-adapter_sd15_light.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0747d08db670535bfa286452a77d93cebad5c677b46d038543f9f2de8690bb26 +size 44642768 diff --git a/loras/SD1.5/animatediff/v3_sd15_adapter.ckpt b/loras/SD1.5/animatediff/v3_sd15_adapter.ckpt new file mode 100644 index 0000000000000000000000000000000000000000..cceb41ce12f2efa1027d4a9e7157d0ecb7e4750c --- /dev/null +++ b/loras/SD1.5/animatediff/v3_sd15_adapter.ckpt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fd2d8e26480f6ab013c1e6af86fdf1dedbb1ed5baf850ccd5f365f39d6c3472c +size 102134097 diff --git a/loras/Zenon-Retagged-10.safetensors b/loras/Zenon-Retagged-10.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1b3103855c2ee9ab3bd7a5510acac6017559ce17 --- /dev/null +++ b/loras/Zenon-Retagged-10.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a28f9488113bfc5341edb0fa98d0973c777ded7cdb359f59630abb0a1ae6c47a +size 18994288 diff --git a/loras/hermes-horses-10.safetensors b/loras/hermes-horses-10.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9f0b9598213b8afe142d878023b22f173c7d0856 --- /dev/null +++ b/loras/hermes-horses-10.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:866c3c969d2341c2d819a4c4369e2c6a93b2e306a976e6515b4d4df1a7d40691 +size 19008688 diff --git a/loras/lcm/SD1.5/pytorch_lora_weights.safetensors b/loras/lcm/SD1.5/pytorch_lora_weights.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ee197fb9f23830b246864db7b9de672995bcbad7 --- /dev/null +++ b/loras/lcm/SD1.5/pytorch_lora_weights.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8f90d840e075ff588a58e22c6586e2ae9a6f7922996ee6649a7f01072333afe4 +size 134621556 diff --git a/loras/lcm_lora_sd15.safetensors b/loras/lcm_lora_sd15.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ee197fb9f23830b246864db7b9de672995bcbad7 --- /dev/null +++ b/loras/lcm_lora_sd15.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8f90d840e075ff588a58e22c6586e2ae9a6f7922996ee6649a7f01072333afe4 +size 134621556 diff --git a/photomaker/put_photomaker_models_here b/photomaker/put_photomaker_models_here new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/sams/sam_vit_b_01ec64.pth b/sams/sam_vit_b_01ec64.pth new file mode 100644 index 0000000000000000000000000000000000000000..ab7d111e57bd052a76fe669986560e3555e9c8f6 --- /dev/null +++ b/sams/sam_vit_b_01ec64.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec2df62732614e57411cdcf32a23ffdf28910380d03139ee0f4fcbe91eb8c912 +size 375042383 diff --git a/style_models/put_t2i_style_model_here b/style_models/put_t2i_style_model_here new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/ultralytics/bbox/face_yolov8m.pt b/ultralytics/bbox/face_yolov8m.pt new file mode 100644 index 0000000000000000000000000000000000000000..3581945a1f3342c5c48d0b7b339b56dd1676008a --- /dev/null +++ b/ultralytics/bbox/face_yolov8m.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f02b8a23e6f12bd2c1b1f6714f66f984c728fa41ed749d033e7d6dea511ef70c +size 52026019 diff --git a/ultralytics/bbox/hand_yolov8s.pt b/ultralytics/bbox/hand_yolov8s.pt new file mode 100644 index 0000000000000000000000000000000000000000..21091d538a48b1afd5e9910e2d3863a1d4974799 --- /dev/null +++ b/ultralytics/bbox/hand_yolov8s.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5c4faf8d17286ace2c3d3346c6d0d4a0c8d62404955263a7ae95c1dd7eb877af +size 22507707 diff --git a/ultralytics/segm/person_yolov8m-seg.pt b/ultralytics/segm/person_yolov8m-seg.pt new file mode 100644 index 0000000000000000000000000000000000000000..8ae56d795a72bb6b19938117adf3c50ee70c21fb --- /dev/null +++ b/ultralytics/segm/person_yolov8m-seg.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9d881ec50b831f546e37977081b18f4e3bf65664aec163f97a311b0955499795 +size 54827683 diff --git a/unet/put_unet_files_here b/unet/put_unet_files_here new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/upscale_models/4x-UltraSharp.pth b/upscale_models/4x-UltraSharp.pth new file mode 100644 index 0000000000000000000000000000000000000000..9f3bb839bebd6cd26c94122b7651261d0b346a50 --- /dev/null +++ b/upscale_models/4x-UltraSharp.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a5812231fc936b42af08a5edba784195495d303d5b3248c24489ef0c4021fe01 +size 66961958 diff --git a/upscale_models/4x_foolhardy_Remacri.pth b/upscale_models/4x_foolhardy_Remacri.pth new file mode 100644 index 0000000000000000000000000000000000000000..7dc76f8e1e196da607cabbbf9f58fa4acea73930 --- /dev/null +++ b/upscale_models/4x_foolhardy_Remacri.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e1a73bd89c2da1ae494774746398689048b5a892bd9653e146713f9df8bca86a +size 67025055 diff --git a/upscale_models/put_esrgan_and_other_upscale_models_here b/upscale_models/put_esrgan_and_other_upscale_models_here new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/vae/clearvae_v23.safetensors b/vae/clearvae_v23.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9bace1d68075d2ce7eee426734e1a4fbe26e35d2 --- /dev/null +++ b/vae/clearvae_v23.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ab33ad55c62bd7ff05480ad1adbea50e8bbeb9225281ec00e990d2ce363f7abf +size 334640988 diff --git a/vae/put_vae_here b/vae/put_vae_here new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/vae/sdxl_vae.safetensors b/vae/sdxl_vae.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5c5bfd44fb0132b3cbd9d122244f502a7cd2bb24 --- /dev/null +++ b/vae/sdxl_vae.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:63aeecb90ff7bc1c115395962d3e803571385b61938377bc7089b36e81e92e2e +size 334641164 diff --git a/vae/vae-ft-mse-840000-ema-pruned.safetensors b/vae/vae-ft-mse-840000-ema-pruned.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..14a39ba28ca5d7ffb8efcf9a24ce5fb31120200b --- /dev/null +++ b/vae/vae-ft-mse-840000-ema-pruned.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:735e4c3a447a3255760d7f86845f09f937809baa529c17370d83e4c3758f3c75 +size 334641190 diff --git a/vae_approx/put_taesd_encoder_pth_and_taesd_decoder_pth_here b/vae_approx/put_taesd_encoder_pth_and_taesd_decoder_pth_here new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/vae_approx/taesd_decoder.pth b/vae_approx/taesd_decoder.pth new file mode 100644 index 0000000000000000000000000000000000000000..38c5e36020be07c8bc17de603b3af60662a30c87 --- /dev/null +++ b/vae_approx/taesd_decoder.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:32f3128746187ac07694f8457c571df342d3243c383087a94d2bd9f95dfe5e87 +size 4915980 diff --git a/vae_approx/taesd_encoder.pth b/vae_approx/taesd_encoder.pth new file mode 100644 index 0000000000000000000000000000000000000000..a8541181eeade0c6fc1d933b19f0229aa9eb0be9 --- /dev/null +++ b/vae_approx/taesd_encoder.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c816e25db3138ebce17d2c561c5a0594849d07c3f67536eb3c19933d6fe5c965 +size 4915788 diff --git a/vae_approx/taesdxl_decoder.pth b/vae_approx/taesdxl_decoder.pth new file mode 100644 index 0000000000000000000000000000000000000000..d511afe942adb89a0e18a6f62fedcd8cff695458 --- /dev/null +++ b/vae_approx/taesdxl_decoder.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e68e771b88ec0278a3b306d8e0fd1c7448471fb91035a842d899e14390303587 +size 4915221 diff --git a/vae_approx/taesdxl_encoder.pth b/vae_approx/taesdxl_encoder.pth new file mode 100644 index 0000000000000000000000000000000000000000..f57e327592c449ce740ffaa2fea41de45f4040ba --- /dev/null +++ b/vae_approx/taesdxl_encoder.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d32e529e67a1f3152afcf755dc1360b221421c988d7eeb6a8a8c8a839432a989 +size 4915029