|
[ARB] |
|
bucket_no_upscale = false |
|
bucket_reso_steps = 64 |
|
enable_bucket = true |
|
max_bucket_reso = 1584 |
|
min_bucket_reso = 327 |
|
|
|
[Attention] |
|
mem_eff_attn = false |
|
xformers = true |
|
|
|
[Basics] |
|
clip_skip = 2 |
|
max_train_epochs = 30 |
|
max_train_steps = 1073741824 |
|
pretrained_model_name_or_path = "******" |
|
resolution = "720,720" |
|
seed = 676138216 |
|
train_data_dir = "******" |
|
|
|
[Cache_latents] |
|
cache_latents = true |
|
vae_batch_size = 1 |
|
cache_latents_to_disk = true |
|
|
|
[Captions] |
|
shuffle_caption = true |
|
caption_extension = ".txt" |
|
keep_tokens = 1 |
|
caption_dropout_rate = 0.05 |
|
caption_dropout_every_n_epochs = 0 |
|
caption_tag_dropout_rate = 0.0 |
|
max_token_length = 150 |
|
weighted_captions = false |
|
token_warmup_min = 1 |
|
token_warmup_step = 0 |
|
|
|
[Data_augmentation] |
|
color_aug = false |
|
flip_aug = false |
|
random_crop = false |
|
|
|
[Dataset] |
|
max_data_loader_n_workers = 8 |
|
persistent_data_loader_workers = true |
|
dataset_repeats = 1 |
|
|
|
[Debugging] |
|
debug_dataset = false |
|
|
|
[Deprecated] |
|
use_8bit_adam = false |
|
use_lion_optimizer = false |
|
learning_rate = 0.0002 |
|
|
|
[Further_improvement] |
|
min_snr_gamma = 0 |
|
multires_noise_discount = 0.3 |
|
multires_noise_iterations = 6 |
|
|
|
[Huggingface] |
|
save_state_to_huggingface = false |
|
resume_from_huggingface = false |
|
async_upload = false |
|
|
|
[Logging] |
|
logging_dir = "******" |
|
log_with = "tensorboard" |
|
log_prefix = "lora_" |
|
|
|
[Lr_scheduler] |
|
lr_scheduler_type = "" |
|
lr_scheduler = "constant" |
|
lr_warmup_steps = 0 |
|
lr_scheduler_num_cycles = 1 |
|
lr_scheduler_power = 1.0 |
|
|
|
[LyCORIS] |
|
network_module = "lycoris.kohya" |
|
network_args = [ "preset=attn-mlp", "algo=lora",] |
|
|
|
[Network_setup] |
|
dim_from_weights = false |
|
network_alpha = 2 |
|
network_dim = 8 |
|
network_dropout = 0 |
|
network_train_text_encoder_only = false |
|
network_train_unet_only = true |
|
resume = false |
|
|
|
[Optimizer] |
|
gradient_accumulation_steps = 1 |
|
gradient_checkpointing = true |
|
max_grad_norm = 1.0 |
|
optimizer_args = [ "weight_decay=0.1", "betas=0.9,0.99",] |
|
optimizer_type = "AdamW8bit" |
|
text_encoder_lr = 0.0006 |
|
train_batch_size = 8 |
|
unet_lr = 0.0006 |
|
|
|
[Others] |
|
lowram = false |
|
training_comment = "narugo1992's automated LoRA training, based on nebulae's config." |
|
|
|
[Regularization] |
|
prior_loss_weight = 1.0 |
|
|
|
[SDv2] |
|
v2 = false |
|
v_parameterization = false |
|
scale_v_pred_loss_like_noise_pred = false |
|
|
|
[Sampling_during_training] |
|
sample_sampler = "ddim" |
|
|
|
[Save] |
|
output_dir = "******" |
|
output_name = "saber_fatestaynightufotable" |
|
save_every_n_epochs = 2 |
|
save_every_n_steps = 1073741824 |
|
save_last_n_steps = 200 |
|
save_last_n_steps_state = 1 |
|
save_model_as = "safetensors" |
|
save_precision = "fp16" |
|
save_state = false |
|
|
|
[Training_preciscion] |
|
mixed_precision = "fp16" |
|
full_fp16 = false |
|
full_bf16 = false |
|
|