the MODEL NAME is referring to PonyXL but it's trained on Animagine cause we forgot to fix it
b5626b9
verified
[sdxl_arguments] | |
cache_text_encoder_outputs = false | |
no_half_vae = true | |
min_timestep = 0 | |
max_timestep = 1000 | |
shuffle_caption = true | |
lowram = true | |
[model_arguments] | |
pretrained_model_name_or_path = "Linaqruf/animagine-xl" | |
vae = "/content/vae/sdxl_vae.safetensors" | |
[dataset_arguments] | |
debug_dataset = false | |
in_json = "/content/LoRA/meta_lat.json" | |
train_data_dir = "/content/LoRA/train_data" | |
dataset_repeats = 5 | |
keep_tokens = 1 | |
resolution = "1024,1024" | |
color_aug = false | |
token_warmup_min = 1 | |
token_warmup_step = 0 | |
[training_arguments] | |
output_dir = "/content/drive/MyDrive/kohya-trainer/output/PoltergeistStyle_ponyxl" | |
output_name = "PoltergeistStyle_ponyxl" | |
save_precision = "fp16" | |
save_every_n_epochs = 1 | |
train_batch_size = 4 | |
max_token_length = 225 | |
mem_eff_attn = false | |
sdpa = true | |
xformers = false | |
max_train_epochs = 5 | |
max_data_loader_n_workers = 8 | |
persistent_data_loader_workers = true | |
gradient_checkpointing = true | |
gradient_accumulation_steps = 1 | |
mixed_precision = "fp16" | |
[logging_arguments] | |
log_with = "wandb" | |
log_tracker_name = "PoltergeistStyle_ponyxl" | |
logging_dir = "/content/LoRA/logs" | |
[sample_prompt_arguments] | |
sample_every_n_epochs = 1 | |
sample_sampler = "k_dpm_2_a" | |
[saving_arguments] | |
save_model_as = "safetensors" | |
[optimizer_arguments] | |
optimizer_type = "AdaFactor" | |
learning_rate = 0.0005 | |
max_grad_norm = 0 | |
optimizer_args = [ "scale_parameter=False", "relative_step=False", "warmup_init=False",] | |
lr_scheduler = "cosine_with_restarts" | |
lr_warmup_steps = 0 | |
lr_scheduler_num_cycles = 3 | |
[additional_network_arguments] | |
no_metadata = false | |
network_module = "networks.lora" | |
network_dim = 32 | |
network_alpha = 16 | |
network_args = [] | |
network_train_unet_only = true | |
[advanced_training_config] | |
save_state = false | |
save_last_n_epochs_state = false | |
multires_noise_iterations = 6 | |
multires_noise_discount = 0.3 | |
caption_dropout_rate = 0 | |
caption_tag_dropout_rate = 0.5 | |
caption_dropout_every_n_epochs = 0 | |
min_snr_gamma = 5 | |