[additional_network_arguments] unet_lr = 0.0005 text_encoder_lr = 0.0001 network_dim = 16 network_alpha = 8 network_module = "networks.lora" [optimizer_arguments] learning_rate = 0.0005 lr_scheduler = "constant_with_warmup" lr_warmup_steps = 10 optimizer_type = "AdamW8bit" [training_arguments] max_train_epochs = 10 save_every_n_epochs = 1 save_last_n_epochs = 10 train_batch_size = 2 clip_skip = 2 min_snr_gamma = 5.0 weighted_captions = false seed = 42 max_token_length = 225 xformers = true lowram = true max_data_loader_n_workers = 8 persistent_data_loader_workers = true save_precision = "fp16" mixed_precision = "fp16" output_dir = "/content/drive/MyDrive/Loras/dog_LoRA/output" logging_dir = "/content/drive/MyDrive/Loras/_logs" output_name = "dog_LoRA" log_prefix = "dog_LoRA" [model_arguments] pretrained_model_name_or_path = "/content/AnyLoRA_noVae_fp16-pruned.ckpt" v2 = false [saving_arguments] save_model_as = "safetensors" [dreambooth_arguments] prior_loss_weight = 1.0 [dataset_arguments] cache_latents = true