--- license: other --- It is a repository for storing as many LECOs as I can think of, emphasizing quantity over quality. Files will continue to be added as needed. Because the guidance_scale parameter is somewhat excessive, these LECOs tend to be very sensitive and too effective; using a weight of -0.1 to -1 is appropriate in most cases. All LECOs are trained with target eq positive, erase settings. The target is a one of among danbooru's GENERAL tags what most frequently used in order from the top to the bottom, and sometimes I also add phrases that I have personally come up with. ``` prompts.yaml - target: "$query" positive: "$query" unconditional: "" neutral: "" action: "erase" guidance_scale: 1.0 resolution: 512 batch_size: 4 ``` ```config.yaml prompts_file: prompts.yaml pretrained_model: name_or_path: "/storage/model-1892-0000-0000.safetensors" v2: false v_pred: false network: type: "lierla" rank: 4 alpha: 1.0 training_method: "full" train: precision: "bfloat16" noise_scheduler: "ddim" iterations: 50 lr: 1 optimizer: "Prodigy" lr_scheduler: "cosine" max_denoising_steps: 50 save: name: "$query" path: "/stable-diffusion-webui/models/Lora/LECO/" per_steps: 50 precision: "float16" logging: use_wandb: false verbose: false other: use_xformers: true ```