|
pretrained_model_name_or_path = "/workspace/models/flux1-dev.safetensors" |
|
ae = "/workspace/models/ae.safetensors" |
|
t5xxl = "/workspace/models/t5xxl_fp16.safetensors" |
|
clip_l = "/workspace/models/clip_l.safetensors" |
|
output_dir = "/workspace/kohya_models/rita-v1" |
|
dataset_config = "/workspace/kohya_models/rita-v1/dataset_config.json" |
|
network_dim = 16 |
|
network_alpha = 16 |
|
train_batch_size = 2 |
|
optimizer_type = "Adamw8bit" |
|
unet_lr = 0.0005 |
|
epoch = 12 |
|
max_train_steps = 1500 |
|
apply_t5_attn_mask = false |
|
cache_latents = true |
|
cache_latents_to_disk = true |
|
cache_text_encoder_outputs = true |
|
cache_text_encoder_outputs_to_disk = true |
|
clip_skip = 1 |
|
discrete_flow_shift = 3.1582 |
|
full_bf16 = true |
|
mixed_precision = "bf16" |
|
gradient_accumulation_steps = 1 |
|
gradient_checkpointing = true |
|
guidance_scale = 1.0 |
|
highvram = true |
|
huber_c = 0.1 |
|
huber_schedule = "snr" |
|
loss_type = "l2" |
|
lr_scheduler = "cosine_with_restarts" |
|
lr_scheduler_args = [] |
|
lr_scheduler_num_cycles = 3 |
|
lr_scheduler_power = 1 |
|
max_data_loader_n_workers = 0 |
|
max_grad_norm = 1 |
|
max_timestep = 1000 |
|
min_snr_gamma = 5 |
|
model_prediction_type = "raw" |
|
network_args = [ "train_double_block_indices=all", "train_single_block_indices=all",] |
|
network_module = "networks.lora_flux" |
|
network_train_unet_only = true |
|
noise_offset = 0.1 |
|
noise_offset_type = "Original" |
|
optimizer_args = [] |
|
prior_loss_weight = 1 |
|
sample_sampler = "euler" |
|
sdpa = true |
|
seed = 42 |
|
t5xxl_max_token_length = 512 |
|
text_encoder_lr = [] |
|
timestep_sampling = "sigmoid" |
|
output_name = "lora" |
|
save_state_to_huggingface = true |
|
save_model_as = "safetensors" |
|
save_every_n_epochs = 1 |
|
save_precision = "bf16" |
|
|