File size: 2,631 Bytes
5bfe3af
581fcec
 
 
 
 
 
 
 
 
 
112b9b4
581fcec
 
 
d2a2d86
6b352dc
 
 
 
 
 
 
ebbf256
6b352dc
 
 
 
 
 
 
d2776c8
6b352dc
 
d2776c8
 
 
 
6b352dc
d2776c8
 
6b352dc
 
 
 
 
 
 
 
 
 
 
d2776c8
6b352dc
 
 
 
 
 
 
d2776c8
6b352dc
 
 
 
 
d2776c8
6b352dc
 
 
 
 
581fcec
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75

MODEL_NAME:str="stabilityai/stable-diffusion-xl-base-1.0"
ADAPTER_NAME:str = "VikramSingh178/sdxl-lora-finetune-product-caption"
ADAPTER_NAME_2:str = "VikramSingh178/Products10k-SDXL-Lora"
VAE_NAME:str= "madebyollin/sdxl-vae-fp16-fix"
DATASET_NAME:str = "hahminlew/kream-product-blip-captions"
PROJECT_NAME:str = "Product Photography"
PRODUCTS_10k_DATASET:str = "VikramSingh178/Products-10k-BLIP-captions"
CAPTIONING_MODEL_NAME:str = "Salesforce/blip-image-captioning-base"
SEGMENTATION_MODEL_NAME:str = "facebook/sam-vit-large"
DETECTION_MODEL_NAME:str = "yolov8l"
ENABLE_COMPILE:bool = False
INPAINTING_MODEL_NAME:str = 'kandinsky-community/kandinsky-2-2-decoder-inpaint'




class Config:
    def __init__(self):
        self.pretrained_model_name_or_path = MODEL_NAME
        self.pretrained_vae_model_name_or_path = VAE_NAME
        self.revision = None
        self.variant = None
        self.dataset_name = PRODUCTS_10k_DATASET
        self.dataset_config_name = None
        self.train_data_dir = None
        self.image_column = 'image'
        self.caption_column = 'text'
        self.validation_prompt = None
        self.num_validation_images = 4
        self.validation_epochs = 1
        self.max_train_samples = 7
        self.output_dir = "output"
        self.cache_dir = None
        self.seed = 42
        self.resolution = 512
        self.center_crop = True
        self.random_flip = True
        self.train_text_encoder = False
        self.train_batch_size = 64
        self.num_train_epochs = 400
        self.max_train_steps = None
        self.checkpointing_steps = 500
        self.checkpoints_total_limit = None
        self.resume_from_checkpoint = None
        self.gradient_accumulation_steps = 1
        self.gradient_checkpointing = False
        self.learning_rate = 1e-4
        self.scale_lr = False
        self.lr_scheduler = "constant"
        self.lr_warmup_steps = 500
        self.snr_gamma = None
        self.allow_tf32 = True
        self.dataloader_num_workers = 0
        self.use_8bit_adam = True
        self.adam_beta1 = 0.9
        self.adam_beta2 = 0.999
        self.adam_weight_decay = 1e-2
        self.adam_epsilon = 1e-08
        self.max_grad_norm = 1.0
        self.push_to_hub = True
        self.hub_token = None
        self.prediction_type = None
        self.hub_model_id = None
        self.logging_dir = "logs"
        self.report_to = "wandb"
        self.mixed_precision = 'fp16'
        self.local_rank = -1
        self.enable_xformers_memory_efficient_attention = False
        self.noise_offset = 0
        self.rank = 4
        self.debug_loss = False