File size: 3,079 Bytes
6cd0f71
 
 
 
 
 
249b307
6cd0f71
 
 
 
 
 
 
 
 
 
249b307
6cd0f71
 
 
 
 
 
 
 
 
 
 
 
 
 
 
41fd312
6cd0f71
249b307
6cd0f71
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
b6a1944
 
6cd0f71
249b307
6cd0f71
 
249b307
6cd0f71
 
b6a1944
6cd0f71
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
293fd5a
6cd0f71
 
 
 
 
249b307
6cd0f71
249b307
 
 
 
b6a1944
6cd0f71
 
 
 
4e1cf15
249b307
6cd0f71
 
 
 
 
 
 
 
 
249b307
b6a1944
6cd0f71
 
 
249b307
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
{
    "seed": 1,
    
    "decoder": {
        "unets": [
            {
                "dim": 416,
                "cond_dim": 512,
                "image_embed_dim": 768,
                "text_embed_dim": 768,
                "cond_on_text_encodings": true,
                "channels": 3,
                "dim_mults": [1, 2, 3, 4],
                "num_resnet_blocks": 4,
                "attn_heads": 8,
                "attn_dim_head": 64,
                "sparse_attn": true,
                "memory_efficient": true
            }
        ],
        "clip": {
            "make": "openai",
            "model": "ViT-L/14"
        },
        "image_sizes": [64],
        "channels": 3,
        "timesteps": 1000,
        "loss_type": "l2",
        "beta_schedule": ["cosine"],
        "learned_variance": true
    },
    "data": {
        "webdataset_base_url": "pipe:aws s3 cp --quiet s3://s-datasets/laion-aesthetic/data/laion2B-en-aesthetic/{}.tar -", 
        "embeddings_url": "s3://s-datasets/laion-aesthetic/ordered_embeddings/",
        "num_workers": 12,
        "batch_size": 21,
        "start_shard": 0,
        "end_shard": 5247,
        "shard_width": 5,
        "index_width": 4,
        "splits": {
            "train": 0.75,
            "val": 0.15,
            "test": 0.1
        },
        "shuffle_train": false,
        "resample_train": true,
        "preprocessing": {
            "RandomResizedCrop": {
                "size": [64, 64],
                "scale": [0.75, 1.0],
                "ratio": [1.0, 1.0]
            },
            "ToTensor": true
        }
    },
    "train": {
        "epochs": 1000,
        "lr":1e-4,
        "wd": 0.01,
        "max_grad_norm": 0.5,
        "save_every_n_samples": 200000,
        "n_sample_images": 10,
        "device": "cuda:0",
        "epoch_samples": 2000000,
        "validation_samples": 100000,
        "use_ema": true,
        "ema_beta": 0.99,
        "save_all": false,
        "save_latest": true,
        "save_best": true,
        "unet_training_mask": [true]
    },
    "evaluate": {
        "n_evaluation_samples": 30,
        "FID": {
            "feature": 64
        },
        "LPIPS": {
            "net_type": "vgg",
            "reduction": "mean"
        }
    },
    "tracker": {
        "data_path": ".tracker-data-2",
        "overwrite_data_path": true,

        "log": {
            "log_type": "wandb",

            "wandb_entity": "nousr_laion",
            "wandb_project": "dalle2_train_decoder",
    
            "wandb_run_id": "5ojoz4bw",

            "wandb_resume": true, 

            "verbose": true
        },

        "load": {
            "load_from": "url",
            "url": "https://huggingface.co/Veldrovive/test_model/resolve/main/eu_latest_checkpoint2.pth" 
        },

        "save": [{
            "save_to": "wandb"
        },
        {
            "save_to": "huggingface",
            "huggingface_repo": "laion/DALLE2-PyTorch",

            "save_latest_to": "latest_{epoch}.pth",
            
            "save_type": "model"
        }]
    }
}