File size: 2,173 Bytes
9512a35 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 |
{
"base_learning_rate": 2e-05,
"cfg_conds": [
"txt"
],
"cfg_scales": [
7.5
],
"channels": 4,
"cond_stage_config": {
"ckpt_path": null,
"target": "ldm.modules.encoders.modules.FrozenCLIPEmbedder"
},
"cond_stage_key": null,
"cond_stage_trainable": false,
"conditioning_key": "hybrid-mv",
"fast_attention": true,
"first_stage_config": {
"params": {
"ddconfig": {
"attn_resolutions": [],
"ch": 128,
"ch_mult": [
1,
2,
4,
4
],
"double_z": true,
"dropout": 0.0,
"in_channels": 3,
"num_res_blocks": 2,
"out_ch": 3,
"resolution": 256,
"z_channels": 4
},
"embed_dim": 4,
"lossconfig": {
"target": "torch.nn.Identity"
},
"monitor": "val/rec_loss"
},
"target": "ldm.models.autoencoder.AutoencoderKL"
},
"first_stage_key": null,
"image_size": 32,
"linear_end": 0.012,
"linear_start": 0.00085,
"log_every_t": 200,
"model_type": "spad",
"monitor": "val/loss_simple_ema",
"num_timesteps_cond": 1,
"resume_path": "data/v1-5-pruned.ckpt",
"scale_factor": 0.18215,
"scheduler_config": {
"params": {
"cycle_lengths": [
10000000000000
],
"f_max": [
1.0
],
"f_min": [
1.0
],
"f_start": [
1e-06
],
"warm_up_steps": [
100
]
},
"target": "ldm.lr_scheduler.LambdaLinearScheduler"
},
"timesteps": 1000,
"transformers_version": "4.42.4",
"unet_config": {
"params": {
"attention_resolutions": [
4,
2,
1
],
"channel_mult": [
1,
2,
4,
4
],
"context_dim": 768,
"image_size": 32,
"in_channels": 4,
"legacy": false,
"model_channels": 320,
"num_heads": 8,
"num_res_blocks": 2,
"out_channels": 4,
"transformer_depth": 1,
"use_checkpoint": true,
"use_spatial_transformer": true
},
"target": "spad.mv_unet.SPADUnetModel"
},
"use_abs_extrinsics": true,
"use_intrinsic": true
}
|