Kendong commited on
Commit
250af62
·
1 Parent(s): ae241d9

Upload 2 files

Browse files
Files changed (2) hide show
  1. zeronvs.ckpt +3 -0
  2. zeronvs_config.yaml +155 -0
zeronvs.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ba72319293a78508c99c177566033ebabc2b93cb7d07f0b7b3e6866c726cc38
3
+ size 15466104283
zeronvs_config.yaml ADDED
@@ -0,0 +1,155 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ base_learning_rate: 0.0001
3
+ target: ldm.models.diffusion.ddpm.LatentDiffusion
4
+ params:
5
+ linear_start: 0.00085
6
+ linear_end: 0.012
7
+ num_timesteps_cond: 1
8
+ log_every_t: 200
9
+ timesteps: 1000
10
+ first_stage_key: image_target
11
+ cond_stage_key: image_cond
12
+ image_size: 32
13
+ channels: 4
14
+ cond_stage_trainable: false
15
+ conditioning_key: hybrid
16
+ monitor: val/loss_simple_ema
17
+ scale_factor: 0.18215
18
+ conditioning_config:
19
+ params:
20
+ mode: 7dof_quantile_scale
21
+ embedding_dim: 19
22
+ depth_model_name: midas
23
+ scheduler_config:
24
+ target: ldm.lr_scheduler.LambdaLinearScheduler
25
+ params:
26
+ warm_up_steps:
27
+ - 100
28
+ cycle_lengths:
29
+ - 10000000000000
30
+ f_start:
31
+ - 1.0e-06
32
+ f_max:
33
+ - 1.0
34
+ f_min:
35
+ - 1.0
36
+ unet_config:
37
+ target: ldm.modules.diffusionmodules.openaimodel.UNetModel
38
+ params:
39
+ image_size: 32
40
+ in_channels: 8
41
+ out_channels: 4
42
+ model_channels: 320
43
+ attention_resolutions:
44
+ - 4
45
+ - 2
46
+ - 1
47
+ num_res_blocks: 2
48
+ channel_mult:
49
+ - 1
50
+ - 2
51
+ - 4
52
+ - 4
53
+ num_heads: 8
54
+ use_spatial_transformer: true
55
+ transformer_depth: 1
56
+ context_dim: 768
57
+ use_checkpoint: true
58
+ legacy: false
59
+ eval_config:
60
+ params:
61
+ scale: 3.0
62
+ ddim_steps: 100
63
+ ddim_eta: 1.0
64
+ lpips_model_path: null
65
+ first_stage_config:
66
+ target: ldm.models.autoencoder.AutoencoderKL
67
+ params:
68
+ embed_dim: 4
69
+ monitor: val/rec_loss
70
+ ddconfig:
71
+ double_z: true
72
+ z_channels: 4
73
+ resolution: 256
74
+ in_channels: 3
75
+ out_ch: 3
76
+ ch: 128
77
+ ch_mult:
78
+ - 1
79
+ - 2
80
+ - 4
81
+ - 4
82
+ num_res_blocks: 2
83
+ attn_resolutions: []
84
+ dropout: 0.0
85
+ lossconfig:
86
+ target: torch.nn.Identity
87
+ cond_stage_config:
88
+ target: ldm.modules.encoders.modules.FrozenCLIPImageEmbedder
89
+ data:
90
+ target: ldm.data.simple.WDSGenericDataModule
91
+ params:
92
+ train_config:
93
+ batch_size: 48
94
+ num_workers: 6
95
+ shuffle_buffer_size: 500
96
+ prefetch_factor: 4
97
+ dataset_config_1:
98
+ dataset_n_shards: 127
99
+ dataset_name: co3d
100
+ views_per_scene: 100
101
+ dataset_n_scenes: 18432
102
+ rate: 0.025
103
+ probability: 0.34
104
+ dataset_url: null
105
+ dataset_config_2:
106
+ dataset_n_shards: 127
107
+ dataset_name: re10k
108
+ views_per_scene: 200
109
+ dataset_n_scenes: 65280
110
+ probability: 0.33
111
+ rate: 0.025
112
+ dataset_url: null
113
+ dataset_config_3:
114
+ dataset_n_shards: 127
115
+ dataset_name: acid
116
+ views_per_scene: 100
117
+ dataset_n_scenes: 12032
118
+ probability: 0.33
119
+ rate: 0.025
120
+ dataset_url: null
121
+ val_config:
122
+ batch_size: 1
123
+ subsample: 1.0
124
+ scene_scale: 1.0
125
+ dataset_n_shards: 1
126
+ dataset_name: co3d
127
+ dataset_n_scenes: 150
128
+ num_workers: 1
129
+ shuffle_buffer_size: 20
130
+ rate: 0.1
131
+ dataset_url: null
132
+ --lightning:
133
+ trainer:
134
+ accumulate_grad_batches: 4
135
+ modelcheckpoint:
136
+ params:
137
+ every_n_train_steps: 2500
138
+ --data:
139
+ params:
140
+ train_config:
141
+ batch_size: 48
142
+ val_config:
143
+ batch_size: 1
144
+ --model:
145
+ params:
146
+ conditioning_config:
147
+ params:
148
+ mode: 7dof_quantile_scale
149
+ embedding_dim: 19
150
+ eval_config:
151
+ params:
152
+ ddim_steps: 100
153
+ base_learning_rate: 0.0001
154
+ --args:
155
+ finetune_from: null