Linoy Tsaban commited on
Commit
0b1c448
1 Parent(s): e62a915

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +11 -8
app.py CHANGED
@@ -64,9 +64,11 @@ def prep(config):
64
  model_key = "stabilityai/stable-diffusion-2-depth"
65
  toy_scheduler = DDIMScheduler.from_pretrained(model_key, subfolder="scheduler")
66
  toy_scheduler.set_timesteps(config["save_steps"])
 
67
  timesteps_to_save, num_inference_steps = get_timesteps(toy_scheduler, num_inference_steps=config["save_steps"],
68
  strength=1.0,
69
  device=device)
 
70
 
71
  # seed_everything(config["seed"])
72
  if not config["frames"]: # original non demo setting
@@ -109,7 +111,7 @@ def preprocess_and_invert(input_video,
109
  randomize_seed,
110
  do_inversion,
111
  # save_dir: str = "latents",
112
- steps: int = 500,
113
  n_timesteps = 50,
114
  batch_size: int = 8,
115
  n_frames: int = 40,
@@ -119,7 +121,7 @@ def preprocess_and_invert(input_video,
119
  sd_version = "2.1"
120
  height = 512
121
  weidth: int = 512
122
-
123
  if do_inversion or randomize_seed:
124
  preprocess_config = {}
125
  preprocess_config['H'] = height
@@ -128,7 +130,7 @@ def preprocess_and_invert(input_video,
128
  preprocess_config['sd_version'] = sd_version
129
  preprocess_config['steps'] = steps
130
  preprocess_config['batch_size'] = batch_size
131
- preprocess_config['save_steps'] = n_timesteps
132
  preprocess_config['n_frames'] = n_frames
133
  preprocess_config['seed'] = seed
134
  preprocess_config['inversion_prompt'] = inversion_prompt
@@ -141,6 +143,8 @@ def preprocess_and_invert(input_video,
141
  seed_everything(seed)
142
 
143
  frames, latents, total_inverted_latents, rgb_reconstruction = prep(preprocess_config)
 
 
144
  frames = gr.State(value=frames)
145
  latents = gr.State(value=latents)
146
  inverted_latents = gr.State(value=total_inverted_latents)
@@ -173,7 +177,7 @@ def edit_with_pnp(input_video,
173
 
174
  config["sd_version"] = "2.1"
175
  config["device"] = device
176
- config["n_timesteps"] = n_timesteps
177
  config["n_frames"] = n_frames
178
  config["batch_size"] = batch_size
179
  config["guidance_scale"] = gudiance_scale
@@ -194,6 +198,7 @@ def edit_with_pnp(input_video,
194
  randomize_seed,
195
  do_inversion,
196
  steps,
 
197
  batch_size,
198
  n_frames,
199
  inversion_prompt)
@@ -272,7 +277,7 @@ with gr.Blocks(css="style.css") as demo:
272
  randomize_seed = gr.Checkbox(label='Randomize seed', value=False)
273
  gudiance_scale = gr.Slider(label='Guidance Scale', minimum=1, maximum=30,
274
  value=7.5, step=0.5, interactive=True)
275
- steps = gr.Slider(label='Inversion steps', minimum=100, maximum=500,
276
  value=500, step=1, interactive=True)
277
 
278
  with gr.Column(min_width=100):
@@ -282,7 +287,7 @@ with gr.Blocks(css="style.css") as demo:
282
  n_frames = gr.Slider(label='Num frames', minimum=2, maximum=200,
283
  value=24, step=1, interactive=True)
284
  n_timesteps = gr.Slider(label='Diffusion steps', minimum=25, maximum=100,
285
- value=30, step=1, interactive=True)
286
  n_fps = gr.Slider(label='Frames per second', minimum=1, maximum=60,
287
  value=10, step=1, interactive=True)
288
 
@@ -351,7 +356,5 @@ with gr.Blocks(css="style.css") as demo:
351
  outputs=[output_video]
352
  )
353
 
354
-
355
-
356
  demo.queue()
357
  demo.launch()
 
64
  model_key = "stabilityai/stable-diffusion-2-depth"
65
  toy_scheduler = DDIMScheduler.from_pretrained(model_key, subfolder="scheduler")
66
  toy_scheduler.set_timesteps(config["save_steps"])
67
+ print("config[save_steps]", config["save_steps"])
68
  timesteps_to_save, num_inference_steps = get_timesteps(toy_scheduler, num_inference_steps=config["save_steps"],
69
  strength=1.0,
70
  device=device)
71
+ print("YOOOO timesteps to save", timesteps_to_save)
72
 
73
  # seed_everything(config["seed"])
74
  if not config["frames"]: # original non demo setting
 
111
  randomize_seed,
112
  do_inversion,
113
  # save_dir: str = "latents",
114
+ steps,
115
  n_timesteps = 50,
116
  batch_size: int = 8,
117
  n_frames: int = 40,
 
121
  sd_version = "2.1"
122
  height = 512
123
  weidth: int = 512
124
+ print("n timesteps", n_timesteps)
125
  if do_inversion or randomize_seed:
126
  preprocess_config = {}
127
  preprocess_config['H'] = height
 
130
  preprocess_config['sd_version'] = sd_version
131
  preprocess_config['steps'] = steps
132
  preprocess_config['batch_size'] = batch_size
133
+ preprocess_config['save_steps'] = int(n_timesteps)
134
  preprocess_config['n_frames'] = n_frames
135
  preprocess_config['seed'] = seed
136
  preprocess_config['inversion_prompt'] = inversion_prompt
 
143
  seed_everything(seed)
144
 
145
  frames, latents, total_inverted_latents, rgb_reconstruction = prep(preprocess_config)
146
+ print(total_inverted_latents.keys())
147
+ print(len(total_inverted_latents.keys()))
148
  frames = gr.State(value=frames)
149
  latents = gr.State(value=latents)
150
  inverted_latents = gr.State(value=total_inverted_latents)
 
177
 
178
  config["sd_version"] = "2.1"
179
  config["device"] = device
180
+ config["n_timesteps"] = int(n_timesteps)
181
  config["n_frames"] = n_frames
182
  config["batch_size"] = batch_size
183
  config["guidance_scale"] = gudiance_scale
 
198
  randomize_seed,
199
  do_inversion,
200
  steps,
201
+ n_timesteps,
202
  batch_size,
203
  n_frames,
204
  inversion_prompt)
 
277
  randomize_seed = gr.Checkbox(label='Randomize seed', value=False)
278
  gudiance_scale = gr.Slider(label='Guidance Scale', minimum=1, maximum=30,
279
  value=7.5, step=0.5, interactive=True)
280
+ steps = gr.Slider(label='Inversion steps', minimum=10, maximum=500,
281
  value=500, step=1, interactive=True)
282
 
283
  with gr.Column(min_width=100):
 
287
  n_frames = gr.Slider(label='Num frames', minimum=2, maximum=200,
288
  value=24, step=1, interactive=True)
289
  n_timesteps = gr.Slider(label='Diffusion steps', minimum=25, maximum=100,
290
+ value=50, step=25, interactive=True)
291
  n_fps = gr.Slider(label='Frames per second', minimum=1, maximum=60,
292
  value=10, step=1, interactive=True)
293
 
 
356
  outputs=[output_video]
357
  )
358
 
 
 
359
  demo.queue()
360
  demo.launch()