Spaces:
Runtime error
Runtime error
minor
Browse files- app.py +3 -2
- t2v_enhanced/model_init.py +1 -0
app.py
CHANGED
@@ -76,16 +76,17 @@ def generate(prompt, num_frames, image, model_name_stage1, model_name_stage2, se
|
|
76 |
|
77 |
n_autoreg_gen = (num_frames-8)//8
|
78 |
|
79 |
-
inference_generator = torch.Generator(device="cuda").manual_seed(seed)
|
80 |
-
|
81 |
if model_name_stage1 == "ModelScopeT2V (text to video)":
|
|
|
82 |
short_video = ms_short_gen(prompt, ms_model, inference_generator, t, device)
|
83 |
elif model_name_stage1 == "AnimateDiff (text to video)":
|
|
|
84 |
short_video = ad_short_gen(prompt, ad_model, inference_generator, t, device)
|
85 |
elif model_name_stage1 == "SVD (image to video)":
|
86 |
# For cached examples
|
87 |
if isinstance(image, dict):
|
88 |
image = image["path"]
|
|
|
89 |
short_video = svd_short_gen(image, prompt, svd_model, sdxl_model, inference_generator, t, device)
|
90 |
|
91 |
stream_long_gen(prompt, short_video, n_autoreg_gen, seed, t, image_guidance, name, stream_cli, stream_model)
|
|
|
76 |
|
77 |
n_autoreg_gen = (num_frames-8)//8
|
78 |
|
|
|
|
|
79 |
if model_name_stage1 == "ModelScopeT2V (text to video)":
|
80 |
+
inference_generator = torch.Generator(device=ms_model.device).manual_seed(seed)
|
81 |
short_video = ms_short_gen(prompt, ms_model, inference_generator, t, device)
|
82 |
elif model_name_stage1 == "AnimateDiff (text to video)":
|
83 |
+
inference_generator = torch.Generator(device=ad_model.device).manual_seed(seed)
|
84 |
short_video = ad_short_gen(prompt, ad_model, inference_generator, t, device)
|
85 |
elif model_name_stage1 == "SVD (image to video)":
|
86 |
# For cached examples
|
87 |
if isinstance(image, dict):
|
88 |
image = image["path"]
|
89 |
+
inference_generator = torch.Generator(device=svd_model.device).manual_seed(seed)
|
90 |
short_video = svd_short_gen(image, prompt, svd_model, sdxl_model, inference_generator, t, device)
|
91 |
|
92 |
stream_long_gen(prompt, short_video, n_autoreg_gen, seed, t, image_guidance, name, stream_cli, stream_model)
|
t2v_enhanced/model_init.py
CHANGED
@@ -119,6 +119,7 @@ def init_v2v_model(cfg, device):
|
|
119 |
|
120 |
pipe_enhance.model.autoencoder = pipe_enhance.model.autoencoder.to(device)
|
121 |
pipe_enhance.model.generator = pipe_enhance.model.generator.to(device)
|
|
|
122 |
pipe_enhance.model.negative_y = pipe_enhance.model.negative_y.to(device)
|
123 |
pipe_enhance.model.cfg.max_frames = 10000
|
124 |
return pipe_enhance
|
|
|
119 |
|
120 |
pipe_enhance.model.autoencoder = pipe_enhance.model.autoencoder.to(device)
|
121 |
pipe_enhance.model.generator = pipe_enhance.model.generator.to(device)
|
122 |
+
pipe_enhance.model.generator = pipe_enhance.model.generator.half()
|
123 |
pipe_enhance.model.negative_y = pipe_enhance.model.negative_y.to(device)
|
124 |
pipe_enhance.model.cfg.max_frames = 10000
|
125 |
return pipe_enhance
|