Spaces:
Paused
Paused
Rename LTX-Video/ltx_video/pipelines/pipeline_ltx_video (1).py to LTX-Video/ltx_video/pipelines/pipeline_ltx_video.py
Browse files
LTX-Video/ltx_video/pipelines/{pipeline_ltx_video (1).py → pipeline_ltx_video.py}
RENAMED
|
@@ -203,6 +203,7 @@ class ConditioningItem:
|
|
| 203 |
media_x: Optional[int] = None
|
| 204 |
media_y: Optional[int] = None
|
| 205 |
|
|
|
|
| 206 |
def encode_conditioning_item(
|
| 207 |
self,
|
| 208 |
raw_item: Union[Image.Image, torch.Tensor],
|
|
@@ -250,6 +251,8 @@ def encode_conditioning_item(
|
|
| 250 |
else:
|
| 251 |
raise TypeError(f"Tipo não suportado: {type(raw_item)}")
|
| 252 |
|
|
|
|
|
|
|
| 253 |
return ConditioningItem(
|
| 254 |
media_item_latents=latents,
|
| 255 |
media_frame_number=frame_number,
|
|
@@ -1455,6 +1458,10 @@ class LTXVideoPipeline(DiffusionPipeline):
|
|
| 1455 |
media_latents = item.media_item_latents.to(
|
| 1456 |
dtype=init_latents.dtype, device=init_latents.device
|
| 1457 |
)
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1458 |
strength = item.conditioning_strength
|
| 1459 |
frame_idx = item.media_frame_number
|
| 1460 |
|
|
|
|
| 203 |
media_x: Optional[int] = None
|
| 204 |
media_y: Optional[int] = None
|
| 205 |
|
| 206 |
+
|
| 207 |
def encode_conditioning_item(
|
| 208 |
self,
|
| 209 |
raw_item: Union[Image.Image, torch.Tensor],
|
|
|
|
| 251 |
else:
|
| 252 |
raise TypeError(f"Tipo não suportado: {type(raw_item)}")
|
| 253 |
|
| 254 |
+
print (f"ltx encode_conditioning_item latents {latents.shape}")
|
| 255 |
+
|
| 256 |
return ConditioningItem(
|
| 257 |
media_item_latents=latents,
|
| 258 |
media_frame_number=frame_number,
|
|
|
|
| 1458 |
media_latents = item.media_item_latents.to(
|
| 1459 |
dtype=init_latents.dtype, device=init_latents.device
|
| 1460 |
)
|
| 1461 |
+
|
| 1462 |
+
|
| 1463 |
+
print (f"prepare_conditioning media_latents {latents.shape}")
|
| 1464 |
+
|
| 1465 |
strength = item.conditioning_strength
|
| 1466 |
frame_idx = item.media_frame_number
|
| 1467 |
|