Spaces:
Runtime error
Runtime error
num_frames = 1 | |
image_size = (256, 256) | |
# Define dataset | |
dataset = dict( | |
type="VideoTextDataset", | |
data_path=None, | |
num_frames=num_frames, | |
frame_interval=1, | |
image_size=image_size, | |
) | |
# Define acceleration | |
num_workers = 16 | |
dtype = "bf16" | |
grad_checkpoint = True | |
plugin = "zero2" | |
# Define model | |
model = dict( | |
type="VideoAutoencoderPipeline", | |
freeze_vae_2d=True, | |
from_pretrained=None, | |
cal_loss=True, | |
vae_2d=dict( | |
type="VideoAutoencoderKL", | |
from_pretrained="PixArt-alpha/pixart_sigma_sdxlvae_T5_diffusers", | |
subfolder="vae", | |
local_files_only=True, | |
), | |
vae_temporal=dict( | |
type="VAE_Temporal_SD", | |
from_pretrained=None, | |
), | |
) | |
# loss weights | |
perceptual_loss_weight = 0.0 # use vgg is not None and more than 0 | |
kl_loss_weight = 1e-6 | |
mixed_image_ratio = 0.1 | |
use_real_rec_loss = False | |
use_z_rec_loss = True | |
use_image_identity_loss = True | |
# Others | |
seed = 42 | |
outputs = "outputs" | |
wandb = False | |
epochs = 100 | |
log_every = 1 | |
ckpt_every = 1000 | |
load = None | |
batch_size = 1 | |
lr = 1e-5 | |
grad_clip = 1.0 | |