Kurokabe's picture
Create README.md
167177d
GANime model trained on the Kimetsu no Yaiba dataset on Tensorflow.
The model internaly uses a Huggingface GPT2-large transformer to generate videos based on the first and last frame.
The parameters used for training are the following:
```yaml
model:
transformer_config:
remaining_frames_method: "own_embeddings"
transformer_type: "gpt2-large"
first_stage_config:
vqvae_config:
beta: 0.25
num_embeddings: 50257
embedding_dim: 128
autoencoder_config:
z_channels: 512
channels: 32
channels_multiplier:
- 2
- 4
- 8
- 8
num_res_blocks: 1
attention_resolution:
- 16
resolution: 128
dropout: 0.0
discriminator_config:
num_layers: 3
filters: 64
loss_config:
discriminator:
loss: "hinge"
factor: 1.0
iter_start: 16200
weight: 0.3
vqvae:
codebook_weight: 1.0
perceptual_weight: 4.0
perceptual_loss: "vgg19"
train:
batch_size: 64
accumulation_size: 1
n_epochs: 10000
len_x_train: 28213
warmup_epoch_percentage: 0.15
lr_start: 1e-5
lr_max: 2.5e-4
perceptual_loss_weight: 1.0
n_frames_before: 1
stop_ground_truth_after_epoch: 1000
```
Implementation and documentation can be found [here](https://github.com/Kurokabe/GANime)