{ "_class_name": "PriorTransformer", "_diffusers_version": "0.18.0.dev0", "added_emb_type": null, "additional_embeddings": 0, "attention_head_dim": 128, "clip_embed_dim": 2048, "dropout": 0.0, "embedding_dim": 1024, "embedding_proj_dim": 1024, "embedding_proj_norm_type": "layer", "encoder_hid_proj_type": null, "norm_in_type": "layer", "num_attention_heads": 8, "num_embeddings": 1024, "num_layers": 24, "time_embed_act_fn": "gelu", "time_embed_dim": 4096 }