{ "_class_name": "PriorTransformer", "_diffusers_version": "0.20.2", "added_emb_type": "prd", "additional_embeddings": 3, "attention_head_dim": 32, "clip_embed_dim": null, "dropout": 0.0, "embedding_dim": 768, "embedding_proj_dim": null, "embedding_proj_norm_type": null, "encoder_hid_proj_type": "linear", "norm_in_type": null, "num_attention_heads": 16, "num_embeddings": 77, "num_layers": 10 }