shap-e / prior /config.json
yiyixuxu
add model
160c523
raw
history blame
475 Bytes
{
"_class_name": "PriorTransformer",
"_diffusers_version": "0.17.0.dev0",
"act_fn": "gelu",
"additional_embeddings": 0,
"attention_head_dim": 64,
"clip_embedding_dim": 768,
"dropout": 0.0,
"embedding_dim": 1024,
"has_encoder_hidden_states_proj": false,
"has_post_process": false,
"has_prd_embedding": false,
"has_pre_norm": true,
"num_attention_heads": 16,
"num_embeddings": 1024,
"num_layers": 24,
"out_dim": 2048,
"time_embed_dim": 4096
}