reza-alipour commited on
Commit
c5e8821
1 Parent(s): 48f31c9

Upload folder using huggingface_hub

Browse files
checkpoint-20000/ema_model/config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "MaskGiTUViT_v2",
3
+ "_version": "0.0.1",
4
+ "add_cond_embeds": true,
5
+ "add_micro_cond_embeds": true,
6
+ "attention_dropout": 0.0,
7
+ "block_num_heads": 12,
8
+ "block_out_channels": [
9
+ 768
10
+ ],
11
+ "codebook_size": 8192,
12
+ "cond_embed_dim": 768,
13
+ "decay": 0.9999,
14
+ "encoder_hidden_size": 768,
15
+ "force_down_up_sample": true,
16
+ "hidden_dropout": 0.0,
17
+ "hidden_size": 1024,
18
+ "in_channels": 768,
19
+ "intermediate_size": 2816,
20
+ "inv_gamma": 1.0,
21
+ "layer_norm_eps": 1e-06,
22
+ "ln_elementwise_affine": true,
23
+ "mask_token_id": 8255,
24
+ "micro_cond_embed_dim": 1280,
25
+ "micro_cond_encode_dim": 256,
26
+ "min_decay": 0.0,
27
+ "norm_type": "rmsnorm",
28
+ "num_attention_heads": 16,
29
+ "num_hidden_layers": 22,
30
+ "num_res_blocks": 3,
31
+ "optimization_step": 20000,
32
+ "power": 0.6666666666666666,
33
+ "update_after_step": 0,
34
+ "use_bias": false,
35
+ "use_ema_warmup": false,
36
+ "use_fused_mlp": false,
37
+ "use_fused_residual_norm": false,
38
+ "vocab_size": 8256
39
+ }
checkpoint-20000/ema_model/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:718d24464fd06b25a963536a414bb82a3d70147f3b79b77d1c67d1c36e1da49a
3
+ size 2433247453
checkpoint-20000/metadata.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"global_step": 20000}
checkpoint-20000/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:53f3eb4abd57a9c56df1ca6567c9e44b23fbabc32a1612e284ae28a51170cb26
3
+ size 4866486533
checkpoint-20000/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3f1e38282bfe65fe3527efc39c65c975b1ed7a4ef1cb9085d0fef5c816f685a
3
+ size 2433254429
checkpoint-20000/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a0152c9c1d1a20236b126f2923e09cf31f0f1b294c130d0c22fbf473220c16e
3
+ size 14727
checkpoint-20000/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e817244ffab4a7ed0b1eefb86058e0e5beed0c56f9485ac4b5a4e2e445b97eb0
3
+ size 627
checkpoint-20000/unwrapped_model/config.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "MaskGiTUViT_v2",
3
+ "_version": "0.0.1",
4
+ "add_cond_embeds": true,
5
+ "add_micro_cond_embeds": true,
6
+ "attention_dropout": 0.0,
7
+ "block_num_heads": 12,
8
+ "block_out_channels": [
9
+ 768
10
+ ],
11
+ "codebook_size": 8192,
12
+ "cond_embed_dim": 768,
13
+ "encoder_hidden_size": 768,
14
+ "force_down_up_sample": true,
15
+ "hidden_dropout": 0.0,
16
+ "hidden_size": 1024,
17
+ "in_channels": 768,
18
+ "intermediate_size": 2816,
19
+ "layer_norm_eps": 1e-06,
20
+ "ln_elementwise_affine": true,
21
+ "mask_token_id": 8255,
22
+ "micro_cond_embed_dim": 1280,
23
+ "micro_cond_encode_dim": 256,
24
+ "norm_type": "rmsnorm",
25
+ "num_attention_heads": 16,
26
+ "num_hidden_layers": 22,
27
+ "num_res_blocks": 3,
28
+ "use_bias": false,
29
+ "use_fused_mlp": false,
30
+ "use_fused_residual_norm": false,
31
+ "vocab_size": 8256
32
+ }
checkpoint-20000/unwrapped_model/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3f1e38282bfe65fe3527efc39c65c975b1ed7a4ef1cb9085d0fef5c816f685a
3
+ size 2433254429
config.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "MaskGiTUViT_v2",
3
+ "_version": "0.0.1",
4
+ "add_cond_embeds": true,
5
+ "add_micro_cond_embeds": true,
6
+ "attention_dropout": 0.0,
7
+ "block_num_heads": 12,
8
+ "block_out_channels": [
9
+ 768
10
+ ],
11
+ "codebook_size": 8192,
12
+ "cond_embed_dim": 768,
13
+ "encoder_hidden_size": 768,
14
+ "force_down_up_sample": true,
15
+ "hidden_dropout": 0.0,
16
+ "hidden_size": 1024,
17
+ "in_channels": 768,
18
+ "intermediate_size": 2816,
19
+ "layer_norm_eps": 1e-06,
20
+ "ln_elementwise_affine": true,
21
+ "mask_token_id": 8255,
22
+ "micro_cond_embed_dim": 1280,
23
+ "micro_cond_encode_dim": 256,
24
+ "norm_type": "rmsnorm",
25
+ "num_attention_heads": 16,
26
+ "num_hidden_layers": 22,
27
+ "num_res_blocks": 3,
28
+ "use_bias": false,
29
+ "use_fused_mlp": false,
30
+ "use_fused_residual_norm": false,
31
+ "vocab_size": 8256
32
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:42266f406d15de5bbb8f79bfd5532543af874bfe641f309a51aa28b429ee605a
3
+ size 2433254429