PGKChaitanya commited on
Commit
6305a97
1 Parent(s): 1cbacdf

Delete Anime_New

Browse files
Anime_New/text_encoder/config.json DELETED
@@ -1,24 +0,0 @@
1
- {
2
- "architectures": [
3
- "CLIPTextModel"
4
- ],
5
- "attention_dropout": 0.0,
6
- "bos_token_id": 0,
7
- "dropout": 0.0,
8
- "eos_token_id": 2,
9
- "hidden_act": "quick_gelu",
10
- "hidden_size": 768,
11
- "initializer_factor": 1.0,
12
- "initializer_range": 0.02,
13
- "intermediate_size": 3072,
14
- "layer_norm_eps": 1e-05,
15
- "max_position_embeddings": 77,
16
- "model_type": "clip_text_model",
17
- "num_attention_heads": 12,
18
- "num_hidden_layers": 12,
19
- "pad_token_id": 1,
20
- "projection_dim": 768,
21
- "torch_dtype": "float32",
22
- "transformers_version": "4.39.3",
23
- "vocab_size": 49408
24
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Anime_New/text_encoder/model.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:9f4a9454ba73c6e616ffe84a5cc3e6686665d8d4876ee49f2c617dc46d8b4c34
3
- size 492265168
 
 
 
 
Anime_New/tokenizer/merges.txt DELETED
The diff for this file is too large to render. See raw diff
 
Anime_New/tokenizer/special_tokens_map.json DELETED
@@ -1,30 +0,0 @@
1
- {
2
- "bos_token": {
3
- "content": "<|startoftext|>",
4
- "lstrip": false,
5
- "normalized": true,
6
- "rstrip": false,
7
- "single_word": false
8
- },
9
- "eos_token": {
10
- "content": "<|endoftext|>",
11
- "lstrip": false,
12
- "normalized": false,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
- "pad_token": {
17
- "content": "<|endoftext|>",
18
- "lstrip": false,
19
- "normalized": false,
20
- "rstrip": false,
21
- "single_word": false
22
- },
23
- "unk_token": {
24
- "content": "<|endoftext|>",
25
- "lstrip": false,
26
- "normalized": false,
27
- "rstrip": false,
28
- "single_word": false
29
- }
30
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Anime_New/tokenizer/tokenizer_config.json DELETED
@@ -1,30 +0,0 @@
1
- {
2
- "add_prefix_space": false,
3
- "added_tokens_decoder": {
4
- "49406": {
5
- "content": "<|startoftext|>",
6
- "lstrip": false,
7
- "normalized": true,
8
- "rstrip": false,
9
- "single_word": false,
10
- "special": true
11
- },
12
- "49407": {
13
- "content": "<|endoftext|>",
14
- "lstrip": false,
15
- "normalized": false,
16
- "rstrip": false,
17
- "single_word": false,
18
- "special": true
19
- }
20
- },
21
- "bos_token": "<|startoftext|>",
22
- "clean_up_tokenization_spaces": true,
23
- "do_lower_case": true,
24
- "eos_token": "<|endoftext|>",
25
- "errors": "replace",
26
- "model_max_length": 77,
27
- "pad_token": "<|endoftext|>",
28
- "tokenizer_class": "CLIPTokenizer",
29
- "unk_token": "<|endoftext|>"
30
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Anime_New/tokenizer/vocab.json DELETED
The diff for this file is too large to render. See raw diff
 
Anime_New/unet/config.json DELETED
@@ -1,67 +0,0 @@
1
- {
2
- "_class_name": "UNet2DConditionModel",
3
- "_diffusers_version": "0.27.2",
4
- "act_fn": "silu",
5
- "addition_embed_type": null,
6
- "addition_embed_type_num_heads": 64,
7
- "addition_time_embed_dim": null,
8
- "attention_head_dim": 8,
9
- "attention_type": "default",
10
- "block_out_channels": [
11
- 320,
12
- 640,
13
- 1280,
14
- 1280
15
- ],
16
- "center_input_sample": false,
17
- "class_embed_type": null,
18
- "class_embeddings_concat": false,
19
- "conv_in_kernel": 3,
20
- "conv_out_kernel": 3,
21
- "cross_attention_dim": 768,
22
- "cross_attention_norm": null,
23
- "down_block_types": [
24
- "CrossAttnDownBlock2D",
25
- "CrossAttnDownBlock2D",
26
- "CrossAttnDownBlock2D",
27
- "DownBlock2D"
28
- ],
29
- "downsample_padding": 1,
30
- "dropout": 0.0,
31
- "dual_cross_attention": false,
32
- "encoder_hid_dim": null,
33
- "encoder_hid_dim_type": null,
34
- "flip_sin_to_cos": true,
35
- "freq_shift": 0,
36
- "in_channels": 4,
37
- "layers_per_block": 2,
38
- "mid_block_only_cross_attention": null,
39
- "mid_block_scale_factor": 1,
40
- "mid_block_type": "UNetMidBlock2DCrossAttn",
41
- "norm_eps": 1e-05,
42
- "norm_num_groups": 32,
43
- "num_attention_heads": null,
44
- "num_class_embeds": null,
45
- "only_cross_attention": false,
46
- "out_channels": 4,
47
- "projection_class_embeddings_input_dim": null,
48
- "resnet_out_scale_factor": 1.0,
49
- "resnet_skip_time_act": false,
50
- "resnet_time_scale_shift": "default",
51
- "reverse_transformer_layers_per_block": null,
52
- "sample_size": 64,
53
- "time_cond_proj_dim": null,
54
- "time_embedding_act_fn": null,
55
- "time_embedding_dim": null,
56
- "time_embedding_type": "positional",
57
- "timestep_post_act": null,
58
- "transformer_layers_per_block": 1,
59
- "up_block_types": [
60
- "UpBlock2D",
61
- "CrossAttnUpBlock2D",
62
- "CrossAttnUpBlock2D",
63
- "CrossAttnUpBlock2D"
64
- ],
65
- "upcast_attention": false,
66
- "use_linear_projection": false
67
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Anime_New/vae/config.json DELETED
@@ -1,33 +0,0 @@
1
- {
2
- "_class_name": "AutoencoderKL",
3
- "_diffusers_version": "0.27.2",
4
- "act_fn": "silu",
5
- "block_out_channels": [
6
- 128,
7
- 256,
8
- 512,
9
- 512
10
- ],
11
- "down_block_types": [
12
- "DownEncoderBlock2D",
13
- "DownEncoderBlock2D",
14
- "DownEncoderBlock2D",
15
- "DownEncoderBlock2D"
16
- ],
17
- "force_upcast": true,
18
- "in_channels": 3,
19
- "latent_channels": 4,
20
- "latents_mean": null,
21
- "latents_std": null,
22
- "layers_per_block": 2,
23
- "norm_num_groups": 32,
24
- "out_channels": 3,
25
- "sample_size": 512,
26
- "scaling_factor": 0.18215,
27
- "up_block_types": [
28
- "UpDecoderBlock2D",
29
- "UpDecoderBlock2D",
30
- "UpDecoderBlock2D",
31
- "UpDecoderBlock2D"
32
- ]
33
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Anime_New/vae/diffusion_pytorch_model.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:2ab67d20c0b8929f301d64ec9b83d939a3e5e57f32d65845b91dac1cd7ebcddc
3
- size 334643268