technillogue commited on
Commit
1664e30
1 Parent(s): cb41f3a

fp16 safetensors

Browse files
Files changed (3) hide show
  1. .gitattributes +1 -0
  2. config.json +15 -18
  3. model.safetensors +3 -0
.gitattributes CHANGED
@@ -29,3 +29,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
29
  *.zip filter=lfs diff=lfs merge=lfs -text
30
  *.zst filter=lfs diff=lfs merge=lfs -text
31
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
29
  *.zip filter=lfs diff=lfs merge=lfs -text
30
  *.zst filter=lfs diff=lfs merge=lfs -text
31
  *tfevents* filter=lfs diff=lfs merge=lfs -text
32
+ model.safetensors filter=lfs diff=lfs merge=lfs -text
config.json CHANGED
@@ -1,7 +1,8 @@
1
  {
2
- "_name_or_path": "clip-vit-large-patch14/",
 
3
  "architectures": [
4
- "SafetyChecker"
5
  ],
6
  "initializer_factor": 1.0,
7
  "logit_scale_init_value": 2.6592,
@@ -13,6 +14,7 @@
13
  "architectures": null,
14
  "attention_dropout": 0.0,
15
  "bad_words_ids": null,
 
16
  "bos_token_id": 0,
17
  "chunk_size_feed_forward": 0,
18
  "cross_attention_hidden_size": null,
@@ -60,14 +62,17 @@
60
  "pad_token_id": 1,
61
  "prefix": null,
62
  "problem_type": null,
 
63
  "pruned_heads": {},
64
  "remove_invalid_values": false,
65
  "repetition_penalty": 1.0,
66
  "return_dict": true,
67
  "return_dict_in_generate": false,
68
  "sep_token_id": null,
 
69
  "task_specific_params": null,
70
  "temperature": 1.0,
 
71
  "tie_encoder_decoder": false,
72
  "tie_word_embeddings": true,
73
  "tokenizer_class": null,
@@ -75,18 +80,12 @@
75
  "top_p": 1.0,
76
  "torch_dtype": null,
77
  "torchscript": false,
78
- "transformers_version": "4.21.0.dev0",
79
  "typical_p": 1.0,
80
  "use_bfloat16": false,
81
  "vocab_size": 49408
82
  },
83
- "text_config_dict": {
84
- "hidden_size": 768,
85
- "intermediate_size": 3072,
86
- "num_attention_heads": 12,
87
- "num_hidden_layers": 12
88
- },
89
- "torch_dtype": "float32",
90
  "transformers_version": null,
91
  "vision_config": {
92
  "_name_or_path": "",
@@ -94,6 +93,7 @@
94
  "architectures": null,
95
  "attention_dropout": 0.0,
96
  "bad_words_ids": null,
 
97
  "bos_token_id": null,
98
  "chunk_size_feed_forward": 0,
99
  "cross_attention_hidden_size": null,
@@ -133,6 +133,7 @@
133
  "num_attention_heads": 16,
134
  "num_beam_groups": 1,
135
  "num_beams": 1,
 
136
  "num_hidden_layers": 24,
137
  "num_return_sequences": 1,
138
  "output_attentions": false,
@@ -142,14 +143,17 @@
142
  "patch_size": 14,
143
  "prefix": null,
144
  "problem_type": null,
 
145
  "pruned_heads": {},
146
  "remove_invalid_values": false,
147
  "repetition_penalty": 1.0,
148
  "return_dict": true,
149
  "return_dict_in_generate": false,
150
  "sep_token_id": null,
 
151
  "task_specific_params": null,
152
  "temperature": 1.0,
 
153
  "tie_encoder_decoder": false,
154
  "tie_word_embeddings": true,
155
  "tokenizer_class": null,
@@ -157,15 +161,8 @@
157
  "top_p": 1.0,
158
  "torch_dtype": null,
159
  "torchscript": false,
160
- "transformers_version": "4.21.0.dev0",
161
  "typical_p": 1.0,
162
  "use_bfloat16": false
163
- },
164
- "vision_config_dict": {
165
- "hidden_size": 1024,
166
- "intermediate_size": 4096,
167
- "num_attention_heads": 16,
168
- "num_hidden_layers": 24,
169
- "patch_size": 14
170
  }
171
  }
 
1
  {
2
+ "_commit_hash": "cb41f3a270d63d454d385fc2e4f571c487c253c5",
3
+ "_name_or_path": "CompVis/stable-diffusion-safety-checker",
4
  "architectures": [
5
+ "StableDiffusionSafetyChecker"
6
  ],
7
  "initializer_factor": 1.0,
8
  "logit_scale_init_value": 2.6592,
 
14
  "architectures": null,
15
  "attention_dropout": 0.0,
16
  "bad_words_ids": null,
17
+ "begin_suppress_tokens": null,
18
  "bos_token_id": 0,
19
  "chunk_size_feed_forward": 0,
20
  "cross_attention_hidden_size": null,
 
62
  "pad_token_id": 1,
63
  "prefix": null,
64
  "problem_type": null,
65
+ "projection_dim": 512,
66
  "pruned_heads": {},
67
  "remove_invalid_values": false,
68
  "repetition_penalty": 1.0,
69
  "return_dict": true,
70
  "return_dict_in_generate": false,
71
  "sep_token_id": null,
72
+ "suppress_tokens": null,
73
  "task_specific_params": null,
74
  "temperature": 1.0,
75
+ "tf_legacy_loss": false,
76
  "tie_encoder_decoder": false,
77
  "tie_word_embeddings": true,
78
  "tokenizer_class": null,
 
80
  "top_p": 1.0,
81
  "torch_dtype": null,
82
  "torchscript": false,
83
+ "transformers_version": "4.30.2",
84
  "typical_p": 1.0,
85
  "use_bfloat16": false,
86
  "vocab_size": 49408
87
  },
88
+ "torch_dtype": "float16",
 
 
 
 
 
 
89
  "transformers_version": null,
90
  "vision_config": {
91
  "_name_or_path": "",
 
93
  "architectures": null,
94
  "attention_dropout": 0.0,
95
  "bad_words_ids": null,
96
+ "begin_suppress_tokens": null,
97
  "bos_token_id": null,
98
  "chunk_size_feed_forward": 0,
99
  "cross_attention_hidden_size": null,
 
133
  "num_attention_heads": 16,
134
  "num_beam_groups": 1,
135
  "num_beams": 1,
136
+ "num_channels": 3,
137
  "num_hidden_layers": 24,
138
  "num_return_sequences": 1,
139
  "output_attentions": false,
 
143
  "patch_size": 14,
144
  "prefix": null,
145
  "problem_type": null,
146
+ "projection_dim": 512,
147
  "pruned_heads": {},
148
  "remove_invalid_values": false,
149
  "repetition_penalty": 1.0,
150
  "return_dict": true,
151
  "return_dict_in_generate": false,
152
  "sep_token_id": null,
153
+ "suppress_tokens": null,
154
  "task_specific_params": null,
155
  "temperature": 1.0,
156
+ "tf_legacy_loss": false,
157
  "tie_encoder_decoder": false,
158
  "tie_word_embeddings": true,
159
  "tokenizer_class": null,
 
161
  "top_p": 1.0,
162
  "torch_dtype": null,
163
  "torchscript": false,
164
+ "transformers_version": "4.30.2",
165
  "typical_p": 1.0,
166
  "use_bfloat16": false
 
 
 
 
 
 
 
167
  }
168
  }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08902f19b1cfebd7c989f152fc0507bef6898c706a91d666509383122324b511
3
+ size 608018440