Upload model
Browse files- config.json +7 -7
- model.safetensors +2 -2
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "
|
3 |
"architectures": [
|
4 |
"VisionEncoderDecoderModel"
|
5 |
],
|
@@ -124,8 +124,8 @@
|
|
124 |
"1": "LABEL_1"
|
125 |
},
|
126 |
"image_size": [
|
127 |
-
|
128 |
-
|
129 |
],
|
130 |
"initializer_range": 0.02,
|
131 |
"is_decoder": false,
|
@@ -179,15 +179,15 @@
|
|
179 |
"torch_dtype": null,
|
180 |
"torchscript": false,
|
181 |
"typical_p": 1.0,
|
182 |
-
"use_2d_embeddings":
|
183 |
"use_absolute_embeddings": false,
|
184 |
"use_bfloat16": false,
|
185 |
-
"window_size":
|
186 |
},
|
187 |
"is_encoder_decoder": true,
|
188 |
"model_type": "vision-encoder-decoder",
|
189 |
"pad_token_id": 1,
|
190 |
"tie_word_embeddings": false,
|
191 |
-
"torch_dtype": "
|
192 |
-
"transformers_version": "4.
|
193 |
}
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "texify/checkpoint-12000",
|
3 |
"architectures": [
|
4 |
"VisionEncoderDecoderModel"
|
5 |
],
|
|
|
124 |
"1": "LABEL_1"
|
125 |
},
|
126 |
"image_size": [
|
127 |
+
420,
|
128 |
+
420
|
129 |
],
|
130 |
"initializer_range": 0.02,
|
131 |
"is_decoder": false,
|
|
|
179 |
"torch_dtype": null,
|
180 |
"torchscript": false,
|
181 |
"typical_p": 1.0,
|
182 |
+
"use_2d_embeddings": false,
|
183 |
"use_absolute_embeddings": false,
|
184 |
"use_bfloat16": false,
|
185 |
+
"window_size": 5
|
186 |
},
|
187 |
"is_encoder_decoder": true,
|
188 |
"model_type": "vision-encoder-decoder",
|
189 |
"pad_token_id": 1,
|
190 |
"tie_word_embeddings": false,
|
191 |
+
"torch_dtype": "float16",
|
192 |
+
"transformers_version": "4.36.0"
|
193 |
}
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a0c3a086c03a4fde4fc57f548b37ef25fa233dc58080f5ee80aa7065f91c9916
|
3 |
+
size 625067248
|