add model
Browse files- config.json +1 -8
- pytorch_model.bin +1 -1
config.json
CHANGED
@@ -2,7 +2,6 @@
|
|
2 |
"architectures": [
|
3 |
"VisionEncoderDecoderModel"
|
4 |
],
|
5 |
-
"block_size": 100,
|
6 |
"decoder": {
|
7 |
"_name_or_path": "distilgpt2",
|
8 |
"_num_labels": 1,
|
@@ -93,7 +92,6 @@
|
|
93 |
"vocab_size": 50257
|
94 |
},
|
95 |
"decoder_start_token_id": 50256,
|
96 |
-
"early_stopping": true,
|
97 |
"encoder": {
|
98 |
"_name_or_path": "Migga/ViT-chess-V4",
|
99 |
"add_cross_attention": false,
|
@@ -4105,15 +4103,10 @@
|
|
4105 |
"use_bfloat16": false
|
4106 |
},
|
4107 |
"is_encoder_decoder": true,
|
4108 |
-
"length_penalty": 2.0,
|
4109 |
-
"max_length": 32,
|
4110 |
"model_type": "vision-encoder-decoder",
|
4111 |
-
"no_repeat_ngram_size": 3,
|
4112 |
-
"num_beams": 4,
|
4113 |
"pad_token_id": 50256,
|
4114 |
"tie_word_embeddings": false,
|
4115 |
"torch_dtype": "float32",
|
4116 |
"transformers_version": null,
|
4117 |
-
"use_auth_token": true
|
4118 |
-
"vocab_size": 50257
|
4119 |
}
|
|
|
2 |
"architectures": [
|
3 |
"VisionEncoderDecoderModel"
|
4 |
],
|
|
|
5 |
"decoder": {
|
6 |
"_name_or_path": "distilgpt2",
|
7 |
"_num_labels": 1,
|
|
|
92 |
"vocab_size": 50257
|
93 |
},
|
94 |
"decoder_start_token_id": 50256,
|
|
|
95 |
"encoder": {
|
96 |
"_name_or_path": "Migga/ViT-chess-V4",
|
97 |
"add_cross_attention": false,
|
|
|
4103 |
"use_bfloat16": false
|
4104 |
},
|
4105 |
"is_encoder_decoder": true,
|
|
|
|
|
4106 |
"model_type": "vision-encoder-decoder",
|
|
|
|
|
4107 |
"pad_token_id": 50256,
|
4108 |
"tie_word_embeddings": false,
|
4109 |
"torch_dtype": "float32",
|
4110 |
"transformers_version": null,
|
4111 |
+
"use_auth_token": true
|
|
|
4112 |
}
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 977708805
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c95beffecf274cb13229d959083c5d09279df1e395171406b2f777adc718affb
|
3 |
size 977708805
|