valhalla commited on
Commit
ec24f93
1 Parent(s): a51f290

add pt model, tokenizer, feat extractor

Browse files
config.json CHANGED
@@ -1,10 +1,12 @@
1
  {
 
2
  "architectures": [
3
- "HybridCLIP"
4
  ],
5
  "freeze_backbones": true,
6
  "initializer_factor": 1.0,
7
- "model_type": "hybrid-clip",
 
8
  "projection_dim": 512,
9
  "seed": 42,
10
  "text_config": {
@@ -17,6 +19,8 @@
17
  "bad_words_ids": null,
18
  "bos_token_id": null,
19
  "chunk_size_feed_forward": 0,
 
 
20
  "decoder_start_token_id": null,
21
  "diversity_penalty": 0.0,
22
  "do_sample": false,
@@ -76,12 +80,13 @@
76
  "top_p": 1.0,
77
  "torch_dtype": null,
78
  "torchscript": false,
79
- "transformers_version": "4.9.0.dev0",
80
  "type_vocab_size": 2,
81
  "use_bfloat16": false,
82
  "use_cache": true,
83
  "vocab_size": 32102
84
  },
 
85
  "transformers_version": null,
86
  "vision_config": {
87
  "_name_or_path": "",
@@ -91,6 +96,7 @@
91
  "bad_words_ids": null,
92
  "bos_token_id": null,
93
  "chunk_size_feed_forward": 0,
 
94
  "decoder_start_token_id": null,
95
  "diversity_penalty": 0.0,
96
  "do_sample": false,
@@ -151,7 +157,7 @@
151
  "top_p": 1.0,
152
  "torch_dtype": null,
153
  "torchscript": false,
154
- "transformers_version": "4.9.0.dev0",
155
  "use_bfloat16": false
156
  }
157
  }
1
  {
2
+ "_name_or_path": "clip-italian",
3
  "architectures": [
4
+ "VisionTextDualEncoderModel"
5
  ],
6
  "freeze_backbones": true,
7
  "initializer_factor": 1.0,
8
+ "logit_scale_init_value": 1,
9
+ "model_type": "vision-text-dual-encoder",
10
  "projection_dim": 512,
11
  "seed": 42,
12
  "text_config": {
19
  "bad_words_ids": null,
20
  "bos_token_id": null,
21
  "chunk_size_feed_forward": 0,
22
+ "classifier_dropout": null,
23
+ "cross_attention_hidden_size": null,
24
  "decoder_start_token_id": null,
25
  "diversity_penalty": 0.0,
26
  "do_sample": false,
80
  "top_p": 1.0,
81
  "torch_dtype": null,
82
  "torchscript": false,
83
+ "transformers_version": "4.13.0.dev0",
84
  "type_vocab_size": 2,
85
  "use_bfloat16": false,
86
  "use_cache": true,
87
  "vocab_size": 32102
88
  },
89
+ "torch_dtype": "float32",
90
  "transformers_version": null,
91
  "vision_config": {
92
  "_name_or_path": "",
96
  "bad_words_ids": null,
97
  "bos_token_id": null,
98
  "chunk_size_feed_forward": 0,
99
+ "cross_attention_hidden_size": null,
100
  "decoder_start_token_id": null,
101
  "diversity_penalty": 0.0,
102
  "do_sample": false,
157
  "top_p": 1.0,
158
  "torch_dtype": null,
159
  "torchscript": false,
160
+ "transformers_version": "4.13.0.dev0",
161
  "use_bfloat16": false
162
  }
163
  }
flax_model.msgpack CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:58b92aa4ee1096e2059c8ab8e88334def7e9cdc142e82a564427be2db150b430
3
- size 795766586
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d33679dc0f9fdfbcdca2c747fd8298590dd52a86668201ce85d4bd1a5685e1f
3
+ size 795766616
preprocessor_config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_size": 224,
3
+ "do_center_crop": true,
4
+ "do_normalize": true,
5
+ "do_resize": true,
6
+ "feature_extractor_type": "CLIPFeatureExtractor",
7
+ "image_mean": [
8
+ 0.48145466,
9
+ 0.4578275,
10
+ 0.40821073
11
+ ],
12
+ "image_std": [
13
+ 0.26862954,
14
+ 0.26130258,
15
+ 0.27577711
16
+ ],
17
+ "resample": 3,
18
+ "size": 224
19
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a41ee21fc1b9415547cb3aa93273a302f6a4bb92a1c1a7ec1c0f99ee1b3d8bce
3
+ size 795915483
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
1
+ {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
1
+ {"do_lower_case": true, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "max_len": 512, "special_tokens_map_file": null, "name_or_path": "dbmdz/bert-base-italian-xxl-uncased", "do_basic_tokenize": true, "never_split": null, "tokenizer_class": "BertTokenizer"}
vocab.txt ADDED
The diff for this file is too large to render. See raw diff