Transformers
PyTorch
clip
Inference Endpoints
visheratin commited on
Commit
64bd21d
1 Parent(s): 2339f17

Upload model

Browse files
Files changed (2) hide show
  1. config.json +21 -0
  2. pytorch_model.bin +3 -0
config.json ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "./hf/nllb-clip-base",
3
+ "architectures": [
4
+ "NLLBCLIPModel"
5
+ ],
6
+ "initializer_factor": 1.0,
7
+ "logit_scale_init_value": 2.6592,
8
+ "model_type": "clip",
9
+ "projection_dim": 512,
10
+ "text_config": {
11
+ "encoder_layerdrop": 0,
12
+ "model_type": "clip_text_model",
13
+ "vocab_size": 256206
14
+ },
15
+ "torch_dtype": "float32",
16
+ "transformers_version": "4.33.1",
17
+ "vision_config": {
18
+ "dropout": 0.0,
19
+ "model_type": "clip_vision_model"
20
+ }
21
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15beffd1b3727aacce10f73c924430b3524563331cf54cef5224f21af660d33b
3
+ size 2007680393