zackli4ai commited on
Commit
2389363
·
verified ·
1 Parent(s): 5426b84

Upload 10 files

Browse files
EmbedNeuralText.mlmodelc/analytics/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c8a3bcbb214b2f51ff88baaa8af3b98cd71a5e78c12ce89a6543a42a780954d
3
+ size 243
EmbedNeuralText.mlmodelc/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9fc594b4a59999068071b784583bcac50b0eca48a793fdd15972a213f9bdefa5
3
+ size 478
EmbedNeuralText.mlmodelc/metadata.json ADDED
@@ -0,0 +1,111 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "metadataOutputVersion" : "3.0",
4
+ "storagePrecision" : "Float16",
5
+ "outputSchema" : [
6
+ {
7
+ "hasShapeFlexibility" : "0",
8
+ "isOptional" : "0",
9
+ "dataType" : "Float16",
10
+ "formattedType" : "MultiArray (Float16 1 × 768)",
11
+ "shortDescription" : "",
12
+ "shape" : "[1, 768]",
13
+ "name" : "text_embeddings",
14
+ "type" : "MultiArray"
15
+ }
16
+ ],
17
+ "modelParameters" : [
18
+
19
+ ],
20
+ "specificationVersion" : 7,
21
+ "mlProgramOperationTypeHistogram" : {
22
+ "Transpose" : 84,
23
+ "Squeeze" : 434,
24
+ "Ios16.gather" : 1,
25
+ "Ios16.softmax" : 144,
26
+ "Ios16.add" : 169,
27
+ "Concat" : 12,
28
+ "Ios16.realDiv" : 2,
29
+ "Tile" : 1,
30
+ "Ios16.reduceSum" : 2,
31
+ "ExpandDims" : 1,
32
+ "Ios16.sub" : 144,
33
+ "Ios16.cast" : 2,
34
+ "Ios16.conv" : 468,
35
+ "Ios16.matmul" : 288,
36
+ "Ios16.gelu" : 12,
37
+ "Ios16.layerNorm" : 25,
38
+ "SliceByIndex" : 36,
39
+ "Ios16.maximum" : 1,
40
+ "Ios16.reshape" : 60,
41
+ "Ios16.mul" : 157,
42
+ "Ios16.reduceL2Norm" : 1
43
+ },
44
+ "computePrecision" : "Mixed (Float16, Int32)",
45
+ "isUpdatable" : "0",
46
+ "stateSchema" : [
47
+
48
+ ],
49
+ "availability" : {
50
+ "macOS" : "13.0",
51
+ "tvOS" : "16.0",
52
+ "visionOS" : "1.0",
53
+ "watchOS" : "9.0",
54
+ "iOS" : "16.0",
55
+ "macCatalyst" : "16.0"
56
+ },
57
+ "modelType" : {
58
+ "name" : "MLModelType_mlProgram"
59
+ },
60
+ "userDefinedMetadata" : {
61
+ "com.github.apple.coremltools.conversion_date" : "2025-11-10",
62
+ "com.github.apple.coremltools.source" : "torch==2.9.0",
63
+ "com.github.apple.coremltools.version" : "9.0",
64
+ "com.github.apple.coremltools.source_dialect" : "TorchScript"
65
+ },
66
+ "inputSchema" : [
67
+ {
68
+ "hasShapeFlexibility" : "0",
69
+ "isOptional" : "0",
70
+ "dataType" : "Int32",
71
+ "formattedType" : "MultiArray (Int32 1 × 256)",
72
+ "shortDescription" : "",
73
+ "shape" : "[1, 256]",
74
+ "name" : "input_ids",
75
+ "type" : "MultiArray"
76
+ },
77
+ {
78
+ "hasShapeFlexibility" : "0",
79
+ "isOptional" : "0",
80
+ "dataType" : "Float16",
81
+ "formattedType" : "MultiArray (Float16 1 × 1 × 1 × 256)",
82
+ "shortDescription" : "",
83
+ "shape" : "[1, 1, 1, 256]",
84
+ "name" : "padding_mask",
85
+ "type" : "MultiArray"
86
+ },
87
+ {
88
+ "hasShapeFlexibility" : "0",
89
+ "isOptional" : "0",
90
+ "dataType" : "Int32",
91
+ "formattedType" : "MultiArray (Int32 1 × 256)",
92
+ "shortDescription" : "",
93
+ "shape" : "[1, 256]",
94
+ "name" : "attention_mask",
95
+ "type" : "MultiArray"
96
+ },
97
+ {
98
+ "hasShapeFlexibility" : "0",
99
+ "isOptional" : "0",
100
+ "dataType" : "Float16",
101
+ "formattedType" : "MultiArray (Float16 1 × 12 × 256 × 256)",
102
+ "shortDescription" : "",
103
+ "shape" : "[1, 12, 256, 256]",
104
+ "name" : "linear_biases",
105
+ "type" : "MultiArray"
106
+ }
107
+ ],
108
+ "generatedClassName" : "JinaClipText",
109
+ "method" : "predict"
110
+ }
111
+ ]
EmbedNeuralText.mlmodelc/model.mil ADDED
The diff for this file is too large to render. See raw diff
 
EmbedNeuralText.mlmodelc/weights/weight.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ef7385e3d958fd39f7c17f77ca4692adce956b0f26567255b4c00e6a2dda77c
3
+ size 274008128
EmbedNeuralVision.mlmodelc/analytics/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f2d067c9c1eddbd0a19aa04670a844b74c418fb82c5ab9aa1dd0563ad57e704
3
+ size 243
EmbedNeuralVision.mlmodelc/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0de02aeeac3c0758909d91f8345d2649f373dc86124706d29d1bc4781c49744
3
+ size 444
EmbedNeuralVision.mlmodelc/metadata.json ADDED
@@ -0,0 +1,99 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "metadataOutputVersion" : "3.0",
4
+ "storagePrecision" : "Float16",
5
+ "outputSchema" : [
6
+ {
7
+ "hasShapeFlexibility" : "0",
8
+ "isOptional" : "0",
9
+ "dataType" : "Float16",
10
+ "formattedType" : "MultiArray (Float16 1 × 768)",
11
+ "shortDescription" : "",
12
+ "shape" : "[1, 768]",
13
+ "name" : "image_embeddings",
14
+ "type" : "MultiArray"
15
+ }
16
+ ],
17
+ "modelParameters" : [
18
+
19
+ ],
20
+ "specificationVersion" : 7,
21
+ "mlProgramOperationTypeHistogram" : {
22
+ "Tile" : 1,
23
+ "Concat" : 301,
24
+ "Ios16.realDiv" : 1,
25
+ "Ios16.mul" : 1020,
26
+ "Ios16.layerNorm" : 49,
27
+ "SliceByIndex" : 577,
28
+ "Stack" : 288,
29
+ "Transpose" : 517,
30
+ "Ios16.conv" : 481,
31
+ "Ios16.add" : 601,
32
+ "Squeeze" : 1152,
33
+ "Ios16.matmul" : 288,
34
+ "Ios16.softmax" : 144,
35
+ "Ios16.reduceL2Norm" : 1,
36
+ "ExpandDims" : 432,
37
+ "Ios16.maximum" : 1,
38
+ "Split" : 288,
39
+ "Ios16.reshape" : 661,
40
+ "Ios16.silu" : 12
41
+ },
42
+ "computePrecision" : "Mixed (Float16, Int32)",
43
+ "isUpdatable" : "0",
44
+ "stateSchema" : [
45
+
46
+ ],
47
+ "availability" : {
48
+ "macOS" : "13.0",
49
+ "tvOS" : "16.0",
50
+ "visionOS" : "1.0",
51
+ "watchOS" : "9.0",
52
+ "iOS" : "16.0",
53
+ "macCatalyst" : "16.0"
54
+ },
55
+ "modelType" : {
56
+ "name" : "MLModelType_mlProgram"
57
+ },
58
+ "userDefinedMetadata" : {
59
+ "com.github.apple.coremltools.conversion_date" : "2025-11-10",
60
+ "com.github.apple.coremltools.source" : "torch==2.9.0",
61
+ "com.github.apple.coremltools.version" : "9.0",
62
+ "com.github.apple.coremltools.source_dialect" : "TorchScript"
63
+ },
64
+ "inputSchema" : [
65
+ {
66
+ "hasShapeFlexibility" : "0",
67
+ "isOptional" : "0",
68
+ "dataType" : "Float16",
69
+ "formattedType" : "MultiArray (Float16 1 × 3 × 224 × 224)",
70
+ "shortDescription" : "",
71
+ "shape" : "[1, 3, 224, 224]",
72
+ "name" : "pixel_values",
73
+ "type" : "MultiArray"
74
+ },
75
+ {
76
+ "hasShapeFlexibility" : "0",
77
+ "isOptional" : "0",
78
+ "dataType" : "Float16",
79
+ "formattedType" : "MultiArray (Float16 1 × 1 × 196 × 64)",
80
+ "shortDescription" : "",
81
+ "shape" : "[1, 1, 196, 64]",
82
+ "name" : "rope_cos",
83
+ "type" : "MultiArray"
84
+ },
85
+ {
86
+ "hasShapeFlexibility" : "0",
87
+ "isOptional" : "0",
88
+ "dataType" : "Float16",
89
+ "formattedType" : "MultiArray (Float16 1 × 1 × 196 × 64)",
90
+ "shortDescription" : "",
91
+ "shape" : "[1, 1, 196, 64]",
92
+ "name" : "rope_sin",
93
+ "type" : "MultiArray"
94
+ }
95
+ ],
96
+ "generatedClassName" : "JinaClipVision",
97
+ "method" : "predict"
98
+ }
99
+ ]
EmbedNeuralVision.mlmodelc/model.mil ADDED
The diff for this file is too large to render. See raw diff
 
EmbedNeuralVision.mlmodelc/weights/weight.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed205327ab4c9947622ce6fa0ac7b53eeb3c1c12cc050a10fec114fcfdbac551
3
+ size 171797440