tharakaade commited on
Commit
aa42dfe
1 Parent(s): 993b36b

Upload LayoutLMv3ForSequenceClassification

Browse files
Files changed (2) hide show
  1. config.json +210 -0
  2. pytorch_model.bin +3 -0
config.json ADDED
@@ -0,0 +1,210 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "microsoft/layoutlmv3-base",
3
+ "architectures": [
4
+ "LayoutLMv3ForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "coordinate_size": 128,
10
+ "eos_token_id": 2,
11
+ "has_relative_attention_bias": true,
12
+ "has_spatial_attention_bias": true,
13
+ "hidden_act": "gelu",
14
+ "hidden_dropout_prob": 0.1,
15
+ "hidden_size": 768,
16
+ "id2label": {
17
+ "0": "Temp P1",
18
+ "1": "Temp P10",
19
+ "2": "Temp P11",
20
+ "3": "Temp P12",
21
+ "4": "Temp P13",
22
+ "5": "Temp P14",
23
+ "6": "Temp P15",
24
+ "7": "Temp P16",
25
+ "8": "Temp P17",
26
+ "9": "Temp P18",
27
+ "10": "Temp P19",
28
+ "11": "Temp P2",
29
+ "12": "Temp P20",
30
+ "13": "Temp P21",
31
+ "14": "Temp P22",
32
+ "15": "Temp P23",
33
+ "16": "Temp P24",
34
+ "17": "Temp P25",
35
+ "18": "Temp P26",
36
+ "19": "Temp P27",
37
+ "20": "Temp P28",
38
+ "21": "Temp P29",
39
+ "22": "Temp P3",
40
+ "23": "Temp P30",
41
+ "24": "Temp P31",
42
+ "25": "Temp P32",
43
+ "26": "Temp P33",
44
+ "27": "Temp P34",
45
+ "28": "Temp P35",
46
+ "29": "Temp P36",
47
+ "30": "Temp P37",
48
+ "31": "Temp P38",
49
+ "32": "Temp P39",
50
+ "33": "Temp P4",
51
+ "34": "Temp P40",
52
+ "35": "Temp P41",
53
+ "36": "Temp P42",
54
+ "37": "Temp P43",
55
+ "38": "Temp P44",
56
+ "39": "Temp P45",
57
+ "40": "Temp P46",
58
+ "41": "Temp P47",
59
+ "42": "Temp P48",
60
+ "43": "Temp P49",
61
+ "44": "Temp P5",
62
+ "45": "Temp P50",
63
+ "46": "Temp P51",
64
+ "47": "Temp P52",
65
+ "48": "Temp P53",
66
+ "49": "Temp P54",
67
+ "50": "Temp P55",
68
+ "51": "Temp P56",
69
+ "52": "Temp P57",
70
+ "53": "Temp P58",
71
+ "54": "Temp P59",
72
+ "55": "Temp P6",
73
+ "56": "Temp P60",
74
+ "57": "Temp P61",
75
+ "58": "Temp P62",
76
+ "59": "Temp P63",
77
+ "60": "Temp P64",
78
+ "61": "Temp P65",
79
+ "62": "Temp P66",
80
+ "63": "Temp P67",
81
+ "64": "Temp P68",
82
+ "65": "Temp P69",
83
+ "66": "Temp P7",
84
+ "67": "Temp P70",
85
+ "68": "Temp P71",
86
+ "69": "Temp P72",
87
+ "70": "Temp P73",
88
+ "71": "Temp P74",
89
+ "72": "Temp P75",
90
+ "73": "Temp P76",
91
+ "74": "Temp P77",
92
+ "75": "Temp P78",
93
+ "76": "Temp P79",
94
+ "77": "Temp P8",
95
+ "78": "Temp P80",
96
+ "79": "Temp P81",
97
+ "80": "Temp P82",
98
+ "81": "Temp P83",
99
+ "82": "Temp P9"
100
+ },
101
+ "initializer_range": 0.02,
102
+ "input_size": 224,
103
+ "intermediate_size": 3072,
104
+ "label2id": {
105
+ "Temp P1": 0,
106
+ "Temp P10": 1,
107
+ "Temp P11": 2,
108
+ "Temp P12": 3,
109
+ "Temp P13": 4,
110
+ "Temp P14": 5,
111
+ "Temp P15": 6,
112
+ "Temp P16": 7,
113
+ "Temp P17": 8,
114
+ "Temp P18": 9,
115
+ "Temp P19": 10,
116
+ "Temp P2": 11,
117
+ "Temp P20": 12,
118
+ "Temp P21": 13,
119
+ "Temp P22": 14,
120
+ "Temp P23": 15,
121
+ "Temp P24": 16,
122
+ "Temp P25": 17,
123
+ "Temp P26": 18,
124
+ "Temp P27": 19,
125
+ "Temp P28": 20,
126
+ "Temp P29": 21,
127
+ "Temp P3": 22,
128
+ "Temp P30": 23,
129
+ "Temp P31": 24,
130
+ "Temp P32": 25,
131
+ "Temp P33": 26,
132
+ "Temp P34": 27,
133
+ "Temp P35": 28,
134
+ "Temp P36": 29,
135
+ "Temp P37": 30,
136
+ "Temp P38": 31,
137
+ "Temp P39": 32,
138
+ "Temp P4": 33,
139
+ "Temp P40": 34,
140
+ "Temp P41": 35,
141
+ "Temp P42": 36,
142
+ "Temp P43": 37,
143
+ "Temp P44": 38,
144
+ "Temp P45": 39,
145
+ "Temp P46": 40,
146
+ "Temp P47": 41,
147
+ "Temp P48": 42,
148
+ "Temp P49": 43,
149
+ "Temp P5": 44,
150
+ "Temp P50": 45,
151
+ "Temp P51": 46,
152
+ "Temp P52": 47,
153
+ "Temp P53": 48,
154
+ "Temp P54": 49,
155
+ "Temp P55": 50,
156
+ "Temp P56": 51,
157
+ "Temp P57": 52,
158
+ "Temp P58": 53,
159
+ "Temp P59": 54,
160
+ "Temp P6": 55,
161
+ "Temp P60": 56,
162
+ "Temp P61": 57,
163
+ "Temp P62": 58,
164
+ "Temp P63": 59,
165
+ "Temp P64": 60,
166
+ "Temp P65": 61,
167
+ "Temp P66": 62,
168
+ "Temp P67": 63,
169
+ "Temp P68": 64,
170
+ "Temp P69": 65,
171
+ "Temp P7": 66,
172
+ "Temp P70": 67,
173
+ "Temp P71": 68,
174
+ "Temp P72": 69,
175
+ "Temp P73": 70,
176
+ "Temp P74": 71,
177
+ "Temp P75": 72,
178
+ "Temp P76": 73,
179
+ "Temp P77": 74,
180
+ "Temp P78": 75,
181
+ "Temp P79": 76,
182
+ "Temp P8": 77,
183
+ "Temp P80": 78,
184
+ "Temp P81": 79,
185
+ "Temp P82": 80,
186
+ "Temp P83": 81,
187
+ "Temp P9": 82
188
+ },
189
+ "layer_norm_eps": 1e-05,
190
+ "max_2d_position_embeddings": 1024,
191
+ "max_position_embeddings": 514,
192
+ "max_rel_2d_pos": 256,
193
+ "max_rel_pos": 128,
194
+ "model_type": "layoutlmv3",
195
+ "num_attention_heads": 12,
196
+ "num_channels": 3,
197
+ "num_hidden_layers": 12,
198
+ "pad_token_id": 1,
199
+ "patch_size": 16,
200
+ "rel_2d_pos_bins": 64,
201
+ "rel_pos_bins": 32,
202
+ "second_input_size": 112,
203
+ "shape_size": 128,
204
+ "text_embed": true,
205
+ "torch_dtype": "float32",
206
+ "transformers_version": "4.28.1",
207
+ "type_vocab_size": 1,
208
+ "visual_embed": true,
209
+ "vocab_size": 50265
210
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87a7b9c38991dc2561d580d3e7d8f4d461a2b1f34926a6dded4226af52584490
3
+ size 504001585