Javtor commited on
Commit
803c944
1 Parent(s): 778a03e

Training in progress, epoch 1

Browse files
.gitignore ADDED
@@ -0,0 +1 @@
 
 
1
+ checkpoint-*/
config.json ADDED
@@ -0,0 +1,279 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "Javtor/biomedical-topic-categorization-2022only",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 1024,
12
+ "id2label": {
13
+ "0": "T001",
14
+ "1": "T002",
15
+ "2": "T004",
16
+ "3": "T005",
17
+ "4": "T007",
18
+ "5": "T008",
19
+ "6": "T010",
20
+ "7": "T011",
21
+ "8": "T012",
22
+ "9": "T013",
23
+ "10": "T014",
24
+ "11": "T015",
25
+ "12": "T016",
26
+ "13": "T017",
27
+ "14": "T018",
28
+ "15": "T019",
29
+ "16": "T020",
30
+ "17": "T022",
31
+ "18": "T023",
32
+ "19": "T024",
33
+ "20": "T025",
34
+ "21": "T026",
35
+ "22": "T028",
36
+ "23": "T029",
37
+ "24": "T030",
38
+ "25": "T031",
39
+ "26": "T032",
40
+ "27": "T033",
41
+ "28": "T034",
42
+ "29": "T037",
43
+ "30": "T038",
44
+ "31": "T039",
45
+ "32": "T040",
46
+ "33": "T041",
47
+ "34": "T042",
48
+ "35": "T043",
49
+ "36": "T044",
50
+ "37": "T045",
51
+ "38": "T046",
52
+ "39": "T047",
53
+ "40": "T048",
54
+ "41": "T049",
55
+ "42": "T050",
56
+ "43": "T051",
57
+ "44": "T052",
58
+ "45": "T053",
59
+ "46": "T054",
60
+ "47": "T055",
61
+ "48": "T056",
62
+ "49": "T057",
63
+ "50": "T058",
64
+ "51": "T059",
65
+ "52": "T060",
66
+ "53": "T061",
67
+ "54": "T062",
68
+ "55": "T063",
69
+ "56": "T064",
70
+ "57": "T065",
71
+ "58": "T066",
72
+ "59": "T067",
73
+ "60": "T068",
74
+ "61": "T069",
75
+ "62": "T070",
76
+ "63": "T071",
77
+ "64": "T072",
78
+ "65": "T073",
79
+ "66": "T074",
80
+ "67": "T075",
81
+ "68": "T077",
82
+ "69": "T078",
83
+ "70": "T079",
84
+ "71": "T080",
85
+ "72": "T081",
86
+ "73": "T082",
87
+ "74": "T083",
88
+ "75": "T085",
89
+ "76": "T086",
90
+ "77": "T087",
91
+ "78": "T089",
92
+ "79": "T090",
93
+ "80": "T091",
94
+ "81": "T092",
95
+ "82": "T093",
96
+ "83": "T094",
97
+ "84": "T095",
98
+ "85": "T096",
99
+ "86": "T097",
100
+ "87": "T098",
101
+ "88": "T099",
102
+ "89": "T100",
103
+ "90": "T101",
104
+ "91": "T102",
105
+ "92": "T103",
106
+ "93": "T104",
107
+ "94": "T109",
108
+ "95": "T114",
109
+ "96": "T116",
110
+ "97": "T120",
111
+ "98": "T121",
112
+ "99": "T122",
113
+ "100": "T123",
114
+ "101": "T125",
115
+ "102": "T126",
116
+ "103": "T127",
117
+ "104": "T129",
118
+ "105": "T130",
119
+ "106": "T131",
120
+ "107": "T167",
121
+ "108": "T168",
122
+ "109": "T169",
123
+ "110": "T170",
124
+ "111": "T171",
125
+ "112": "T184",
126
+ "113": "T185",
127
+ "114": "T190",
128
+ "115": "T191",
129
+ "116": "T192",
130
+ "117": "T194",
131
+ "118": "T195",
132
+ "119": "T196",
133
+ "120": "T197",
134
+ "121": "T200",
135
+ "122": "T201",
136
+ "123": "T204"
137
+ },
138
+ "initializer_range": 0.02,
139
+ "intermediate_size": 4096,
140
+ "label2id": {
141
+ "T001": 0,
142
+ "T002": 1,
143
+ "T004": 2,
144
+ "T005": 3,
145
+ "T007": 4,
146
+ "T008": 5,
147
+ "T010": 6,
148
+ "T011": 7,
149
+ "T012": 8,
150
+ "T013": 9,
151
+ "T014": 10,
152
+ "T015": 11,
153
+ "T016": 12,
154
+ "T017": 13,
155
+ "T018": 14,
156
+ "T019": 15,
157
+ "T020": 16,
158
+ "T022": 17,
159
+ "T023": 18,
160
+ "T024": 19,
161
+ "T025": 20,
162
+ "T026": 21,
163
+ "T028": 22,
164
+ "T029": 23,
165
+ "T030": 24,
166
+ "T031": 25,
167
+ "T032": 26,
168
+ "T033": 27,
169
+ "T034": 28,
170
+ "T037": 29,
171
+ "T038": 30,
172
+ "T039": 31,
173
+ "T040": 32,
174
+ "T041": 33,
175
+ "T042": 34,
176
+ "T043": 35,
177
+ "T044": 36,
178
+ "T045": 37,
179
+ "T046": 38,
180
+ "T047": 39,
181
+ "T048": 40,
182
+ "T049": 41,
183
+ "T050": 42,
184
+ "T051": 43,
185
+ "T052": 44,
186
+ "T053": 45,
187
+ "T054": 46,
188
+ "T055": 47,
189
+ "T056": 48,
190
+ "T057": 49,
191
+ "T058": 50,
192
+ "T059": 51,
193
+ "T060": 52,
194
+ "T061": 53,
195
+ "T062": 54,
196
+ "T063": 55,
197
+ "T064": 56,
198
+ "T065": 57,
199
+ "T066": 58,
200
+ "T067": 59,
201
+ "T068": 60,
202
+ "T069": 61,
203
+ "T070": 62,
204
+ "T071": 63,
205
+ "T072": 64,
206
+ "T073": 65,
207
+ "T074": 66,
208
+ "T075": 67,
209
+ "T077": 68,
210
+ "T078": 69,
211
+ "T079": 70,
212
+ "T080": 71,
213
+ "T081": 72,
214
+ "T082": 73,
215
+ "T083": 74,
216
+ "T085": 75,
217
+ "T086": 76,
218
+ "T087": 77,
219
+ "T089": 78,
220
+ "T090": 79,
221
+ "T091": 80,
222
+ "T092": 81,
223
+ "T093": 82,
224
+ "T094": 83,
225
+ "T095": 84,
226
+ "T096": 85,
227
+ "T097": 86,
228
+ "T098": 87,
229
+ "T099": 88,
230
+ "T100": 89,
231
+ "T101": 90,
232
+ "T102": 91,
233
+ "T103": 92,
234
+ "T104": 93,
235
+ "T109": 94,
236
+ "T114": 95,
237
+ "T116": 96,
238
+ "T120": 97,
239
+ "T121": 98,
240
+ "T122": 99,
241
+ "T123": 100,
242
+ "T125": 101,
243
+ "T126": 102,
244
+ "T127": 103,
245
+ "T129": 104,
246
+ "T130": 105,
247
+ "T131": 106,
248
+ "T167": 107,
249
+ "T168": 108,
250
+ "T169": 109,
251
+ "T170": 110,
252
+ "T171": 111,
253
+ "T184": 112,
254
+ "T185": 113,
255
+ "T190": 114,
256
+ "T191": 115,
257
+ "T192": 116,
258
+ "T194": 117,
259
+ "T195": 118,
260
+ "T196": 119,
261
+ "T197": 120,
262
+ "T200": 121,
263
+ "T201": 122,
264
+ "T204": 123
265
+ },
266
+ "layer_norm_eps": 1e-12,
267
+ "max_position_embeddings": 512,
268
+ "model_type": "bert",
269
+ "num_attention_heads": 16,
270
+ "num_hidden_layers": 24,
271
+ "pad_token_id": 0,
272
+ "position_embedding_type": "absolute",
273
+ "problem_type": "multi_label_classification",
274
+ "torch_dtype": "float32",
275
+ "transformers_version": "4.23.1",
276
+ "type_vocab_size": 2,
277
+ "use_cache": true,
278
+ "vocab_size": 28895
279
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18d2bb82b6bf2bb8e8cfcc0b3ead3b282594158821573b0c07bf56d99dc17d99
3
+ size 1334550581
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "do_basic_tokenize": true,
4
+ "do_lower_case": true,
5
+ "mask_token": "[MASK]",
6
+ "name_or_path": "Javtor/biomedical-topic-categorization-2022only",
7
+ "never_split": null,
8
+ "pad_token": "[PAD]",
9
+ "sep_token": "[SEP]",
10
+ "special_tokens_map_file": null,
11
+ "strip_accents": null,
12
+ "tokenize_chinese_chars": true,
13
+ "tokenizer_class": "BertTokenizer",
14
+ "unk_token": "[UNK]"
15
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:12e160b8dd2fb1c42d0fa9ba231f667d52ba8bbbd169b612ea15748aad3b5697
3
+ size 3451
vocab.txt ADDED
The diff for this file is too large to render. See raw diff