jamarticus commited on
Commit
a1e319a
1 Parent(s): 9087a04

YurtsAI/ner-document-context

Browse files
README.md ADDED
@@ -0,0 +1,347 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: roberta-base
3
+ datasets:
4
+ - YurtsAI/named_entity_recognition_document_context
5
+ language:
6
+ - en
7
+ library_name: span-marker
8
+ metrics:
9
+ - precision
10
+ - recall
11
+ - f1
12
+ pipeline_tag: token-classification
13
+ tags:
14
+ - span-marker
15
+ - token-classification
16
+ - ner
17
+ - named-entity-recognition
18
+ - generated_from_span_marker_trainer
19
+ widget:
20
+ - text: We have Kanye West, Beyoncé, and Taylor Swift performing at the beachside
21
+ park on the island of Maui.
22
+ - text: This book, published by Epic Games and sponsored by the University of Hawaii,
23
+ features recipes inspired by the popular game League of Legends and a foreword
24
+ by renowned food scholar, Dr. Thomas Johnson, a professor at Harvard University.
25
+ - text: The National Institute of Technology has partnered with CafeCorp to provide
26
+ a menu planning template for businesses in the downtown area.
27
+ - text: The marketing efforts for the Chicago Bulls basketball team in Wrigley Park
28
+ were a huge success, with 80% of attendees speaking Spanish.
29
+ - text: The most important thing was to try using the coconut oil from a tiny store
30
+ near the river, and a sprinkle of Japanese spices I learned from my friend who
31
+ speaks fluent Japanese.
32
+ model-index:
33
+ - name: SpanMarker with roberta-base on YurtsAI/named_entity_recognition_document_context
34
+ results:
35
+ - task:
36
+ type: token-classification
37
+ name: Named Entity Recognition
38
+ dataset:
39
+ name: Unknown
40
+ type: YurtsAI/named_entity_recognition_document_context
41
+ split: eval
42
+ metrics:
43
+ - type: f1
44
+ value: 0.3902777777777778
45
+ name: F1
46
+ - type: precision
47
+ value: 0.6189427312775331
48
+ name: Precision
49
+ - type: recall
50
+ value: 0.28498985801217036
51
+ name: Recall
52
+ ---
53
+
54
+ # SpanMarker with roberta-base on YurtsAI/named_entity_recognition_document_context
55
+
56
+ This is a [SpanMarker](https://github.com/tomaarsen/SpanMarkerNER) model trained on the [YurtsAI/named_entity_recognition_document_context](https://huggingface.co/datasets/YurtsAI/named_entity_recognition_document_context) dataset that can be used for Named Entity Recognition. This SpanMarker model uses [roberta-base](https://huggingface.co/roberta-base) as the underlying encoder.
57
+
58
+ ## Model Details
59
+
60
+ ### Model Description
61
+ - **Model Type:** SpanMarker
62
+ - **Encoder:** [roberta-base](https://huggingface.co/roberta-base)
63
+ - **Maximum Sequence Length:** 256 tokens
64
+ - **Maximum Entity Length:** 11 words
65
+ - **Training Dataset:** [YurtsAI/named_entity_recognition_document_context](https://huggingface.co/datasets/YurtsAI/named_entity_recognition_document_context)
66
+ - **Language:** en
67
+ <!-- - **License:** Unknown -->
68
+
69
+ ### Model Sources
70
+
71
+ - **Repository:** [SpanMarker on GitHub](https://github.com/tomaarsen/SpanMarkerNER)
72
+ - **Thesis:** [SpanMarker For Named Entity Recognition](https://raw.githubusercontent.com/tomaarsen/SpanMarkerNER/main/thesis.pdf)
73
+
74
+ ### Model Labels
75
+ | Label | Examples |
76
+ |:-----------------------------------------|:------------------------------------------------------------------------------------------------------------------------------|
77
+ | art-broadcastprogram | "television program", "Origin of the Gods", "reality show" |
78
+ | art-film | "a video of a successful grant proposal", "'The Matrix '", "film crew" |
79
+ | art-music | "a new album by Beyoncé", "Yesterday by The Beatles", "favorite music CD" |
80
+ | art-other | "art therapy", "play", "Mona Lisa" |
81
+ | art-painting | "vibrant street art scene", "through art", "painting" |
82
+ | art-writtenart | "'The Lost Gods '", "Book 1", "environmental science book" |
83
+ | building-airport | "airport", "major airport", "an airport" |
84
+ | building-hospital | "New York hospital", "local hospital", "hospital" |
85
+ | building-hotel | "hotel", "new hotel in Austin", "a giant hotel" |
86
+ | building-library | "new library", "library", "new , state-of-the-art library" |
87
+ | building-other | "10-story building", "headquarters building", "factory building" |
88
+ | building-restaurant | "new restaurant", "our upscale restaurant", "restaurant" |
89
+ | building-sportsfacility | "sports facility", "Union Park Sports Complex", "city 's sports center" |
90
+ | building-theater | "the local theater", "theater in downtown", "theater" |
91
+ | datetime-absolute | "January 10 , 2020", "January 17 , 2025 at 14:00", "March 25th" |
92
+ | datetime-authored | "2023-02-22", "2019-04-15", "2020-02-15" |
93
+ | datetime-range | "2010-2015", "Q4 2019", "Friday to Sunday" |
94
+ | datetime-relative | "next week 's appointment", "last Saturday", "next week" |
95
+ | event-attack/battle/war/militaryconflict | "attacks/wars", "The", "A" |
96
+ | event-disaster | "My", "To", "disaster" |
97
+ | event-election | "the election for the mayor", "upcoming election", "election season" |
98
+ | event-other | "conference", "annual 4th of july BBQ", "charity gala" |
99
+ | event-protest | "protest", "protest last saturday", "protest rally" |
100
+ | event-sportsevent | "sports event", "annual tennis tournament", "biggest sports event of the year" |
101
+ | location-bodiesofwater | "ocean", "Lake Como", "Lake Michigan" |
102
+ | location-gpe | "Italy", "Texas", "city" |
103
+ | location-island | "Island Radio", "Caribbean island", "island" |
104
+ | location-mountain | "mountain terrain", "the mountain", "mountain" |
105
+ | location-other | "low-lying areas of the city", "advertising hub", "backyard" |
106
+ | location-park | "park", "location-park", "the park" |
107
+ | location-road/railway/highway/transit | "Greyhound network", "road", "train journey" |
108
+ | organization-company | "local company", "Verizon", "a company" |
109
+ | organization-education | "Harvard University", "UW", "University of Arizona" |
110
+ | organization-government/governmentagency | "Red Cross", "local government", "SEC" |
111
+ | organization-media/newspaper | "The New York Times", "media organizations", "Army Times" |
112
+ | organization-other | "Cognizant", "Better World Foundation", "conservation organization" |
113
+ | organization-politicalparty | "Spaceship of Progress Party", "Libertarian Party", "Green Party" |
114
+ | organization-religion | "local church", "the power of prayer", "diamatists" |
115
+ | organization-showorganization | "Royal Shakespeare Company", "Earth 's Edge Theater Company", "Cosmic Theater group" |
116
+ | organization-sportsleague | "International Swimming Federation", "NBA league", "NFL" |
117
+ | organization-sportsteam | "soccer team", "Syracuse Orange football team", "Seattle Seahawks" |
118
+ | other-astronomything | "latest discoveries in the field of astronomy", "Galactic Conference Best Recipe Award-winning recipe book", "astronomy camp" |
119
+ | other-award | "other-award", "annual tech show awards", "Nobel Peace Prize" |
120
+ | other-biologything | "salmon 's gene for cold adaptation", "terrain", "the forces that drive you" |
121
+ | other-chemicalthing | "Overall", "The", "In" |
122
+ | other-currency | "US dollars", "Japanese Yen", "$ 500,000" |
123
+ | other-disease | "malaria", "type 1 diabetes", "the common cold" |
124
+ | other-educationaldegree | "master 's degree", "thesis", "Ph.D in food science" |
125
+ | other-god | "Peter Pan", "divine", "Zeus the god" |
126
+ | other-language | "English", "Amharic", "Sanskrit" |
127
+ | other-law | "legislation", "professorial separation laws", "Clean Air Act" |
128
+ | other-livingthing | "We", "To", "flowers" |
129
+ | other-medical | "antibiotics", "medical treatment", "necessary testing protocols" |
130
+ | person-actor | "Emma Stone", "Dr. Steven Spielberg", "Jennifer Lawrence" |
131
+ | person-artist/author | "Chuck Close", "artist 's new album", "Jane Smith" |
132
+ | person-athlete | "athlete friend", "LeBron James", "John and Sally" |
133
+ | person-director | "John Oliver", "favorite director", "Dr. Johnson" |
134
+ | person-other | "your", "HR representative", "therapist or counselor" |
135
+ | person-politician | "To", "At", "Secretary of State" |
136
+ | person-scholar | "Dr. John Smith", "Dr. Johnson", "a scholar of comparative religion" |
137
+ | person-soldier | "veterans", "the brave soldiers", "a soldier" |
138
+ | product-airplane | "Cessna 172", "company 's fleet of private airplanes", "airline" |
139
+ | product-car | "leased car", "your car", "car" |
140
+ | product-food | "StarBites", "food truck business", "ice cream" |
141
+ | product-game | "the 'Train to Nowhere ' game", "board game", "screen protector" |
142
+ | product-other | "new medicine", "acting software", "table" |
143
+ | product-ship | "research ship", "ship", "a ship" |
144
+ | product-software | "software", "instruction manual", "pizza ordering app" |
145
+ | product-train | "Universal Sonicator", "train", "the train" |
146
+ | product-weapon | "Flip Flops", "Sno Blaster", "SecurityFirst" |
147
+
148
+ ## Evaluation
149
+
150
+ ### Metrics
151
+ | Label | Precision | Recall | F1 |
152
+ |:-----------------------------------------|:----------|:-------|:-------|
153
+ | **all** | 0.6189 | 0.2850 | 0.3903 |
154
+ | art-broadcastprogram | 0.0 | 0.0 | 0.0 |
155
+ | art-film | 0.0 | 0.0 | 0.0 |
156
+ | art-music | 0.6667 | 0.2 | 0.3077 |
157
+ | art-other | 0.0 | 0.0 | 0.0 |
158
+ | art-painting | 0.0 | 0.0 | 0.0 |
159
+ | art-writtenart | 0.0 | 0.0 | 0.0 |
160
+ | building-airport | 0.7143 | 0.7692 | 0.7407 |
161
+ | building-hospital | 0.6667 | 0.7778 | 0.7179 |
162
+ | building-hotel | 0.7857 | 0.6875 | 0.7333 |
163
+ | building-library | 0.8182 | 0.75 | 0.7826 |
164
+ | building-other | 0.0 | 0.0 | 0.0 |
165
+ | building-restaurant | 0.8571 | 0.375 | 0.5217 |
166
+ | building-sportsfacility | 0.6667 | 0.5 | 0.5714 |
167
+ | building-theater | 0.9 | 0.5625 | 0.6923 |
168
+ | datetime-absolute | 0.3333 | 0.0769 | 0.125 |
169
+ | datetime-authored | 0.55 | 0.8462 | 0.6667 |
170
+ | datetime-range | 0.75 | 0.5 | 0.6 |
171
+ | datetime-relative | 0.0 | 0.0 | 0.0 |
172
+ | event-attack/battle/war/militaryconflict | 0.8 | 0.2857 | 0.4211 |
173
+ | event-disaster | 0.5385 | 0.5 | 0.5185 |
174
+ | event-election | 0.75 | 0.5 | 0.6 |
175
+ | event-other | 0.0 | 0.0 | 0.0 |
176
+ | event-protest | 0.5455 | 0.4615 | 0.5000 |
177
+ | event-sportsevent | 0.625 | 0.3846 | 0.4762 |
178
+ | location-bodiesofwater | 0.8333 | 0.3571 | 0.5 |
179
+ | location-gpe | 0.375 | 0.2143 | 0.2727 |
180
+ | location-island | 0.7143 | 0.3333 | 0.4545 |
181
+ | location-mountain | 0.5882 | 0.625 | 0.6061 |
182
+ | location-other | 0.0 | 0.0 | 0.0 |
183
+ | location-park | 0.6667 | 0.5 | 0.5714 |
184
+ | location-road/railway/highway/transit | 0.8 | 0.5333 | 0.64 |
185
+ | organization-company | 0.0 | 0.0 | 0.0 |
186
+ | organization-education | 0.3077 | 0.2857 | 0.2963 |
187
+ | organization-government/governmentagency | 0.25 | 0.0909 | 0.1333 |
188
+ | organization-media/newspaper | 0.5833 | 0.4667 | 0.5185 |
189
+ | organization-other | 1.0 | 0.0769 | 0.1429 |
190
+ | organization-politicalparty | 0.75 | 0.2727 | 0.4000 |
191
+ | organization-religion | 1.0 | 0.3077 | 0.4706 |
192
+ | organization-showorganization | 0.75 | 0.25 | 0.375 |
193
+ | organization-sportsleague | 0.8571 | 0.4286 | 0.5714 |
194
+ | organization-sportsteam | 0.4286 | 0.5 | 0.4615 |
195
+ | other-astronomything | 0.0 | 0.0 | 0.0 |
196
+ | other-award | 1.0 | 0.2143 | 0.3529 |
197
+ | other-biologything | 0.0 | 0.0 | 0.0 |
198
+ | other-chemicalthing | 0.4 | 0.3077 | 0.3478 |
199
+ | other-currency | 1.0 | 0.2143 | 0.3529 |
200
+ | other-disease | 0.5714 | 0.3077 | 0.4 |
201
+ | other-educationaldegree | 0.5833 | 0.5833 | 0.5833 |
202
+ | other-god | 0.8 | 0.2222 | 0.3478 |
203
+ | other-language | 0.8 | 0.2857 | 0.4211 |
204
+ | other-law | 0.6667 | 0.5 | 0.5714 |
205
+ | other-livingthing | 0.0 | 0.0 | 0.0 |
206
+ | other-medical | 0.0 | 0.0 | 0.0 |
207
+ | person-actor | 0.3448 | 0.5 | 0.4082 |
208
+ | person-artist/author | 0.6667 | 0.1429 | 0.2353 |
209
+ | person-athlete | 0.6667 | 0.2353 | 0.3478 |
210
+ | person-director | 0.2 | 0.0714 | 0.1053 |
211
+ | person-other | 0.0 | 0.0 | 0.0 |
212
+ | person-politician | 0.6667 | 0.0952 | 0.1667 |
213
+ | person-scholar | 0.4118 | 0.4667 | 0.4375 |
214
+ | person-soldier | 0.0 | 0.0 | 0.0 |
215
+ | product-airplane | 0.75 | 0.3333 | 0.4615 |
216
+ | product-car | 1.0 | 0.2143 | 0.3529 |
217
+ | product-food | 0.0 | 0.0 | 0.0 |
218
+ | product-game | 1.0 | 0.1333 | 0.2353 |
219
+ | product-other | 0.5 | 0.0909 | 0.1538 |
220
+ | product-ship | 0.75 | 0.3 | 0.4286 |
221
+ | product-software | 1.0 | 0.4167 | 0.5882 |
222
+ | product-train | 0.5556 | 0.3571 | 0.4348 |
223
+ | product-weapon | 0.3333 | 0.0625 | 0.1053 |
224
+
225
+ ## Uses
226
+
227
+ ### Direct Use for Inference
228
+
229
+ ```python
230
+ from span_marker import SpanMarkerModel
231
+
232
+ # Download from the 🤗 Hub
233
+ model = SpanMarkerModel.from_pretrained("YurtsAI/named_entity_recognition_document_context")
234
+ # Run inference
235
+ entities = model.predict("We have Kanye West, Beyoncé, and Taylor Swift performing at the beachside park on the island of Maui.")
236
+ ```
237
+
238
+ ### Downstream Use
239
+ You can finetune this model on your own dataset.
240
+
241
+ <details><summary>Click to expand</summary>
242
+
243
+ ```python
244
+ from span_marker import SpanMarkerModel, Trainer
245
+
246
+ # Download from the 🤗 Hub
247
+ model = SpanMarkerModel.from_pretrained("YurtsAI/named_entity_recognition_document_context")
248
+
249
+ # Specify a Dataset with "tokens" and "ner_tag" columns
250
+ dataset = load_dataset("conll2003") # For example CoNLL2003
251
+
252
+ # Initialize a Trainer using the pretrained model & dataset
253
+ trainer = Trainer(
254
+ model=model,
255
+ train_dataset=dataset["train"],
256
+ eval_dataset=dataset["validation"],
257
+ )
258
+ trainer.train()
259
+ trainer.save_model("YurtsAI/named_entity_recognition_document_context-finetuned")
260
+ ```
261
+ </details>
262
+
263
+ <!--
264
+ ### Out-of-Scope Use
265
+
266
+ *List how the model may foreseeably be misused and address what users ought not to do with the model.*
267
+ -->
268
+
269
+ <!--
270
+ ## Bias, Risks and Limitations
271
+
272
+ *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
273
+ -->
274
+
275
+ <!--
276
+ ### Recommendations
277
+
278
+ *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
279
+ -->
280
+
281
+ ## Training Details
282
+
283
+ ### Training Set Metrics
284
+ | Training set | Min | Median | Max |
285
+ |:----------------------|:----|:--------|:----|
286
+ | Sentence length | 1 | 18.4126 | 309 |
287
+ | Entities per sentence | 0 | 0.9794 | 5 |
288
+
289
+ ### Training Hyperparameters
290
+ - learning_rate: 1e-05
291
+ - train_batch_size: 4
292
+ - eval_batch_size: 4
293
+ - seed: 42
294
+ - gradient_accumulation_steps: 2
295
+ - total_train_batch_size: 8
296
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
297
+ - lr_scheduler_type: linear
298
+ - lr_scheduler_warmup_ratio: 0.1
299
+ - num_epochs: 3
300
+
301
+ ### Training Results
302
+ | Epoch | Step | Validation Loss | Validation Precision | Validation Recall | Validation F1 | Validation Accuracy |
303
+ |:------:|:----:|:---------------:|:--------------------:|:-----------------:|:-------------:|:-------------------:|
304
+ | 0.4322 | 500 | 0.0503 | 0.0 | 0.0 | 0.0 | 0.8898 |
305
+ | 0.8643 | 1000 | 0.0435 | 1.0 | 0.0010 | 0.0020 | 0.8900 |
306
+ | 1.2965 | 1500 | 0.0383 | 0.2841 | 0.0254 | 0.0466 | 0.8908 |
307
+ | 1.7286 | 2000 | 0.0326 | 0.5556 | 0.0710 | 0.1259 | 0.8951 |
308
+ | 2.1608 | 2500 | 0.0294 | 0.5806 | 0.1826 | 0.2778 | 0.9032 |
309
+ | 2.5929 | 3000 | 0.0278 | 0.6259 | 0.2698 | 0.3770 | 0.9109 |
310
+
311
+ ### Framework Versions
312
+ - Python: 3.12.2
313
+ - SpanMarker: 1.5.0
314
+ - Transformers: 4.41.2
315
+ - PyTorch: 2.3.1
316
+ - Datasets: 2.20.0
317
+ - Tokenizers: 0.19.1
318
+
319
+ ## Citation
320
+
321
+ ### BibTeX
322
+ ```
323
+ @software{Aarsen_SpanMarker,
324
+ author = {Aarsen, Tom},
325
+ license = {Apache-2.0},
326
+ title = {{SpanMarker for Named Entity Recognition}},
327
+ url = {https://github.com/tomaarsen/SpanMarkerNER}
328
+ }
329
+ ```
330
+
331
+ <!--
332
+ ## Glossary
333
+
334
+ *Clearly define terms in order to be accessible across audiences.*
335
+ -->
336
+
337
+ <!--
338
+ ## Model Card Authors
339
+
340
+ *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
341
+ -->
342
+
343
+ <!--
344
+ ## Model Card Contact
345
+
346
+ *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
347
+ -->
added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "<end>": 50266,
3
+ "<start>": 50265
4
+ }
config.json ADDED
@@ -0,0 +1,665 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "SpanMarkerModel"
4
+ ],
5
+ "encoder": {
6
+ "_name_or_path": "roberta-base",
7
+ "add_cross_attention": false,
8
+ "architectures": [
9
+ "RobertaForMaskedLM"
10
+ ],
11
+ "attention_probs_dropout_prob": 0.1,
12
+ "bad_words_ids": null,
13
+ "begin_suppress_tokens": null,
14
+ "bos_token_id": 0,
15
+ "chunk_size_feed_forward": 0,
16
+ "classifier_dropout": null,
17
+ "cross_attention_hidden_size": null,
18
+ "decoder_start_token_id": null,
19
+ "diversity_penalty": 0.0,
20
+ "do_sample": false,
21
+ "early_stopping": false,
22
+ "encoder_no_repeat_ngram_size": 0,
23
+ "eos_token_id": 2,
24
+ "exponential_decay_length_penalty": null,
25
+ "finetuning_task": null,
26
+ "forced_bos_token_id": null,
27
+ "forced_eos_token_id": null,
28
+ "hidden_act": "gelu",
29
+ "hidden_dropout_prob": 0.1,
30
+ "hidden_size": 768,
31
+ "id2label": {
32
+ "0": "O",
33
+ "1": "B-art-broadcastprogram",
34
+ "2": "I-art-broadcastprogram",
35
+ "3": "B-art-film",
36
+ "4": "I-art-film",
37
+ "5": "B-art-music",
38
+ "6": "I-art-music",
39
+ "7": "B-art-other",
40
+ "8": "I-art-other",
41
+ "9": "B-art-painting",
42
+ "10": "I-art-painting",
43
+ "11": "B-art-writtenart",
44
+ "12": "I-art-writtenart",
45
+ "13": "B-building-airport",
46
+ "14": "I-building-airport",
47
+ "15": "B-building-hospital",
48
+ "16": "I-building-hospital",
49
+ "17": "B-building-hotel",
50
+ "18": "I-building-hotel",
51
+ "19": "B-building-library",
52
+ "20": "I-building-library",
53
+ "21": "B-building-other",
54
+ "22": "I-building-other",
55
+ "23": "B-building-restaurant",
56
+ "24": "I-building-restaurant",
57
+ "25": "B-building-sportsfacility",
58
+ "26": "I-building-sportsfacility",
59
+ "27": "B-building-theater",
60
+ "28": "I-building-theater",
61
+ "29": "B-event-attack/battle/war/militaryconflict",
62
+ "30": "I-event-attack/battle/war/militaryconflict",
63
+ "31": "B-event-disaster",
64
+ "32": "I-event-disaster",
65
+ "33": "B-event-election",
66
+ "34": "I-event-election",
67
+ "35": "B-event-other",
68
+ "36": "I-event-other",
69
+ "37": "B-event-protest",
70
+ "38": "I-event-protest",
71
+ "39": "B-event-sportsevent",
72
+ "40": "I-event-sportsevent",
73
+ "41": "B-location-gpe",
74
+ "42": "I-location-gpe",
75
+ "43": "B-location-bodiesofwater",
76
+ "44": "I-location-bodiesofwater",
77
+ "45": "B-location-island",
78
+ "46": "I-location-island",
79
+ "47": "B-location-mountain",
80
+ "48": "I-location-mountain",
81
+ "49": "B-location-other",
82
+ "50": "I-location-other",
83
+ "51": "B-location-park",
84
+ "52": "I-location-park",
85
+ "53": "B-location-road/railway/highway/transit",
86
+ "54": "I-location-road/railway/highway/transit",
87
+ "55": "B-organization-company",
88
+ "56": "I-organization-company",
89
+ "57": "B-organization-education",
90
+ "58": "I-organization-education",
91
+ "59": "B-organization-government/governmentagency",
92
+ "60": "I-organization-government/governmentagency",
93
+ "61": "B-organization-media/newspaper",
94
+ "62": "I-organization-media/newspaper",
95
+ "63": "B-organization-other",
96
+ "64": "I-organization-other",
97
+ "65": "B-organization-politicalparty",
98
+ "66": "I-organization-politicalparty",
99
+ "67": "B-organization-religion",
100
+ "68": "I-organization-religion",
101
+ "69": "B-organization-showorganization",
102
+ "70": "I-organization-showorganization",
103
+ "71": "B-organization-sportsleague",
104
+ "72": "I-organization-sportsleague",
105
+ "73": "B-organization-sportsteam",
106
+ "74": "I-organization-sportsteam",
107
+ "75": "B-other-astronomything",
108
+ "76": "I-other-astronomything",
109
+ "77": "B-other-award",
110
+ "78": "I-other-award",
111
+ "79": "B-other-biologything",
112
+ "80": "I-other-biologything",
113
+ "81": "B-other-chemicalthing",
114
+ "82": "I-other-chemicalthing",
115
+ "83": "B-other-currency",
116
+ "84": "I-other-currency",
117
+ "85": "B-other-disease",
118
+ "86": "I-other-disease",
119
+ "87": "B-other-educationaldegree",
120
+ "88": "I-other-educationaldegree",
121
+ "89": "B-other-god",
122
+ "90": "I-other-god",
123
+ "91": "B-other-language",
124
+ "92": "I-other-language",
125
+ "93": "B-other-law",
126
+ "94": "I-other-law",
127
+ "95": "B-other-livingthing",
128
+ "96": "I-other-livingthing",
129
+ "97": "B-other-medical",
130
+ "98": "I-other-medical",
131
+ "99": "B-person-actor",
132
+ "100": "I-person-actor",
133
+ "101": "B-person-artist/author",
134
+ "102": "I-person-artist/author",
135
+ "103": "B-person-athlete",
136
+ "104": "I-person-athlete",
137
+ "105": "B-person-director",
138
+ "106": "I-person-director",
139
+ "107": "B-person-other",
140
+ "108": "I-person-other",
141
+ "109": "B-person-politician",
142
+ "110": "I-person-politician",
143
+ "111": "B-person-scholar",
144
+ "112": "I-person-scholar",
145
+ "113": "B-person-soldier",
146
+ "114": "I-person-soldier",
147
+ "115": "B-product-airplane",
148
+ "116": "I-product-airplane",
149
+ "117": "B-product-car",
150
+ "118": "I-product-car",
151
+ "119": "B-product-food",
152
+ "120": "I-product-food",
153
+ "121": "B-product-game",
154
+ "122": "I-product-game",
155
+ "123": "B-product-other",
156
+ "124": "I-product-other",
157
+ "125": "B-product-ship",
158
+ "126": "I-product-ship",
159
+ "127": "B-product-software",
160
+ "128": "I-product-software",
161
+ "129": "B-product-train",
162
+ "130": "I-product-train",
163
+ "131": "B-product-weapon",
164
+ "132": "I-product-weapon",
165
+ "133": "B-datetime-absolute",
166
+ "134": "I-datetime-absolute",
167
+ "135": "B-datetime-relative",
168
+ "136": "I-datetime-relative",
169
+ "137": "B-datetime-range",
170
+ "138": "I-datetime-range",
171
+ "139": "B-datetime-authored",
172
+ "140": "I-datetime-authored"
173
+ },
174
+ "initializer_range": 0.02,
175
+ "intermediate_size": 3072,
176
+ "is_decoder": false,
177
+ "is_encoder_decoder": false,
178
+ "label2id": {
179
+ "B-art-broadcastprogram": 1,
180
+ "B-art-film": 3,
181
+ "B-art-music": 5,
182
+ "B-art-other": 7,
183
+ "B-art-painting": 9,
184
+ "B-art-writtenart": 11,
185
+ "B-building-airport": 13,
186
+ "B-building-hospital": 15,
187
+ "B-building-hotel": 17,
188
+ "B-building-library": 19,
189
+ "B-building-other": 21,
190
+ "B-building-restaurant": 23,
191
+ "B-building-sportsfacility": 25,
192
+ "B-building-theater": 27,
193
+ "B-datetime-absolute": 133,
194
+ "B-datetime-authored": 139,
195
+ "B-datetime-range": 137,
196
+ "B-datetime-relative": 135,
197
+ "B-event-attack/battle/war/militaryconflict": 29,
198
+ "B-event-disaster": 31,
199
+ "B-event-election": 33,
200
+ "B-event-other": 35,
201
+ "B-event-protest": 37,
202
+ "B-event-sportsevent": 39,
203
+ "B-location-bodiesofwater": 43,
204
+ "B-location-gpe": 41,
205
+ "B-location-island": 45,
206
+ "B-location-mountain": 47,
207
+ "B-location-other": 49,
208
+ "B-location-park": 51,
209
+ "B-location-road/railway/highway/transit": 53,
210
+ "B-organization-company": 55,
211
+ "B-organization-education": 57,
212
+ "B-organization-government/governmentagency": 59,
213
+ "B-organization-media/newspaper": 61,
214
+ "B-organization-other": 63,
215
+ "B-organization-politicalparty": 65,
216
+ "B-organization-religion": 67,
217
+ "B-organization-showorganization": 69,
218
+ "B-organization-sportsleague": 71,
219
+ "B-organization-sportsteam": 73,
220
+ "B-other-astronomything": 75,
221
+ "B-other-award": 77,
222
+ "B-other-biologything": 79,
223
+ "B-other-chemicalthing": 81,
224
+ "B-other-currency": 83,
225
+ "B-other-disease": 85,
226
+ "B-other-educationaldegree": 87,
227
+ "B-other-god": 89,
228
+ "B-other-language": 91,
229
+ "B-other-law": 93,
230
+ "B-other-livingthing": 95,
231
+ "B-other-medical": 97,
232
+ "B-person-actor": 99,
233
+ "B-person-artist/author": 101,
234
+ "B-person-athlete": 103,
235
+ "B-person-director": 105,
236
+ "B-person-other": 107,
237
+ "B-person-politician": 109,
238
+ "B-person-scholar": 111,
239
+ "B-person-soldier": 113,
240
+ "B-product-airplane": 115,
241
+ "B-product-car": 117,
242
+ "B-product-food": 119,
243
+ "B-product-game": 121,
244
+ "B-product-other": 123,
245
+ "B-product-ship": 125,
246
+ "B-product-software": 127,
247
+ "B-product-train": 129,
248
+ "B-product-weapon": 131,
249
+ "I-art-broadcastprogram": 2,
250
+ "I-art-film": 4,
251
+ "I-art-music": 6,
252
+ "I-art-other": 8,
253
+ "I-art-painting": 10,
254
+ "I-art-writtenart": 12,
255
+ "I-building-airport": 14,
256
+ "I-building-hospital": 16,
257
+ "I-building-hotel": 18,
258
+ "I-building-library": 20,
259
+ "I-building-other": 22,
260
+ "I-building-restaurant": 24,
261
+ "I-building-sportsfacility": 26,
262
+ "I-building-theater": 28,
263
+ "I-datetime-absolute": 134,
264
+ "I-datetime-authored": 140,
265
+ "I-datetime-range": 138,
266
+ "I-datetime-relative": 136,
267
+ "I-event-attack/battle/war/militaryconflict": 30,
268
+ "I-event-disaster": 32,
269
+ "I-event-election": 34,
270
+ "I-event-other": 36,
271
+ "I-event-protest": 38,
272
+ "I-event-sportsevent": 40,
273
+ "I-location-bodiesofwater": 44,
274
+ "I-location-gpe": 42,
275
+ "I-location-island": 46,
276
+ "I-location-mountain": 48,
277
+ "I-location-other": 50,
278
+ "I-location-park": 52,
279
+ "I-location-road/railway/highway/transit": 54,
280
+ "I-organization-company": 56,
281
+ "I-organization-education": 58,
282
+ "I-organization-government/governmentagency": 60,
283
+ "I-organization-media/newspaper": 62,
284
+ "I-organization-other": 64,
285
+ "I-organization-politicalparty": 66,
286
+ "I-organization-religion": 68,
287
+ "I-organization-showorganization": 70,
288
+ "I-organization-sportsleague": 72,
289
+ "I-organization-sportsteam": 74,
290
+ "I-other-astronomything": 76,
291
+ "I-other-award": 78,
292
+ "I-other-biologything": 80,
293
+ "I-other-chemicalthing": 82,
294
+ "I-other-currency": 84,
295
+ "I-other-disease": 86,
296
+ "I-other-educationaldegree": 88,
297
+ "I-other-god": 90,
298
+ "I-other-language": 92,
299
+ "I-other-law": 94,
300
+ "I-other-livingthing": 96,
301
+ "I-other-medical": 98,
302
+ "I-person-actor": 100,
303
+ "I-person-artist/author": 102,
304
+ "I-person-athlete": 104,
305
+ "I-person-director": 106,
306
+ "I-person-other": 108,
307
+ "I-person-politician": 110,
308
+ "I-person-scholar": 112,
309
+ "I-person-soldier": 114,
310
+ "I-product-airplane": 116,
311
+ "I-product-car": 118,
312
+ "I-product-food": 120,
313
+ "I-product-game": 122,
314
+ "I-product-other": 124,
315
+ "I-product-ship": 126,
316
+ "I-product-software": 128,
317
+ "I-product-train": 130,
318
+ "I-product-weapon": 132,
319
+ "O": 0
320
+ },
321
+ "layer_norm_eps": 1e-05,
322
+ "length_penalty": 1.0,
323
+ "max_length": 20,
324
+ "max_position_embeddings": 514,
325
+ "min_length": 0,
326
+ "model_type": "roberta",
327
+ "no_repeat_ngram_size": 0,
328
+ "num_attention_heads": 12,
329
+ "num_beam_groups": 1,
330
+ "num_beams": 1,
331
+ "num_hidden_layers": 12,
332
+ "num_return_sequences": 1,
333
+ "output_attentions": false,
334
+ "output_hidden_states": false,
335
+ "output_scores": false,
336
+ "pad_token_id": 1,
337
+ "position_embedding_type": "absolute",
338
+ "prefix": null,
339
+ "problem_type": null,
340
+ "pruned_heads": {},
341
+ "remove_invalid_values": false,
342
+ "repetition_penalty": 1.0,
343
+ "return_dict": true,
344
+ "return_dict_in_generate": false,
345
+ "sep_token_id": null,
346
+ "suppress_tokens": null,
347
+ "task_specific_params": null,
348
+ "temperature": 1.0,
349
+ "tf_legacy_loss": false,
350
+ "tie_encoder_decoder": false,
351
+ "tie_word_embeddings": true,
352
+ "tokenizer_class": null,
353
+ "top_k": 50,
354
+ "top_p": 1.0,
355
+ "torch_dtype": null,
356
+ "torchscript": false,
357
+ "transformers_version": "4.41.2",
358
+ "type_vocab_size": 1,
359
+ "typical_p": 1.0,
360
+ "use_bfloat16": false,
361
+ "use_cache": true,
362
+ "vocab_size": 50272
363
+ },
364
+ "entity_max_length": 11,
365
+ "id2label": {
366
+ "0": "O",
367
+ "1": "art-broadcastprogram",
368
+ "2": "art-film",
369
+ "3": "art-music",
370
+ "4": "art-other",
371
+ "5": "art-painting",
372
+ "6": "art-writtenart",
373
+ "7": "building-airport",
374
+ "8": "building-hospital",
375
+ "9": "building-hotel",
376
+ "10": "building-library",
377
+ "11": "building-other",
378
+ "12": "building-restaurant",
379
+ "13": "building-sportsfacility",
380
+ "14": "building-theater",
381
+ "15": "datetime-absolute",
382
+ "16": "datetime-authored",
383
+ "17": "datetime-range",
384
+ "18": "datetime-relative",
385
+ "19": "event-attack/battle/war/militaryconflict",
386
+ "20": "event-disaster",
387
+ "21": "event-election",
388
+ "22": "event-other",
389
+ "23": "event-protest",
390
+ "24": "event-sportsevent",
391
+ "25": "location-bodiesofwater",
392
+ "26": "location-gpe",
393
+ "27": "location-island",
394
+ "28": "location-mountain",
395
+ "29": "location-other",
396
+ "30": "location-park",
397
+ "31": "location-road/railway/highway/transit",
398
+ "32": "organization-company",
399
+ "33": "organization-education",
400
+ "34": "organization-government/governmentagency",
401
+ "35": "organization-media/newspaper",
402
+ "36": "organization-other",
403
+ "37": "organization-politicalparty",
404
+ "38": "organization-religion",
405
+ "39": "organization-showorganization",
406
+ "40": "organization-sportsleague",
407
+ "41": "organization-sportsteam",
408
+ "42": "other-astronomything",
409
+ "43": "other-award",
410
+ "44": "other-biologything",
411
+ "45": "other-chemicalthing",
412
+ "46": "other-currency",
413
+ "47": "other-disease",
414
+ "48": "other-educationaldegree",
415
+ "49": "other-god",
416
+ "50": "other-language",
417
+ "51": "other-law",
418
+ "52": "other-livingthing",
419
+ "53": "other-medical",
420
+ "54": "person-actor",
421
+ "55": "person-artist/author",
422
+ "56": "person-athlete",
423
+ "57": "person-director",
424
+ "58": "person-other",
425
+ "59": "person-politician",
426
+ "60": "person-scholar",
427
+ "61": "person-soldier",
428
+ "62": "product-airplane",
429
+ "63": "product-car",
430
+ "64": "product-food",
431
+ "65": "product-game",
432
+ "66": "product-other",
433
+ "67": "product-ship",
434
+ "68": "product-software",
435
+ "69": "product-train",
436
+ "70": "product-weapon"
437
+ },
438
+ "id2reduced_id": {
439
+ "0": 0,
440
+ "1": 1,
441
+ "2": 1,
442
+ "3": 2,
443
+ "4": 2,
444
+ "5": 3,
445
+ "6": 3,
446
+ "7": 4,
447
+ "8": 4,
448
+ "9": 5,
449
+ "10": 5,
450
+ "11": 6,
451
+ "12": 6,
452
+ "13": 7,
453
+ "14": 7,
454
+ "15": 8,
455
+ "16": 8,
456
+ "17": 9,
457
+ "18": 9,
458
+ "19": 10,
459
+ "20": 10,
460
+ "21": 11,
461
+ "22": 11,
462
+ "23": 12,
463
+ "24": 12,
464
+ "25": 13,
465
+ "26": 13,
466
+ "27": 14,
467
+ "28": 14,
468
+ "29": 19,
469
+ "30": 19,
470
+ "31": 20,
471
+ "32": 20,
472
+ "33": 21,
473
+ "34": 21,
474
+ "35": 22,
475
+ "36": 22,
476
+ "37": 23,
477
+ "38": 23,
478
+ "39": 24,
479
+ "40": 24,
480
+ "41": 26,
481
+ "42": 26,
482
+ "43": 25,
483
+ "44": 25,
484
+ "45": 27,
485
+ "46": 27,
486
+ "47": 28,
487
+ "48": 28,
488
+ "49": 29,
489
+ "50": 29,
490
+ "51": 30,
491
+ "52": 30,
492
+ "53": 31,
493
+ "54": 31,
494
+ "55": 32,
495
+ "56": 32,
496
+ "57": 33,
497
+ "58": 33,
498
+ "59": 34,
499
+ "60": 34,
500
+ "61": 35,
501
+ "62": 35,
502
+ "63": 36,
503
+ "64": 36,
504
+ "65": 37,
505
+ "66": 37,
506
+ "67": 38,
507
+ "68": 38,
508
+ "69": 39,
509
+ "70": 39,
510
+ "71": 40,
511
+ "72": 40,
512
+ "73": 41,
513
+ "74": 41,
514
+ "75": 42,
515
+ "76": 42,
516
+ "77": 43,
517
+ "78": 43,
518
+ "79": 44,
519
+ "80": 44,
520
+ "81": 45,
521
+ "82": 45,
522
+ "83": 46,
523
+ "84": 46,
524
+ "85": 47,
525
+ "86": 47,
526
+ "87": 48,
527
+ "88": 48,
528
+ "89": 49,
529
+ "90": 49,
530
+ "91": 50,
531
+ "92": 50,
532
+ "93": 51,
533
+ "94": 51,
534
+ "95": 52,
535
+ "96": 52,
536
+ "97": 53,
537
+ "98": 53,
538
+ "99": 54,
539
+ "100": 54,
540
+ "101": 55,
541
+ "102": 55,
542
+ "103": 56,
543
+ "104": 56,
544
+ "105": 57,
545
+ "106": 57,
546
+ "107": 58,
547
+ "108": 58,
548
+ "109": 59,
549
+ "110": 59,
550
+ "111": 60,
551
+ "112": 60,
552
+ "113": 61,
553
+ "114": 61,
554
+ "115": 62,
555
+ "116": 62,
556
+ "117": 63,
557
+ "118": 63,
558
+ "119": 64,
559
+ "120": 64,
560
+ "121": 65,
561
+ "122": 65,
562
+ "123": 66,
563
+ "124": 66,
564
+ "125": 67,
565
+ "126": 67,
566
+ "127": 68,
567
+ "128": 68,
568
+ "129": 69,
569
+ "130": 69,
570
+ "131": 70,
571
+ "132": 70,
572
+ "133": 15,
573
+ "134": 15,
574
+ "135": 18,
575
+ "136": 18,
576
+ "137": 17,
577
+ "138": 17,
578
+ "139": 16,
579
+ "140": 16
580
+ },
581
+ "label2id": {
582
+ "O": 0,
583
+ "art-broadcastprogram": 1,
584
+ "art-film": 2,
585
+ "art-music": 3,
586
+ "art-other": 4,
587
+ "art-painting": 5,
588
+ "art-writtenart": 6,
589
+ "building-airport": 7,
590
+ "building-hospital": 8,
591
+ "building-hotel": 9,
592
+ "building-library": 10,
593
+ "building-other": 11,
594
+ "building-restaurant": 12,
595
+ "building-sportsfacility": 13,
596
+ "building-theater": 14,
597
+ "datetime-absolute": 15,
598
+ "datetime-authored": 16,
599
+ "datetime-range": 17,
600
+ "datetime-relative": 18,
601
+ "event-attack/battle/war/militaryconflict": 19,
602
+ "event-disaster": 20,
603
+ "event-election": 21,
604
+ "event-other": 22,
605
+ "event-protest": 23,
606
+ "event-sportsevent": 24,
607
+ "location-bodiesofwater": 25,
608
+ "location-gpe": 26,
609
+ "location-island": 27,
610
+ "location-mountain": 28,
611
+ "location-other": 29,
612
+ "location-park": 30,
613
+ "location-road/railway/highway/transit": 31,
614
+ "organization-company": 32,
615
+ "organization-education": 33,
616
+ "organization-government/governmentagency": 34,
617
+ "organization-media/newspaper": 35,
618
+ "organization-other": 36,
619
+ "organization-politicalparty": 37,
620
+ "organization-religion": 38,
621
+ "organization-showorganization": 39,
622
+ "organization-sportsleague": 40,
623
+ "organization-sportsteam": 41,
624
+ "other-astronomything": 42,
625
+ "other-award": 43,
626
+ "other-biologything": 44,
627
+ "other-chemicalthing": 45,
628
+ "other-currency": 46,
629
+ "other-disease": 47,
630
+ "other-educationaldegree": 48,
631
+ "other-god": 49,
632
+ "other-language": 50,
633
+ "other-law": 51,
634
+ "other-livingthing": 52,
635
+ "other-medical": 53,
636
+ "person-actor": 54,
637
+ "person-artist/author": 55,
638
+ "person-athlete": 56,
639
+ "person-director": 57,
640
+ "person-other": 58,
641
+ "person-politician": 59,
642
+ "person-scholar": 60,
643
+ "person-soldier": 61,
644
+ "product-airplane": 62,
645
+ "product-car": 63,
646
+ "product-food": 64,
647
+ "product-game": 65,
648
+ "product-other": 66,
649
+ "product-ship": 67,
650
+ "product-software": 68,
651
+ "product-train": 69,
652
+ "product-weapon": 70
653
+ },
654
+ "marker_max_length": 128,
655
+ "max_next_context": 2,
656
+ "max_prev_context": 2,
657
+ "model_max_length": 256,
658
+ "model_max_length_default": 512,
659
+ "model_type": "span-marker",
660
+ "span_marker_version": "1.5.0",
661
+ "torch_dtype": "float32",
662
+ "trained_with_document_context": true,
663
+ "transformers_version": "4.41.2",
664
+ "vocab_size": 50272
665
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64657100004bf75b1955431a51759d761a548206d96bf0f66391e1c0b392c115
3
+ size 499064676
special_tokens_map.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<s>",
3
+ "cls_token": "<s>",
4
+ "eos_token": "</s>",
5
+ "mask_token": {
6
+ "content": "<mask>",
7
+ "lstrip": true,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "pad_token": "<pad>",
13
+ "sep_token": "</s>",
14
+ "unk_token": "<unk>"
15
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,76 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": true,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "<s>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "<pad>",
14
+ "lstrip": false,
15
+ "normalized": true,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "2": {
21
+ "content": "</s>",
22
+ "lstrip": false,
23
+ "normalized": true,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "3": {
29
+ "content": "<unk>",
30
+ "lstrip": false,
31
+ "normalized": true,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": true
35
+ },
36
+ "50264": {
37
+ "content": "<mask>",
38
+ "lstrip": true,
39
+ "normalized": false,
40
+ "rstrip": false,
41
+ "single_word": false,
42
+ "special": true
43
+ },
44
+ "50265": {
45
+ "content": "<start>",
46
+ "lstrip": false,
47
+ "normalized": false,
48
+ "rstrip": false,
49
+ "single_word": false,
50
+ "special": true
51
+ },
52
+ "50266": {
53
+ "content": "<end>",
54
+ "lstrip": false,
55
+ "normalized": false,
56
+ "rstrip": false,
57
+ "single_word": false,
58
+ "special": true
59
+ }
60
+ },
61
+ "bos_token": "<s>",
62
+ "clean_up_tokenization_spaces": true,
63
+ "cls_token": "<s>",
64
+ "entity_max_length": 11,
65
+ "eos_token": "</s>",
66
+ "errors": "replace",
67
+ "mask_token": "<mask>",
68
+ "max_next_context": 2,
69
+ "max_prev_context": 2,
70
+ "model_max_length": 256,
71
+ "pad_token": "<pad>",
72
+ "sep_token": "</s>",
73
+ "tokenizer_class": "RobertaTokenizer",
74
+ "trim_offsets": true,
75
+ "unk_token": "<unk>"
76
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:790460a18d0c9bd7a3a4a8a367f39f52e7bb83e0c39f05f47bc633ca151a4f82
3
+ size 5112
vocab.json ADDED
The diff for this file is too large to render. See raw diff