anismahmahi commited on
Commit
3663394
1 Parent(s): 1e88da6

Add SetFit model

Browse files
1_Pooling/config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 768,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": true,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false
7
+ }
README.md ADDED
@@ -0,0 +1,307 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: setfit
3
+ tags:
4
+ - setfit
5
+ - sentence-transformers
6
+ - text-classification
7
+ - generated_from_setfit_trainer
8
+ metrics:
9
+ - f1
10
+ widget:
11
+ - text: The DOJ surrendered in the matter and accepted the fact that the plans are
12
+ protected under the First Amendment.
13
+ - text: The settlement was approved by a federal judge.
14
+ - text: The project in question is under the banner of the Vox-Pol Network of Excellence,
15
+ which “is designed to comprehensively research, analyse, debate, and critique
16
+ issues surrounding violent online political extremism (VOPE).”
17
+ - text: And the more substantive those debates become, the further away the US discourse
18
+ will move from the mendacious assumptions of Obama’s Iran policy – that the Iranian
19
+ regime is a responsible actor and potential US ally, and that there is nothing
20
+ inherently aggressive or problematic about Iran’s illicit nuclear weapons program.
21
+ - text: Of course, the UN has reached no such conclusion.
22
+ pipeline_tag: text-classification
23
+ inference: true
24
+ model-index:
25
+ - name: SetFit
26
+ results:
27
+ - task:
28
+ type: text-classification
29
+ name: Text Classification
30
+ dataset:
31
+ name: Unknown
32
+ type: unknown
33
+ split: test
34
+ metrics:
35
+ - type: f1
36
+ value: 0.4317180616740088
37
+ name: F1
38
+ ---
39
+
40
+ # SetFit
41
+
42
+ This is a [SetFit](https://github.com/huggingface/setfit) model that can be used for Text Classification. A [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance is used for classification.
43
+
44
+ The model has been trained using an efficient few-shot learning technique that involves:
45
+
46
+ 1. Fine-tuning a [Sentence Transformer](https://www.sbert.net) with contrastive learning.
47
+ 2. Training a classification head with features from the fine-tuned Sentence Transformer.
48
+
49
+ ## Model Details
50
+
51
+ ### Model Description
52
+ - **Model Type:** SetFit
53
+ <!-- - **Sentence Transformer:** [Unknown](https://huggingface.co/unknown) -->
54
+ - **Classification head:** a [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance
55
+ - **Maximum Sequence Length:** 512 tokens
56
+ - **Number of Classes:** 2 classes
57
+ <!-- - **Training Dataset:** [Unknown](https://huggingface.co/datasets/unknown) -->
58
+ <!-- - **Language:** Unknown -->
59
+ <!-- - **License:** Unknown -->
60
+
61
+ ### Model Sources
62
+
63
+ - **Repository:** [SetFit on GitHub](https://github.com/huggingface/setfit)
64
+ - **Paper:** [Efficient Few-Shot Learning Without Prompts](https://arxiv.org/abs/2209.11055)
65
+ - **Blogpost:** [SetFit: Efficient Few-Shot Learning Without Prompts](https://huggingface.co/blog/setfit)
66
+
67
+ ### Model Labels
68
+ | Label | Examples |
69
+ |:------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
70
+ | 0.0 | <ul><li>'Pamela Geller and Robert Spencer co-founded anti-Muslim group Stop Islamization of America.\n'</li><li>'He added: "We condemn all those whose behaviours and views run counter to our shared values and will not stand for extremism in any form."\n'</li><li>'Ms Geller, of the Atlas Shrugs blog, and Mr Spencer, of Jihad Watch, are also co-founders of the American Freedom Defense Initiative, best known for a pro-Israel "Defeat Jihad" poster campaign on the New York subway.\n'</li></ul> |
71
+ | 1.0 | <ul><li>'On both of their blogs the pair called their bans from entering the UK "a striking blow against freedom" and said the "the nation that gave the world the Magna Carta is dead".\n'</li><li>'A researcher with the organisation, Matthew Collins, said it was "delighted" with the decision.\n'</li><li>'Lead attorney Matt Gonzalez has argued that the weapon was a SIG Sauer with a "hair trigger in single-action mode" — a model well-known for accidental discharges even among experienced shooters.\n'</li></ul> |
72
+
73
+ ## Evaluation
74
+
75
+ ### Metrics
76
+ | Label | F1 |
77
+ |:--------|:-------|
78
+ | **all** | 0.4317 |
79
+
80
+ ## Uses
81
+
82
+ ### Direct Use for Inference
83
+
84
+ First install the SetFit library:
85
+
86
+ ```bash
87
+ pip install setfit
88
+ ```
89
+
90
+ Then you can load this model and run inference.
91
+
92
+ ```python
93
+ from setfit import SetFitModel
94
+
95
+ # Download from the 🤗 Hub
96
+ model = SetFitModel.from_pretrained("anismahmahi/improve-G3-setfit-model")
97
+ # Run inference
98
+ preds = model("The settlement was approved by a federal judge.")
99
+ ```
100
+
101
+ <!--
102
+ ### Downstream Use
103
+
104
+ *List how someone could finetune this model on their own dataset.*
105
+ -->
106
+
107
+ <!--
108
+ ### Out-of-Scope Use
109
+
110
+ *List how the model may foreseeably be misused and address what users ought not to do with the model.*
111
+ -->
112
+
113
+ <!--
114
+ ## Bias, Risks and Limitations
115
+
116
+ *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
117
+ -->
118
+
119
+ <!--
120
+ ### Recommendations
121
+
122
+ *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
123
+ -->
124
+
125
+ ## Training Details
126
+
127
+ ### Training Set Metrics
128
+ | Training set | Min | Median | Max |
129
+ |:-------------|:----|:--------|:----|
130
+ | Word count | 1 | 26.2226 | 129 |
131
+
132
+ | Label | Training Sample Count |
133
+ |:------|:----------------------|
134
+ | 0 | 2362 |
135
+ | 1 | 1784 |
136
+
137
+ ### Training Hyperparameters
138
+ - batch_size: (16, 16)
139
+ - num_epochs: (2, 2)
140
+ - max_steps: -1
141
+ - sampling_strategy: oversampling
142
+ - num_iterations: 5
143
+ - body_learning_rate: (2e-05, 1e-05)
144
+ - head_learning_rate: 0.01
145
+ - loss: CosineSimilarityLoss
146
+ - distance_metric: cosine_distance
147
+ - margin: 0.25
148
+ - end_to_end: False
149
+ - use_amp: False
150
+ - warmup_proportion: 0.1
151
+ - seed: 42
152
+ - eval_max_steps: -1
153
+ - load_best_model_at_end: True
154
+
155
+ ### Training Results
156
+ | Epoch | Step | Training Loss | Validation Loss |
157
+ |:-------:|:--------:|:-------------:|:---------------:|
158
+ | 0.0004 | 1 | 0.3949 | - |
159
+ | 0.0193 | 50 | 0.2806 | - |
160
+ | 0.0386 | 100 | 0.2461 | - |
161
+ | 0.0579 | 150 | 0.2522 | - |
162
+ | 0.0772 | 200 | 0.279 | - |
163
+ | 0.0965 | 250 | 0.2149 | - |
164
+ | 0.1157 | 300 | 0.2513 | - |
165
+ | 0.1350 | 350 | 0.2426 | - |
166
+ | 0.1543 | 400 | 0.2696 | - |
167
+ | 0.1736 | 450 | 0.2485 | - |
168
+ | 0.1929 | 500 | 0.2209 | - |
169
+ | 0.2122 | 550 | 0.2412 | - |
170
+ | 0.2315 | 600 | 0.1801 | - |
171
+ | 0.2508 | 650 | 0.197 | - |
172
+ | 0.2701 | 700 | 0.2223 | - |
173
+ | 0.2894 | 750 | 0.1825 | - |
174
+ | 0.3086 | 800 | 0.2067 | - |
175
+ | 0.3279 | 850 | 0.1726 | - |
176
+ | 0.3472 | 900 | 0.2091 | - |
177
+ | 0.3665 | 950 | 0.2159 | - |
178
+ | 0.3858 | 1000 | 0.2433 | - |
179
+ | 0.4051 | 1050 | 0.1102 | - |
180
+ | 0.4244 | 1100 | 0.081 | - |
181
+ | 0.4437 | 1150 | 0.1661 | - |
182
+ | 0.4630 | 1200 | 0.1574 | - |
183
+ | 0.4823 | 1250 | 0.1458 | - |
184
+ | 0.5015 | 1300 | 0.0881 | - |
185
+ | 0.5208 | 1350 | 0.0683 | - |
186
+ | 0.5401 | 1400 | 0.2053 | - |
187
+ | 0.5594 | 1450 | 0.0581 | - |
188
+ | 0.5787 | 1500 | 0.0742 | - |
189
+ | 0.5980 | 1550 | 0.1775 | - |
190
+ | 0.6173 | 1600 | 0.0541 | - |
191
+ | 0.6366 | 1650 | 0.1086 | - |
192
+ | 0.6559 | 1700 | 0.0654 | - |
193
+ | 0.6752 | 1750 | 0.0909 | - |
194
+ | 0.6944 | 1800 | 0.0571 | - |
195
+ | 0.7137 | 1850 | 0.0016 | - |
196
+ | 0.7330 | 1900 | 0.0963 | - |
197
+ | 0.7523 | 1950 | 0.0063 | - |
198
+ | 0.7716 | 2000 | 0.0011 | - |
199
+ | 0.7909 | 2050 | 0.0033 | - |
200
+ | 0.8102 | 2100 | 0.0069 | - |
201
+ | 0.8295 | 2150 | 0.0013 | - |
202
+ | 0.8488 | 2200 | 0.0051 | - |
203
+ | 0.8681 | 2250 | 0.0596 | - |
204
+ | 0.8873 | 2300 | 0.0007 | - |
205
+ | 0.9066 | 2350 | 0.0122 | - |
206
+ | 0.9259 | 2400 | 0.0012 | - |
207
+ | 0.9452 | 2450 | 0.0003 | - |
208
+ | 0.9645 | 2500 | 0.0012 | - |
209
+ | 0.9838 | 2550 | 0.002 | - |
210
+ | **1.0** | **2592** | **-** | **0.2706** |
211
+ | 1.0031 | 2600 | 0.001 | - |
212
+ | 1.0224 | 2650 | 0.0015 | - |
213
+ | 1.0417 | 2700 | 0.0594 | - |
214
+ | 1.0610 | 2750 | 0.0011 | - |
215
+ | 1.0802 | 2800 | 0.0087 | - |
216
+ | 1.0995 | 2850 | 0.0608 | - |
217
+ | 1.1188 | 2900 | 0.0531 | - |
218
+ | 1.1381 | 2950 | 0.0006 | - |
219
+ | 1.1574 | 3000 | 0.001 | - |
220
+ | 1.1767 | 3050 | 0.06 | - |
221
+ | 1.1960 | 3100 | 0.0003 | - |
222
+ | 1.2153 | 3150 | 0.0004 | - |
223
+ | 1.2346 | 3200 | 0.0002 | - |
224
+ | 1.2539 | 3250 | 0.0007 | - |
225
+ | 1.2731 | 3300 | 0.0006 | - |
226
+ | 1.2924 | 3350 | 0.0005 | - |
227
+ | 1.3117 | 3400 | 0.0007 | - |
228
+ | 1.3310 | 3450 | 0.0001 | - |
229
+ | 1.3503 | 3500 | 0.0587 | - |
230
+ | 1.3696 | 3550 | 0.0002 | - |
231
+ | 1.3889 | 3600 | 0.0001 | - |
232
+ | 1.4082 | 3650 | 0.0003 | - |
233
+ | 1.4275 | 3700 | 0.0002 | - |
234
+ | 1.4468 | 3750 | 0.0011 | - |
235
+ | 1.4660 | 3800 | 0.0007 | - |
236
+ | 1.4853 | 3850 | 0.0001 | - |
237
+ | 1.5046 | 3900 | 0.0001 | - |
238
+ | 1.5239 | 3950 | 0.0002 | - |
239
+ | 1.5432 | 4000 | 0.0001 | - |
240
+ | 1.5625 | 4050 | 0.0003 | - |
241
+ | 1.5818 | 4100 | 0.0002 | - |
242
+ | 1.6011 | 4150 | 0.0001 | - |
243
+ | 1.6204 | 4200 | 0.0002 | - |
244
+ | 1.6397 | 4250 | 0.0002 | - |
245
+ | 1.6590 | 4300 | 0.0003 | - |
246
+ | 1.6782 | 4350 | 0.0003 | - |
247
+ | 1.6975 | 4400 | 0.0002 | - |
248
+ | 1.7168 | 4450 | 0.0001 | - |
249
+ | 1.7361 | 4500 | 0.0037 | - |
250
+ | 1.7554 | 4550 | 0.0002 | - |
251
+ | 1.7747 | 4600 | 0.0001 | - |
252
+ | 1.7940 | 4650 | 0.0001 | - |
253
+ | 1.8133 | 4700 | 0.0001 | - |
254
+ | 1.8326 | 4750 | 0.0001 | - |
255
+ | 1.8519 | 4800 | 0.0003 | - |
256
+ | 1.8711 | 4850 | 0.0002 | - |
257
+ | 1.8904 | 4900 | 0.0001 | - |
258
+ | 1.9097 | 4950 | 0.0004 | - |
259
+ | 1.9290 | 5000 | 0.0001 | - |
260
+ | 1.9483 | 5050 | 0.0001 | - |
261
+ | 1.9676 | 5100 | 0.0001 | - |
262
+ | 1.9869 | 5150 | 0.0004 | - |
263
+ | 2.0 | 5184 | - | 0.2802 |
264
+
265
+ * The bold row denotes the saved checkpoint.
266
+ ### Framework Versions
267
+ - Python: 3.10.12
268
+ - SetFit: 1.0.2
269
+ - Sentence Transformers: 2.2.2
270
+ - Transformers: 4.35.2
271
+ - PyTorch: 2.1.0+cu121
272
+ - Datasets: 2.16.1
273
+ - Tokenizers: 0.15.0
274
+
275
+ ## Citation
276
+
277
+ ### BibTeX
278
+ ```bibtex
279
+ @article{https://doi.org/10.48550/arxiv.2209.11055,
280
+ doi = {10.48550/ARXIV.2209.11055},
281
+ url = {https://arxiv.org/abs/2209.11055},
282
+ author = {Tunstall, Lewis and Reimers, Nils and Jo, Unso Eun Seo and Bates, Luke and Korat, Daniel and Wasserblat, Moshe and Pereg, Oren},
283
+ keywords = {Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences},
284
+ title = {Efficient Few-Shot Learning Without Prompts},
285
+ publisher = {arXiv},
286
+ year = {2022},
287
+ copyright = {Creative Commons Attribution 4.0 International}
288
+ }
289
+ ```
290
+
291
+ <!--
292
+ ## Glossary
293
+
294
+ *Clearly define terms in order to be accessible across audiences.*
295
+ -->
296
+
297
+ <!--
298
+ ## Model Card Authors
299
+
300
+ *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
301
+ -->
302
+
303
+ <!--
304
+ ## Model Card Contact
305
+
306
+ *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
307
+ -->
config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "checkpoints/step_2592/",
3
+ "architectures": [
4
+ "MPNetModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "eos_token_id": 2,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-05,
15
+ "max_position_embeddings": 514,
16
+ "model_type": "mpnet",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 1,
20
+ "relative_attention_num_buckets": 32,
21
+ "torch_dtype": "float32",
22
+ "transformers_version": "4.35.2",
23
+ "vocab_size": 30527
24
+ }
config_sentence_transformers.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "2.0.0",
4
+ "transformers": "4.7.0",
5
+ "pytorch": "1.9.0+cu102"
6
+ }
7
+ }
config_setfit.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "labels": [
3
+ 0,
4
+ 1
5
+ ],
6
+ "normalize_embeddings": false
7
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:198837eddacbaeca0e2c946bbddd9b60969ce8253c3792a11d253a4afe10bc9a
3
+ size 437967672
model_head.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:04e49725536ff8bdd8293ca370ebe9c19b6a89f3c9e3e1be1c395e3551c306a9
3
+ size 6975
modules.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ }
14
+ ]
sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 512,
3
+ "do_lower_case": false
4
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "cls_token": {
10
+ "content": "<s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "eos_token": {
17
+ "content": "</s>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "mask_token": {
24
+ "content": "<mask>",
25
+ "lstrip": true,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "pad_token": {
31
+ "content": "<pad>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ },
37
+ "sep_token": {
38
+ "content": "</s>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false
43
+ },
44
+ "unk_token": {
45
+ "content": "[UNK]",
46
+ "lstrip": false,
47
+ "normalized": false,
48
+ "rstrip": false,
49
+ "single_word": false
50
+ }
51
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<s>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<pad>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "</s>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "104": {
28
+ "content": "[UNK]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "30526": {
36
+ "content": "<mask>",
37
+ "lstrip": true,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "bos_token": "<s>",
45
+ "clean_up_tokenization_spaces": true,
46
+ "cls_token": "<s>",
47
+ "do_basic_tokenize": true,
48
+ "do_lower_case": true,
49
+ "eos_token": "</s>",
50
+ "mask_token": "<mask>",
51
+ "max_length": 512,
52
+ "model_max_length": 512,
53
+ "never_split": null,
54
+ "pad_to_multiple_of": null,
55
+ "pad_token": "<pad>",
56
+ "pad_token_type_id": 0,
57
+ "padding_side": "right",
58
+ "sep_token": "</s>",
59
+ "stride": 0,
60
+ "strip_accents": null,
61
+ "tokenize_chinese_chars": true,
62
+ "tokenizer_class": "MPNetTokenizer",
63
+ "truncation_side": "right",
64
+ "truncation_strategy": "longest_first",
65
+ "unk_token": "[UNK]"
66
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff