dendimaki commited on
Commit
59baf27
1 Parent(s): 7e50149

Add SetFit model

Browse files
1_Pooling/config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 768,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": true,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false,
7
+ "pooling_mode_weightedmean_tokens": false,
8
+ "pooling_mode_lasttoken": false,
9
+ "include_prompt": true
10
+ }
README.md ADDED
@@ -0,0 +1,258 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: setfit
3
+ tags:
4
+ - setfit
5
+ - sentence-transformers
6
+ - text-classification
7
+ - generated_from_setfit_trainer
8
+ base_model: sentence-transformers/paraphrase-mpnet-base-v2
9
+ datasets:
10
+ - dendimaki/v1
11
+ metrics:
12
+ - accuracy
13
+ widget:
14
+ - text: so you know you said that layer three maybe sounded interesting
15
+ - text: just this like sense of energy thats aliveness and aliveness tingly aliveness
16
+ - text: id say is pretty or really the dominant state unless i really focus on location
17
+ one and even then
18
+ - text: pervading presence
19
+ - text: nonduality for you
20
+ pipeline_tag: text-classification
21
+ inference: true
22
+ model-index:
23
+ - name: SetFit with sentence-transformers/paraphrase-mpnet-base-v2
24
+ results:
25
+ - task:
26
+ type: text-classification
27
+ name: Text Classification
28
+ dataset:
29
+ name: dendimaki/v1
30
+ type: dendimaki/v1
31
+ split: test
32
+ metrics:
33
+ - type: accuracy
34
+ value: 0.46352941176470586
35
+ name: Accuracy
36
+ ---
37
+
38
+ # SetFit with sentence-transformers/paraphrase-mpnet-base-v2
39
+
40
+ This is a [SetFit](https://github.com/huggingface/setfit) model trained on the [dendimaki/v1](https://huggingface.co/datasets/dendimaki/v1) dataset that can be used for Text Classification. This SetFit model uses [sentence-transformers/paraphrase-mpnet-base-v2](https://huggingface.co/sentence-transformers/paraphrase-mpnet-base-v2) as the Sentence Transformer embedding model. A [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance is used for classification.
41
+
42
+ The model has been trained using an efficient few-shot learning technique that involves:
43
+
44
+ 1. Fine-tuning a [Sentence Transformer](https://www.sbert.net) with contrastive learning.
45
+ 2. Training a classification head with features from the fine-tuned Sentence Transformer.
46
+
47
+ ## Model Details
48
+
49
+ ### Model Description
50
+ - **Model Type:** SetFit
51
+ - **Sentence Transformer body:** [sentence-transformers/paraphrase-mpnet-base-v2](https://huggingface.co/sentence-transformers/paraphrase-mpnet-base-v2)
52
+ - **Classification head:** a [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance
53
+ - **Maximum Sequence Length:** 512 tokens
54
+ - **Number of Classes:** 26 classes
55
+ - **Training Dataset:** [dendimaki/v1](https://huggingface.co/datasets/dendimaki/v1)
56
+ <!-- - **Language:** Unknown -->
57
+ <!-- - **License:** Unknown -->
58
+
59
+ ### Model Sources
60
+
61
+ - **Repository:** [SetFit on GitHub](https://github.com/huggingface/setfit)
62
+ - **Paper:** [Efficient Few-Shot Learning Without Prompts](https://arxiv.org/abs/2209.11055)
63
+ - **Blogpost:** [SetFit: Efficient Few-Shot Learning Without Prompts](https://huggingface.co/blog/setfit)
64
+
65
+ ### Model Labels
66
+ | Label | Examples |
67
+ |:------|:-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
68
+ | 20 | <ul><li>'while the finder feels a deep sense of completeness his or her partner still has a narrativeself that thrives on external validation'</li><li>'disassembled'</li><li>'location four definitely adds a whole new perspective and can decondition a lot especially if one deepens there but yeah save that for when you feel the timing is good'</li></ul> |
69
+ | 26 | <ul><li>'i think the emptiness is a different one'</li><li>'being like a container for whats arising and the stuff thats arising'</li><li>'spaciousness or emptiness'</li></ul> |
70
+ | 27 | <ul><li>'encased in gelatin'</li><li>'feeling full of joy'</li><li>'so if i do if i meditate in a certain way i have meditated and it happens and i drop into more of a kind of equalized more still flat perception i would say or just not not perhaps not maybe not flat but its like dropping into a different dimension if you could say that like thats not really its not about the physical that much anymore as much as its a different its like residing in a different field that is more quiet and peaceful and if i sink in in my day to day life i can also go go pretty quickly to that straight away actually but i again i guess i choose not to because again somewhere along the way i think one of my teachers emphasized also feeling the fullness but thats analysis for something else but yeah ive experienced that quite a few times'</li></ul> |
71
+ | 18 | <ul><li>'mixture of personal and impersonal love'</li><li>'it sounds very plausible i think being lonely is one thing if i just sit there in my apartment you know and become more and more still and around boredom or being boring'</li><li>'popular term for this change in perception is nonduality or not two'</li></ul> |
72
+ | 28 | <ul><li>'but the shift into layer four is you know it can be an intense one and it really is very different than everything that comes before it and so you know lots of strange things can happen on the way to it in the direction of it you know sort of associated with it um and its possible that when you felt like you had made progress in that direction and then you had this other sort of experience come in that it was you know just one of those types of things in that direction'</li><li>'only reality just unfolding'</li><li>'dimensional flatness'</li></ul> |
73
+ | 16 | <ul><li>'the path of freedom remains emotionless the path of humanity'</li><li>'moments and so basically when you come out of the narrative mind you start to fill the mind moments that the narrative mind filled with sensory mind moments and so that can also account for the for the luminosity thing it doesnt necessarily have to be it can be a combination of what you said but when you when you were talking about it i was like oh it could be a mind moment thing just because you know theres more moments of sensory experience in the conscious experience'</li><li>'path of humanity'</li></ul> |
74
+ | 17 | <ul><li>'seer'</li><li>'seems like the looker is there looking out your eyes'</li><li>'with recalling memories that related to their'</li></ul> |
75
+ | 25 | <ul><li>'fluid or experiencing one layer'</li><li>'layer one level'</li><li>'pulled back to probably layer one'</li></ul> |
76
+ | 19 | <ul><li>'an example of one potential reason relates to personal love for ones child'</li><li>'or an all pervasive consciousness'</li><li>'it was when my dad died and you know i was like crying but i was like well this is just love so this is okay i wouldnt say this is i want it to stop'</li></ul> |
77
+ | 15 | <ul><li>'the thing the thing to keep in mind is that for a system for a layer four location four especially but youre sort of close enough you know youre like a hair away from the thing type system what reading those books will do is basically prime you basically primes the system'</li><li>'the peace is of a different order than that of any other layer because it is not dependent on any positionality such as i am awareness or i am'</li><li>'deeper into layer 4 in later locations the sense of unfolding diminishes until everything feels instantaneous and total '</li></ul> |
78
+ | 8 | <ul><li>'strong psychological triggers such as the death of a loved one can still cause a reaction in the system but for the most part there is persistent equanimity and joy'</li></ul> |
79
+ | 14 | <ul><li>'layer 3 can remain accessible in location 4 though usually only the deepest centerless aspects of it'</li><li>'dont have that mental abstraction'</li><li>'the subjective experience is emmeshed with deep beliefs about what is ultimately real and transitioning to and deepening into location 4 can be disconcerting'</li></ul> |
80
+ | 22 | <ul><li>'fundamentalist beliefs'</li><li>'fundamental wellbeing kind of gets more and more boring in a way'</li><li>'curcumin supplement'</li></ul> |
81
+ | 3 | <ul><li>'the boundaries between work and play blur in location 1 layer 4 each act imbued with purpose and the joy of being'</li><li>'in location 1 layer 4 the setting sun doesnt signify an end but a gentle closure a pause for reflection and gratitude'</li><li>'i can still get triggered but negative emotions fall off much faster like glimpsing into layer four by doing unprovoked happiness'</li></ul> |
82
+ | 4 | <ul><li>'memories also tend to arise less because there is an increased focus of attention on the present and because the past is no longer valued as defining the sense of self'</li><li>'when youre describing like a deeper nonduality is the absence of layer one'</li></ul> |
83
+ | 6 | <ul><li>'so you cant stay in location two but youre not able to access the depth of a layout to possibly and certainly layer three that youre able to with your eyes closed'</li><li>'cosmic love'</li><li>'layer 3 is highly accessible in location 2 however it remains relatively rare for finders to reach layer 3 persistently when they do it is often taken to be end of the path in terms of deepening further into fundamental wellbeing '</li></ul> |
84
+ | 21 | <ul><li>'psychic intuitive empathic'</li><li>'darkness'</li><li>'psychedelics'</li></ul> |
85
+ | 10 | <ul><li>'the main thing was a sense of a kind of strong gravitational pull'</li></ul> |
86
+ | 24 | <ul><li>'since 2017 was when i did finders course and transitioned'</li></ul> |
87
+ | 0 | <ul><li>'environment under trigger its more like 11 and then kind of off on my own doing my thing'</li><li>'very attached to my mind'</li></ul> |
88
+ | 11 | <ul><li>'this is partly because one is unable to deepen into it and stabilize in it and partly because it cannot be known objectivelyor even subjectively in the usual sense'</li><li>'the unfolding does not happen in anything rather it is total and complete in itself'</li></ul> |
89
+ | 1 | <ul><li>'only location one layer two seemed to get a graphic and the bird looks a little confused'</li></ul> |
90
+ | 9 | <ul><li>'feeling like youre dissolving into it'</li><li>'in location three there was a certain clarity that i dont have now because it was like less commotion or deadness because like the love would infuse every thought so a thought would come up and instead of me where i am right now i dont want to deal with it it would just be like oh its okay its lets lets just sit with it and the loving feeling would just infuse every thought and then certain judgments that id have oh well i dont really need to look at it that way i can well i can just put love in this or i can just love it so that that id say that was like the most stark contrast'</li></ul> |
91
+ | 5 | <ul><li>'something into this experience of two so my experience of this has its just now releasing a lot of the as of a couple of days ago thought it might be wise to look at this yeah so ive been experiencing you know this very strange weird nonduality type'</li><li>'shifting into layer two'</li><li>'things are seen with more distance and objectivity and one typically becomes less reactive the downside of this is that it can be a great place to escape the mind and disassociate from psychological conditioning this is usually whats meant when people speak about spiritual bypassing '</li></ul> |
92
+ | 12 | <ul><li>'this can lead to a wide range of outcomes from extraordinary life results to some of the amoral behavior observed in late location teachers'</li><li>'mind is very quiet'</li><li>'essentially this is a metaawareness of what is happening in the mind but there is no sense of being able to engage with it like there is in previous locations '</li></ul> |
93
+ | 23 | <ul><li>'until youre feeling deeper or more stable in fundamental wellbeing'</li><li>' an event in fundamental wellbeing for a while'</li></ul> |
94
+
95
+ ## Evaluation
96
+
97
+ ### Metrics
98
+ | Label | Accuracy |
99
+ |:--------|:---------|
100
+ | **all** | 0.4635 |
101
+
102
+ ## Uses
103
+
104
+ ### Direct Use for Inference
105
+
106
+ First install the SetFit library:
107
+
108
+ ```bash
109
+ pip install setfit
110
+ ```
111
+
112
+ Then you can load this model and run inference.
113
+
114
+ ```python
115
+ from setfit import SetFitModel
116
+
117
+ # Download from the 🤗 Hub
118
+ model = SetFitModel.from_pretrained("dendimaki/fewshot-model")
119
+ # Run inference
120
+ preds = model("pervading presence")
121
+ ```
122
+
123
+ <!--
124
+ ### Downstream Use
125
+
126
+ *List how someone could finetune this model on their own dataset.*
127
+ -->
128
+
129
+ <!--
130
+ ### Out-of-Scope Use
131
+
132
+ *List how the model may foreseeably be misused and address what users ought not to do with the model.*
133
+ -->
134
+
135
+ <!--
136
+ ## Bias, Risks and Limitations
137
+
138
+ *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
139
+ -->
140
+
141
+ <!--
142
+ ### Recommendations
143
+
144
+ *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
145
+ -->
146
+
147
+ ## Training Details
148
+
149
+ ### Training Set Metrics
150
+ | Training set | Min | Median | Max |
151
+ |:-------------|:----|:--------|:----|
152
+ | Word count | 1 | 21.9052 | 247 |
153
+
154
+ | Label | Training Sample Count |
155
+ |:------|:----------------------|
156
+ | 0 | 2 |
157
+ | 1 | 1 |
158
+ | 3 | 5 |
159
+ | 4 | 2 |
160
+ | 5 | 4 |
161
+ | 6 | 11 |
162
+ | 8 | 1 |
163
+ | 9 | 2 |
164
+ | 10 | 1 |
165
+ | 11 | 2 |
166
+ | 12 | 3 |
167
+ | 14 | 4 |
168
+ | 15 | 8 |
169
+ | 16 | 8 |
170
+ | 17 | 11 |
171
+ | 18 | 28 |
172
+ | 19 | 25 |
173
+ | 20 | 14 |
174
+ | 21 | 4 |
175
+ | 22 | 7 |
176
+ | 23 | 2 |
177
+ | 24 | 1 |
178
+ | 25 | 13 |
179
+ | 26 | 30 |
180
+ | 27 | 36 |
181
+ | 28 | 7 |
182
+
183
+ ### Training Hyperparameters
184
+ - batch_size: (16, 16)
185
+ - num_epochs: (1, 1)
186
+ - max_steps: -1
187
+ - sampling_strategy: oversampling
188
+ - num_iterations: 20
189
+ - body_learning_rate: (2e-05, 2e-05)
190
+ - head_learning_rate: 2e-05
191
+ - loss: CosineSimilarityLoss
192
+ - distance_metric: cosine_distance
193
+ - margin: 0.25
194
+ - end_to_end: False
195
+ - use_amp: False
196
+ - warmup_proportion: 0.1
197
+ - seed: 42
198
+ - eval_max_steps: -1
199
+ - load_best_model_at_end: False
200
+
201
+ ### Training Results
202
+ | Epoch | Step | Training Loss | Validation Loss |
203
+ |:------:|:----:|:-------------:|:---------------:|
204
+ | 0.0017 | 1 | 0.252 | - |
205
+ | 0.0862 | 50 | 0.1891 | - |
206
+ | 0.1724 | 100 | 0.1793 | - |
207
+ | 0.2586 | 150 | 0.1848 | - |
208
+ | 0.3448 | 200 | 0.1033 | - |
209
+ | 0.4310 | 250 | 0.0473 | - |
210
+ | 0.5172 | 300 | 0.1213 | - |
211
+ | 0.6034 | 350 | 0.0343 | - |
212
+ | 0.6897 | 400 | 0.0276 | - |
213
+ | 0.7759 | 450 | 0.0262 | - |
214
+ | 0.8621 | 500 | 0.0425 | - |
215
+ | 0.9483 | 550 | 0.0482 | - |
216
+
217
+ ### Framework Versions
218
+ - Python: 3.10.12
219
+ - SetFit: 1.0.3
220
+ - Sentence Transformers: 2.7.0
221
+ - Transformers: 4.40.1
222
+ - PyTorch: 2.2.1+cu121
223
+ - Datasets: 2.19.0
224
+ - Tokenizers: 0.19.1
225
+
226
+ ## Citation
227
+
228
+ ### BibTeX
229
+ ```bibtex
230
+ @article{https://doi.org/10.48550/arxiv.2209.11055,
231
+ doi = {10.48550/ARXIV.2209.11055},
232
+ url = {https://arxiv.org/abs/2209.11055},
233
+ author = {Tunstall, Lewis and Reimers, Nils and Jo, Unso Eun Seo and Bates, Luke and Korat, Daniel and Wasserblat, Moshe and Pereg, Oren},
234
+ keywords = {Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences},
235
+ title = {Efficient Few-Shot Learning Without Prompts},
236
+ publisher = {arXiv},
237
+ year = {2022},
238
+ copyright = {Creative Commons Attribution 4.0 International}
239
+ }
240
+ ```
241
+
242
+ <!--
243
+ ## Glossary
244
+
245
+ *Clearly define terms in order to be accessible across audiences.*
246
+ -->
247
+
248
+ <!--
249
+ ## Model Card Authors
250
+
251
+ *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
252
+ -->
253
+
254
+ <!--
255
+ ## Model Card Contact
256
+
257
+ *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
258
+ -->
config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "sentence-transformers/paraphrase-mpnet-base-v2",
3
+ "architectures": [
4
+ "MPNetModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "eos_token_id": 2,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-05,
15
+ "max_position_embeddings": 514,
16
+ "model_type": "mpnet",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 1,
20
+ "relative_attention_num_buckets": 32,
21
+ "torch_dtype": "float32",
22
+ "transformers_version": "4.40.1",
23
+ "vocab_size": 30527
24
+ }
config_sentence_transformers.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "2.0.0",
4
+ "transformers": "4.7.0",
5
+ "pytorch": "1.9.0+cu102"
6
+ },
7
+ "prompts": {},
8
+ "default_prompt_name": null
9
+ }
config_setfit.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "normalize_embeddings": false,
3
+ "labels": null
4
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f19277bddf334d61ed77d8bc959868cb16e2468aa900fb88f0d0eb9f6b83286
3
+ size 437967672
model_head.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca818e7848f7965b7b584b1cb5076033e393f18efc5821b4d3c7d9a221a57001
3
+ size 160999
modules.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ }
14
+ ]
sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 512,
3
+ "do_lower_case": false
4
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "cls_token": {
10
+ "content": "<s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "eos_token": {
17
+ "content": "</s>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "mask_token": {
24
+ "content": "<mask>",
25
+ "lstrip": true,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "pad_token": {
31
+ "content": "<pad>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ },
37
+ "sep_token": {
38
+ "content": "</s>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false
43
+ },
44
+ "unk_token": {
45
+ "content": "[UNK]",
46
+ "lstrip": false,
47
+ "normalized": false,
48
+ "rstrip": false,
49
+ "single_word": false
50
+ }
51
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,59 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<s>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<pad>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "</s>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "104": {
28
+ "content": "[UNK]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "30526": {
36
+ "content": "<mask>",
37
+ "lstrip": true,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "bos_token": "<s>",
45
+ "clean_up_tokenization_spaces": true,
46
+ "cls_token": "<s>",
47
+ "do_basic_tokenize": true,
48
+ "do_lower_case": true,
49
+ "eos_token": "</s>",
50
+ "mask_token": "<mask>",
51
+ "model_max_length": 512,
52
+ "never_split": null,
53
+ "pad_token": "<pad>",
54
+ "sep_token": "</s>",
55
+ "strip_accents": null,
56
+ "tokenize_chinese_chars": true,
57
+ "tokenizer_class": "MPNetTokenizer",
58
+ "unk_token": "[UNK]"
59
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff