carlesoctav commited on
Commit
023d036
1 Parent(s): 74e03c8

Add SetFit model

Browse files
1_Pooling/config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 768,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": true,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false,
7
+ "pooling_mode_weightedmean_tokens": false,
8
+ "pooling_mode_lasttoken": false,
9
+ "include_prompt": true
10
+ }
README.md ADDED
@@ -0,0 +1,249 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language: en
3
+ license: apache-2.0
4
+ library_name: setfit
5
+ tags:
6
+ - setfit
7
+ - sentence-transformers
8
+ - text-classification
9
+ - generated_from_setfit_trainer
10
+ base_model: sentence-transformers/paraphrase-mpnet-base-v2
11
+ metrics:
12
+ - accuracy
13
+ widget:
14
+ - text: great movie, so close to perfection let me get this straight. this is a brilliant
15
+ brilliant refreshingly brutal movie.i'm glad they didn't soften the general malevolence,
16
+ but i feel they missed out on what i consider the most pivotal point of the book.paul
17
+ drinks the water of life. with it his genetic memory is unlocked, he can foresee
18
+ the actions of people in the future. the golden path is laid out. and so pursues
19
+ the mind altering awakening, leaving him a husk; trapped in one path of fate -
20
+ trapped between his own ego and the true path needed for humanity. in the movie,
21
+ paul drink bad, paul wake up. paul president with superpower!i understand that
22
+ it's a very hard thing to portray for an audience but i think i was just really
23
+ hoping for villeneuve to emphasise the importance of that part and it felt quite
24
+ rushed in that regard.but i doubt they'll make a movie about a big virgin worm
25
+ so prescience might not matter too much.
26
+ - text: absolutely breathtaking the movie is the complete cinematic experience. i
27
+ loved every single line every moment every little thing that makes this movie.the
28
+ only thing that is bothering me is the thirst so bad for the next part.i felt
29
+ like i was in the movie riding a sand worm, i was a fremen. i felt the pain the
30
+ wonder the joy the anger. this felt like reading the book and you just can't stop.
31
+ the excellence of this movie is not only the cast or the story it is the very
32
+ making of it. i loved every dialogue that was uttered. its just a masterpiece.though
33
+ there is a stagnant pace in between it doesn't seem to matter. because most of
34
+ the second part of the movie is such a cliff hanger. 6 out of 10 found this helpful.
35
+ was this review helpful? sign in to vote. permalink
36
+ - text: 'let''s be serious, guys.. appreciate that everyone is entitled to their opinion,
37
+ so here''s mine: anyone giving this less than a solid 9 needs to re-evaluate themselves
38
+ as a person. because you either have no imagination or are just generally a negative
39
+ human. this film has everything and is a modern day great. easily the best cinematic
40
+ experience i''ve ever had, comparable to films like the dark knight trilogy and
41
+ the original star wars films.for a nearly three hour long film, basically nobody
42
+ got up to go for a toilet break and the entire time i felt totally present, gripped
43
+ by it.don''t listen to anyone on here leaving poor reviews. go and watch the film
44
+ and see the magic for yourself. 8 out of 13 found this helpful. was this review
45
+ helpful? sign in to vote. permalink'
46
+ - text: phenomenal this movie was particularly gorgeous and exciting giving all the
47
+ key moments and suspense that anybody of the sort would love, this movie brings
48
+ the suspense and excitement to keep you engaged and always cautious of what's
49
+ next, a truly wonderful story that is acted so perfectly and well, this adaptation
50
+ has brung the story alive and in the spotlight proving there is not only a lot
51
+ to it but also that it has a lot more to come and personally i want to see it
52
+ all. i left the theater thoroughly wanting even more for the story and continuing
53
+ on that i can't wait for what is to come of this movie. it is truly a must watch
54
+ masterpiece. 4 out of 6 found this helpful. was this review helpful? sign in to
55
+ vote. permalink
56
+ - text: film of the decade i've always wished to watch films like lord of the rings
57
+ and star wars in theaters, but i was simply born too late. dune 2 made me feel
58
+ like i was watching those movies in theaters, the epic sweaping shots, the massive
59
+ amount of extras, the attention to detail, the costumes, every single fight looked
60
+ like they spent days choreographing it. the soundtrack was the best i heard since
61
+ interstellar, and it matched the mood at every point. honestly i thought film
62
+ was going down, disney is losing it and they own almost everything. but dune 2
63
+ restored my hope in movies and actually made me want to pursue a career in film.
64
+ overall, this movie was epic and easily deserves a 10 star rating. 1 out of 1
65
+ found this helpful. was this review helpful? sign in to vote. permalink
66
+ pipeline_tag: text-classification
67
+ inference: true
68
+ ---
69
+
70
+ # SetFit with sentence-transformers/paraphrase-mpnet-base-v2 on data/raw/15239678.jsonl
71
+
72
+ This is a [SetFit](https://github.com/huggingface/setfit) model that can be used for Text Classification. This SetFit model uses [sentence-transformers/paraphrase-mpnet-base-v2](https://huggingface.co/sentence-transformers/paraphrase-mpnet-base-v2) as the Sentence Transformer embedding model. A [SetFitHead](huggingface.co/docs/setfit/reference/main#setfit.SetFitHead) instance is used for classification.
73
+
74
+ The model has been trained using an efficient few-shot learning technique that involves:
75
+
76
+ 1. Fine-tuning a [Sentence Transformer](https://www.sbert.net) with contrastive learning.
77
+ 2. Training a classification head with features from the fine-tuned Sentence Transformer.
78
+
79
+ ## Model Details
80
+
81
+ ### Model Description
82
+ - **Model Type:** SetFit
83
+ - **Sentence Transformer body:** [sentence-transformers/paraphrase-mpnet-base-v2](https://huggingface.co/sentence-transformers/paraphrase-mpnet-base-v2)
84
+ - **Classification head:** a [SetFitHead](huggingface.co/docs/setfit/reference/main#setfit.SetFitHead) instance
85
+ - **Maximum Sequence Length:** 512 tokens
86
+ - **Number of Classes:** 2 classes
87
+ <!-- - **Training Dataset:** [Unknown](https://huggingface.co/datasets/unknown) -->
88
+ - **Language:** en
89
+ - **License:** apache-2.0
90
+
91
+ ### Model Sources
92
+
93
+ - **Repository:** [SetFit on GitHub](https://github.com/huggingface/setfit)
94
+ - **Paper:** [Efficient Few-Shot Learning Without Prompts](https://arxiv.org/abs/2209.11055)
95
+ - **Blogpost:** [SetFit: Efficient Few-Shot Learning Without Prompts](https://huggingface.co/blog/setfit)
96
+
97
+ ### Model Labels
98
+ | Label | Examples |
99
+ |:---------|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
100
+ | positive | <ul><li>"good sequel dune part ii is a very good sequel. the world building in this movie is great and really establishes the fremen culture. i love how the movie dives deep into how they view culture and religion and the split that they have over their belief in paul. timothee chalamet is excellent as paul atreides. his character arc is amazing. he starts off as someone who just wants to help and then through his time with the fremen he starts to use them and their faith to get his own revenge. zendaya was also great and her character's dynamic with paul was very fleshed out. i loved how she went from loving him to despising what he represents. florence pugh was a good addition here although she didn't have much to do. austin butler stole the show in my opinion. he was a perfect villain and his pure psychotic nature was frightening. the costumes, visual effects, and set design all looked great. i especially love the design of the black and white planet. there were a lot of cool things they did with it, like the fireworks for example. the action scenes are pretty good. the cinematography is very good and denis villeneuve crushed it directing. hans zimmer delivers a killer score here as he always does. i only have a few minor issues with the movie. just like the first movie i think the pacing was a little off and despite rebecca ferguson giving a great performance i didn't really think her storyline was that interesting or well explained. overall dune part ii is a really good movie and if you liked the first one you'll have a great time here."</li><li>"am i the only one who's not seeing it? i mean, yeah, it's very entertaining and, of course, very visually stunning. the set pieces, the cinematography, the use of visual effects and lights, the sound design and music, all, absolutely amazing and almost literally stunning!but then? i'm not really seeing much after that. as i have not read the books, this movie was a total mystery to me. there's barely any dialog--at least not any that would explain anything what's going on at all. the world and the technology etc just doesn't make much sense to me.none of the characters are particularly interesting, to be honest. they don't really have that much personality to them, and even if they did, they didn't really make me care about them all that much.i don't know, i'm a bit conflicted, it wasn't a bad movie and, as i said, it was entertaining and visually mesmerizing, but it lacked the depth that i was expecting of a world this size and this rich with lore and history. maybe the movie makers assumed everyone has read the books? as someone to who the world is not familiar at all, it just seems rather confusing and strange. i feel like they just focused on making it as visually awesome as they can (in which they arguably succeeded), but left the story on the pages of the books."</li><li>'dune: part two it\'s totally amazing best sf movie i just saw the new movie "dune: part two" and i was speechless.it was amazing, full of creativity and an unforgettable action.from the sensational footage, which i will constantly praise, to the story and everything that the dune universe means.shocking scenes, dramatic scenes, romantic scenes but also full scenes, full of action, carnage and explosions.something i\'ve been waiting for so long and it was more perfect than i expected...in imax, the sound was so good that at any explosion or need for bass, it creates a shock, an earthquake that gives incredible pleasure.you must see it and i declare that it is the best film of 2024 and i don\'t think that any of the upcoming movies will be better. waiting for oscars and all the wins possible .'</li></ul> |
101
+ | negative | <ul><li>'boring, wow, so very boring ...i walked out about 2hrs in because i just couldn\'t anymore.i don\'t understand how a movie with so much time and space could possibly exclude so much and move the plot so little.no storytelling: it feels like an extended montage where paul\'s cheerleading base grows along with his relationship with chani but we don\'t know how or why.poorly written: female leads felt like wallpaper rather than the badass characters who carry paul through the book. some reviews here are saying the movie was true to the book, which i don\'t understand because they also don\'t even touch on the space guild.vapid acting: paul. bautista can only scream? everyone seemed demoralized in their roles (i would be too if i was zendaya, rebecca ferguson, or anyone else who actually wanted to embrace the role of a lifetime and was so absolutely diminished)."cinematography": i must be hard to get more than a handful of creative shots of sand. what was happening on the harkonnen home world? honestly the blown out black and white scene felt really lazy and cheap. at least after that there was more desert.christopher walken: was his casting supposed to be some sort of meta joke? even his performance rang flat. i thought the emperor was supposed to be eternally young due to spice consumption. how this character is supposed to be the center / leader of the universe is anyone\'s guess.i\'ll wait until it\'s streaming and finish at my convenience to see if any of this gets salvaged, i just can\'t imagine how it could be.'</li><li>"great i you didn't read the books short spoiler-free version: a was really looking forward to this movie. having read the books multiple times, i left the cinema feeling cheated en confused. granted the visuals and the music are astounding. the actors perform very well and the story is fine if haven't read the books. and there's the problem. i you, like me, have read the book you will be thinking 'why?' every 10 minutes or so. villeneuve has made so many changes to the story, i hardly recognize it as dune anymore. and that makes me sad. i try to view the movie and the book as separate things. but how many changes can one make in the story before it deviates to much? after all it is a movie adaptation of the book, there should be enough of it you can recognize. in this the director has, sadly, mist the marc.after the movie i kept sitting staring as the credits rolled by, i was the last one to leave an i felt cheated and sad. my advice is, if you love the books, wait for the streaming services to pick it up. if you haven't read the books, go an see it, you'll love it although perhaps it might be a bit long for your taste.longer version (spoilers!) so, what went wrong? is do understand that you have tot make some choices in what you keep en leave out in the story when making a story. you merge some characters, leave some out. delete scenes that a nice but not really necessary in the main plotline. (like the banket-scene in the first movie). but villenueve and the writers have made al lot of changes that impact the story in so much that is totally deviates from the book. as a fan i can not get my head around soms of them: chani is not supporting at all of paul, she is mostly angry and against all that paul wants tot do. the actors try there best but i miss the chemistry from the previous adaptations and the books. in the ending there is no 'we are concubines' scene. chani leaves paul and rides out in the dessert. why?jessica in the books is a strong well trained bene gesserit. in the movie she goes from a weak, sniveling, manipulative woman. not the strong reverent mother who takes care of het people en supports paul and grows closer tot chani. i can't understand why.stilgar in the books is a strong and wise leader who teaches paul and gradually gives him his trust. in the movie he is transformed into an religious zealot from the south of arrakis, were all fremen blindly believe in the prophecy. (in the north the fremen are non-believers, essentially making it 2 different tribes of fremen, again why?).and then there is no mention of the ecological theme, instead of the water of life they threaten with nukes on the spicefields, no spicing guild and a feeble weak emperor (poor christhoper walken, he tries)) i can get why feyd is changed into a psychopath (not much change from the books) and why hawat and alia are left out completely (sad but understandable). but the rest? as i stated, i feld robbed, sad and very disappointed. it could have been better, but is was a mocking of the real thing."</li><li>"the whole movie happens in slow motion once you realize that the movie is happening in slow motion, you can't un-see it:slowly pan over beautiful landscape, close-up of someone's face. talk. very. slowly. one. word ..... at a time. pan out. bwaaaap/loud noise, next scene, rinse and repeat, stir in the occasional action scene.there are a lot of dialog scenes in this movie, but not much is said. i understand that the book is both thick and dense, so showing everything in slow motion seems an odd choice. i honestly think, if you watch this at home, you could watch it at 2x speed and it might be a decent way to spend an hour and 15 minutes.the battle scenes were also kinda dumb. this is ~8000 years in the future, inter-galactic space flight exists, planes, helicopters, nuclear warheads, guns, lazers, bazookas and more all exist, yet people decide to go to battle as if it's the middle ages - fighting with knives! you don't need to be a messiah to realize that you could bring guns to a knife fight. i'll give them poetic liscense on this one and perhaps we can write it off to the aesthetics, but it did make some of the scenes a little laughable once you realize what's going on. maybe they explain this in the book?i did not enjoy dune 1 for much the same reasons, but so many people were saying this one was better. fool me once, shame on you. fool me twice, shame on me! i will not be fooled a third time!"</li></ul> |
102
+
103
+ ## Uses
104
+
105
+ ### Direct Use for Inference
106
+
107
+ First install the SetFit library:
108
+
109
+ ```bash
110
+ pip install setfit
111
+ ```
112
+
113
+ Then you can load this model and run inference.
114
+
115
+ ```python
116
+ from setfit import SetFitModel
117
+
118
+ # Download from the 🤗 Hub
119
+ model = SetFitModel.from_pretrained("carlesoctav/SentimentClassifierDune")
120
+ # Run inference
121
+ preds = model("phenomenal this movie was particularly gorgeous and exciting giving all the key moments and suspense that anybody of the sort would love, this movie brings the suspense and excitement to keep you engaged and always cautious of what's next, a truly wonderful story that is acted so perfectly and well, this adaptation has brung the story alive and in the spotlight proving there is not only a lot to it but also that it has a lot more to come and personally i want to see it all. i left the theater thoroughly wanting even more for the story and continuing on that i can't wait for what is to come of this movie. it is truly a must watch masterpiece. 4 out of 6 found this helpful. was this review helpful? sign in to vote. permalink")
122
+ ```
123
+
124
+ <!--
125
+ ### Downstream Use
126
+
127
+ *List how someone could finetune this model on their own dataset.*
128
+ -->
129
+
130
+ <!--
131
+ ### Out-of-Scope Use
132
+
133
+ *List how the model may foreseeably be misused and address what users ought not to do with the model.*
134
+ -->
135
+
136
+ <!--
137
+ ## Bias, Risks and Limitations
138
+
139
+ *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
140
+ -->
141
+
142
+ <!--
143
+ ### Recommendations
144
+
145
+ *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
146
+ -->
147
+
148
+ ## Training Details
149
+
150
+ ### Training Set Metrics
151
+ | Training set | Min | Median | Max |
152
+ |:-------------|:----|:---------|:----|
153
+ | Word count | 107 | 215.2273 | 972 |
154
+
155
+ | Label | Training Sample Count |
156
+ |:---------|:----------------------|
157
+ | negative | 99 |
158
+ | positive | 99 |
159
+
160
+ ### Training Hyperparameters
161
+ - batch_size: (16, 16)
162
+ - num_epochs: (1, 1)
163
+ - max_steps: -1
164
+ - sampling_strategy: oversampling
165
+ - body_learning_rate: (2e-05, 1e-05)
166
+ - head_learning_rate: 0.01
167
+ - loss: CosineSimilarityLoss
168
+ - distance_metric: cosine_distance
169
+ - margin: 0.25
170
+ - end_to_end: False
171
+ - use_amp: False
172
+ - warmup_proportion: 0.1
173
+ - seed: 42
174
+ - eval_max_steps: -1
175
+ - load_best_model_at_end: True
176
+
177
+ ### Training Results
178
+ | Epoch | Step | Training Loss | Validation Loss |
179
+ |:-------:|:--------:|:-------------:|:---------------:|
180
+ | 0.0008 | 1 | 0.2606 | - |
181
+ | 0.0404 | 50 | 0.1578 | - |
182
+ | 0.0808 | 100 | 0.0066 | - |
183
+ | 0.1212 | 150 | 0.0004 | - |
184
+ | 0.1616 | 200 | 0.0003 | - |
185
+ | 0.2019 | 250 | 0.0005 | - |
186
+ | 0.2423 | 300 | 0.0002 | - |
187
+ | 0.2827 | 350 | 0.0003 | - |
188
+ | 0.3231 | 400 | 0.0001 | - |
189
+ | 0.3635 | 450 | 0.0001 | - |
190
+ | 0.4039 | 500 | 0.0001 | - |
191
+ | 0.4443 | 550 | 0.0001 | - |
192
+ | 0.4847 | 600 | 0.0 | - |
193
+ | 0.5250 | 650 | 0.0 | - |
194
+ | 0.5654 | 700 | 0.0 | - |
195
+ | 0.6058 | 750 | 0.0 | - |
196
+ | 0.6462 | 800 | 0.0 | - |
197
+ | 0.6866 | 850 | 0.0 | - |
198
+ | 0.7270 | 900 | 0.0 | - |
199
+ | 0.7674 | 950 | 0.0 | - |
200
+ | 0.8078 | 1000 | 0.0 | - |
201
+ | 0.8481 | 1050 | 0.0 | - |
202
+ | 0.8885 | 1100 | 0.0 | - |
203
+ | 0.9289 | 1150 | 0.0 | - |
204
+ | 0.9693 | 1200 | 0.0 | - |
205
+ | **1.0** | **1238** | **-** | **0.1555** |
206
+
207
+ * The bold row denotes the saved checkpoint.
208
+ ### Framework Versions
209
+ - Python: 3.10.11
210
+ - SetFit: 1.0.3
211
+ - Sentence Transformers: 2.5.1
212
+ - Transformers: 4.38.2
213
+ - PyTorch: 2.0.1
214
+ - Datasets: 2.18.0
215
+ - Tokenizers: 0.15.2
216
+
217
+ ## Citation
218
+
219
+ ### BibTeX
220
+ ```bibtex
221
+ @article{https://doi.org/10.48550/arxiv.2209.11055,
222
+ doi = {10.48550/ARXIV.2209.11055},
223
+ url = {https://arxiv.org/abs/2209.11055},
224
+ author = {Tunstall, Lewis and Reimers, Nils and Jo, Unso Eun Seo and Bates, Luke and Korat, Daniel and Wasserblat, Moshe and Pereg, Oren},
225
+ keywords = {Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences},
226
+ title = {Efficient Few-Shot Learning Without Prompts},
227
+ publisher = {arXiv},
228
+ year = {2022},
229
+ copyright = {Creative Commons Attribution 4.0 International}
230
+ }
231
+ ```
232
+
233
+ <!--
234
+ ## Glossary
235
+
236
+ *Clearly define terms in order to be accessible across audiences.*
237
+ -->
238
+
239
+ <!--
240
+ ## Model Card Authors
241
+
242
+ *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
243
+ -->
244
+
245
+ <!--
246
+ ## Model Card Contact
247
+
248
+ *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
249
+ -->
config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "exp/Dune2Classifier/step_1238",
3
+ "architectures": [
4
+ "MPNetModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "eos_token_id": 2,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-05,
15
+ "max_position_embeddings": 514,
16
+ "model_type": "mpnet",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 1,
20
+ "relative_attention_num_buckets": 32,
21
+ "torch_dtype": "float32",
22
+ "transformers_version": "4.38.2",
23
+ "vocab_size": 30527
24
+ }
config_sentence_transformers.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "2.0.0",
4
+ "transformers": "4.7.0",
5
+ "pytorch": "1.9.0+cu102"
6
+ },
7
+ "prompts": {},
8
+ "default_prompt_name": null
9
+ }
config_setfit.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "labels": [
3
+ "negative",
4
+ "positive"
5
+ ],
6
+ "normalize_embeddings": false
7
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02b256aa2e310b239aa4b3920a575c03fc53d075f1631e1edf214349d9789a7e
3
+ size 437967672
model_head.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:034f8755f06e904c8b226849b7d18e3df523dfbfd93710233551dc41e600eaeb
3
+ size 7640
modules.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ }
14
+ ]
sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 512,
3
+ "do_lower_case": false
4
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "cls_token": {
10
+ "content": "<s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "eos_token": {
17
+ "content": "</s>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "mask_token": {
24
+ "content": "<mask>",
25
+ "lstrip": true,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "pad_token": {
31
+ "content": "<pad>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ },
37
+ "sep_token": {
38
+ "content": "</s>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false
43
+ },
44
+ "unk_token": {
45
+ "content": "[UNK]",
46
+ "lstrip": false,
47
+ "normalized": false,
48
+ "rstrip": false,
49
+ "single_word": false
50
+ }
51
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<s>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<pad>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "</s>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "104": {
28
+ "content": "[UNK]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "30526": {
36
+ "content": "<mask>",
37
+ "lstrip": true,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "bos_token": "<s>",
45
+ "clean_up_tokenization_spaces": true,
46
+ "cls_token": "<s>",
47
+ "do_basic_tokenize": true,
48
+ "do_lower_case": true,
49
+ "eos_token": "</s>",
50
+ "mask_token": "<mask>",
51
+ "max_length": 512,
52
+ "model_max_length": 512,
53
+ "never_split": null,
54
+ "pad_to_multiple_of": null,
55
+ "pad_token": "<pad>",
56
+ "pad_token_type_id": 0,
57
+ "padding_side": "right",
58
+ "sep_token": "</s>",
59
+ "stride": 0,
60
+ "strip_accents": null,
61
+ "tokenize_chinese_chars": true,
62
+ "tokenizer_class": "MPNetTokenizer",
63
+ "truncation_side": "right",
64
+ "truncation_strategy": "longest_first",
65
+ "unk_token": "[UNK]"
66
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff