Add SetFit model
Browse files- 1_Pooling/config.json +10 -0
- README.md +293 -0
- config.json +32 -0
- config_sentence_transformers.json +10 -0
- config_setfit.json +4 -0
- model.safetensors +3 -0
- model_head.pkl +3 -0
- modules.json +20 -0
- sentence_bert_config.json +4 -0
- special_tokens_map.json +37 -0
- tokenizer.json +0 -0
- tokenizer_config.json +57 -0
- vocab.txt +0 -0
1_Pooling/config.json
ADDED
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"word_embedding_dimension": 768,
|
3 |
+
"pooling_mode_cls_token": true,
|
4 |
+
"pooling_mode_mean_tokens": false,
|
5 |
+
"pooling_mode_max_tokens": false,
|
6 |
+
"pooling_mode_mean_sqrt_len_tokens": false,
|
7 |
+
"pooling_mode_weightedmean_tokens": false,
|
8 |
+
"pooling_mode_lasttoken": false,
|
9 |
+
"include_prompt": true
|
10 |
+
}
|
README.md
ADDED
@@ -0,0 +1,293 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
base_model: BAAI/bge-base-en-v1.5
|
3 |
+
library_name: setfit
|
4 |
+
metrics:
|
5 |
+
- accuracy
|
6 |
+
pipeline_tag: text-classification
|
7 |
+
tags:
|
8 |
+
- setfit
|
9 |
+
- sentence-transformers
|
10 |
+
- text-classification
|
11 |
+
- generated_from_setfit_trainer
|
12 |
+
widget:
|
13 |
+
- text: 'Reasoning:
|
14 |
+
|
15 |
+
1. Context Grounding: The provided answer reflects the key identifiers of a funnel
|
16 |
+
spider found in the document, such as the dark brown or black body, hard shiny
|
17 |
+
carapace, and large fangs.
|
18 |
+
|
19 |
+
2. Relevance: The answer directly addresses the question of how to identify a
|
20 |
+
funnel spider with relevant details.
|
21 |
+
|
22 |
+
3. Conciseness: The answer is clear, with pertinent details, and avoids unnecessary
|
23 |
+
information.
|
24 |
+
|
25 |
+
|
26 |
+
Final evaluation:'
|
27 |
+
- text: 'Evaluation:
|
28 |
+
|
29 |
+
|
30 |
+
1. Context Grounding:
|
31 |
+
|
32 |
+
The answer is well-supported by the provided document. It correctly mentions creating
|
33 |
+
margins, using a 12-point font, double-spacing text, creating a running header,
|
34 |
+
typing the heading in the upper left corner, and starting the body of the paper
|
35 |
+
with left-aligned text.
|
36 |
+
|
37 |
+
|
38 |
+
2. Relevance:
|
39 |
+
|
40 |
+
The answer is relevant to the question asked and specifically addresses how to
|
41 |
+
write a paper in MLA format. It does not deviate into unrelated topics.
|
42 |
+
|
43 |
+
|
44 |
+
3. Conciseness:
|
45 |
+
|
46 |
+
The answer is clear and to the point, avoiding unnecessary information, and covers
|
47 |
+
the key steps required to write in MLA format.
|
48 |
+
|
49 |
+
|
50 |
+
Final Result:'
|
51 |
+
- text: 'The answer provided is partially correct but lacks several important details
|
52 |
+
covered in the document. It highlights the importance of grades in specific subjects
|
53 |
+
and getting clinical experience, but misses other key steps such as involving
|
54 |
+
oneself in extracurricular activities, understanding application procedures, preparing
|
55 |
+
extensively for the MCAT, and engaging with advisors and peers.
|
56 |
+
|
57 |
+
|
58 |
+
**Reasoning:**
|
59 |
+
|
60 |
+
1. **Context Grounding:** The answer does cover some points mentioned in the document
|
61 |
+
such as focusing on grades in specific subjects (BCPM) and getting clinical experience.
|
62 |
+
However, it omits several other detailed recommendations provided in the document.
|
63 |
+
|
64 |
+
2. **Relevance:** The answer is somewhat relevant to the question but fails to
|
65 |
+
address the full scope of steps required to get into medical school as outlined
|
66 |
+
in the document.
|
67 |
+
|
68 |
+
3. **Conciseness:** While the answer is concise, it lacks the critical breadth
|
69 |
+
needed to be truly adequate and comprehensive for getting into medical school.
|
70 |
+
|
71 |
+
|
72 |
+
**Final Evaluation: **'
|
73 |
+
- text: 'Evaluation:
|
74 |
+
|
75 |
+
|
76 |
+
1. Context Grounding: The answer is largely grounded in the provided document
|
77 |
+
but adds some minor details that were not explicitly mentioned, such as "putting
|
78 |
+
the clothes on top of you to cover your body."
|
79 |
+
|
80 |
+
2. Relevance: The answer is highly relevant to the question asked and focuses
|
81 |
+
on strategies and techniques for playing hide and seek, which is directly related
|
82 |
+
to the document''s content.
|
83 |
+
|
84 |
+
3. Conciseness: The answer is slightly long-winded but generally clear. It could
|
85 |
+
be more concise by removing extraneous details.
|
86 |
+
|
87 |
+
|
88 |
+
Final result:'
|
89 |
+
- text: "**Evaluation Reasoning:**\n\n1. **Context Grounding:**\n - **Weak Grounding:**\
|
90 |
+
\ The provided instructions for making a saline solution and administering it\
|
91 |
+
\ are not entirely accurate based on the document. The document specifies different\
|
92 |
+
\ proportions (1 cup water, 1/2 teaspoon salt, 1/2 teaspoon baking soda) and advises\
|
93 |
+
\ against overly inserting the bulb.\n \n2. **Relevance:**\n - **Partially\
|
94 |
+
\ Relevant:** The answer attempts to address the question about treating a baby's\
|
95 |
+
\ cough but contains inaccuracies and some deviations that make it only partially\
|
96 |
+
\ relevant.\n\n3. **Conciseness:**\n - **Problems with Conciseness:** The answer\
|
97 |
+
\ includes extraneous details not needed to succinctly treat a baby’s cough, especially\
|
98 |
+
\ the inclusion of incorrect proportion information which adds confusion rather\
|
99 |
+
\ than clarity.\n\n**Final Evaluation: **"
|
100 |
+
inference: true
|
101 |
+
model-index:
|
102 |
+
- name: SetFit with BAAI/bge-base-en-v1.5
|
103 |
+
results:
|
104 |
+
- task:
|
105 |
+
type: text-classification
|
106 |
+
name: Text Classification
|
107 |
+
dataset:
|
108 |
+
name: Unknown
|
109 |
+
type: unknown
|
110 |
+
split: test
|
111 |
+
metrics:
|
112 |
+
- type: accuracy
|
113 |
+
value: 0.8648648648648649
|
114 |
+
name: Accuracy
|
115 |
+
---
|
116 |
+
|
117 |
+
# SetFit with BAAI/bge-base-en-v1.5
|
118 |
+
|
119 |
+
This is a [SetFit](https://github.com/huggingface/setfit) model that can be used for Text Classification. This SetFit model uses [BAAI/bge-base-en-v1.5](https://huggingface.co/BAAI/bge-base-en-v1.5) as the Sentence Transformer embedding model. A [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance is used for classification.
|
120 |
+
|
121 |
+
The model has been trained using an efficient few-shot learning technique that involves:
|
122 |
+
|
123 |
+
1. Fine-tuning a [Sentence Transformer](https://www.sbert.net) with contrastive learning.
|
124 |
+
2. Training a classification head with features from the fine-tuned Sentence Transformer.
|
125 |
+
|
126 |
+
## Model Details
|
127 |
+
|
128 |
+
### Model Description
|
129 |
+
- **Model Type:** SetFit
|
130 |
+
- **Sentence Transformer body:** [BAAI/bge-base-en-v1.5](https://huggingface.co/BAAI/bge-base-en-v1.5)
|
131 |
+
- **Classification head:** a [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance
|
132 |
+
- **Maximum Sequence Length:** 512 tokens
|
133 |
+
- **Number of Classes:** 2 classes
|
134 |
+
<!-- - **Training Dataset:** [Unknown](https://huggingface.co/datasets/unknown) -->
|
135 |
+
<!-- - **Language:** Unknown -->
|
136 |
+
<!-- - **License:** Unknown -->
|
137 |
+
|
138 |
+
### Model Sources
|
139 |
+
|
140 |
+
- **Repository:** [SetFit on GitHub](https://github.com/huggingface/setfit)
|
141 |
+
- **Paper:** [Efficient Few-Shot Learning Without Prompts](https://arxiv.org/abs/2209.11055)
|
142 |
+
- **Blogpost:** [SetFit: Efficient Few-Shot Learning Without Prompts](https://huggingface.co/blog/setfit)
|
143 |
+
|
144 |
+
### Model Labels
|
145 |
+
| Label | Examples |
|
146 |
+
|:------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
|
147 |
+
| 1 | <ul><li>"The answer provided is comprehensive and directly addresses the question. Here is the reasoning:\n\n1. **Context Grounding:** The answer precisely matches the details provided in the document. Patricia Wallace's various roles, including managing a clothing closet, overseeing a food pantry, coordinating the food backpack program, and leading the Intervention Support Team, are well-supported by the text.\n \n2. **Relevance:** The answer is entirely relevant to the question, as it lists the specific roles and responsibilities of Patricia Wallace at Oak View Elementary as outlined in the document.\n\n3. **Conciseness:** The answer is clear and focused, listing the relevant roles and responsibilities without unnecessaryinformation.\n\nTherefore, the evaluation is:"</li><li>"Reasoning:\n1. Context Grounding: The answer is well supported by the document provided. It details the necessary steps to administer a saline solution to a baby, which is a method found within the source text.\n2. Relevance: The answer focuses specifically on treating a baby's cough, directly addressing the question asked.\n3. Conciseness: The answer is clear and to the point, providing concrete steps without unnecessary information. However, the answer could have been made even more concise by avoiding repetitions about the saline solution preparation.\n\nFinal evaluation:"</li><li>'Reasoning:\n1. Context Grounding: The answer provided accurately reflects the information in the document, describing the symptoms, risk factors, and necessary actions if toxic shock syndrome (TSS) is suspected.\n2. Relevance: The answer directly addresses the question asked, focusing on how to recognize TSS and what to do if you suspect you have it.\n3. Conciseness: The answer effectively condenses the necessary information into a coherent, straightforward explanation without extraneous details.\n\nFinal Evaluation:'</li></ul> |
|
148 |
+
| 0 | <ul><li>'Evaluation:\nThe answer provided incorrectly identifies the creation of a "literature hall" instead of a "science hall" as mentioned in the document. The answer also correctly attributes the oversight to Fr. Zahm, but this information is related to the wrong type of hall as per the document.\n\n1. Context Grounding: The document specifically states that a "Science Hall" was built under the direction of Fr. Zahm in 1883, not a literature hall.\n2. Relevance: The answer partially addresses the question correctly by mentioning Fr. Zahm, but it misidentifies the type of hall constructed.\n3. Conciseness: The answer is concise but includesincorrect information.\n\nThe final evaluation:'</li><li>'Reasoning:\n1. Context Grounding: The document supports that Gregory Johnson is the CEO of Franklin Templeton Investments and provides sufficient context about his role and relation to the company.\n2. Relevance: The answer directly addresses the question about the CEO of Franklin Templeton Investments.\n3. Conciseness: The answer presents the information clearly and succinctly without unnecessary details.\n\nFinal Result: ****'</li><li>'The answer correctly identifies that retired priests and brothers live at Fatima House. However, the additional information about the rare collection of ancient religious manuscripts at Fatima House is not supported by the document, making it an irrelevant addition. This deviates from the principle of conciseness and relevance to the specific question asked.\n\nFinal evaluation:'</li></ul> |
|
149 |
+
|
150 |
+
## Evaluation
|
151 |
+
|
152 |
+
### Metrics
|
153 |
+
| Label | Accuracy |
|
154 |
+
|:--------|:---------|
|
155 |
+
| **all** | 0.8649 |
|
156 |
+
|
157 |
+
## Uses
|
158 |
+
|
159 |
+
### Direct Use for Inference
|
160 |
+
|
161 |
+
First install the SetFit library:
|
162 |
+
|
163 |
+
```bash
|
164 |
+
pip install setfit
|
165 |
+
```
|
166 |
+
|
167 |
+
Then you can load this model and run inference.
|
168 |
+
|
169 |
+
```python
|
170 |
+
from setfit import SetFitModel
|
171 |
+
|
172 |
+
# Download from the 🤗 Hub
|
173 |
+
model = SetFitModel.from_pretrained("Netta1994/setfit_baai_wikisum_gpt-4o_cot-few_shot-instructions_remove_final_evaluation_e1_large")
|
174 |
+
# Run inference
|
175 |
+
preds = model("Reasoning:
|
176 |
+
1. Context Grounding: The provided answer reflects the key identifiers of a funnel spider found in the document, such as the dark brown or black body, hard shiny carapace, and large fangs.
|
177 |
+
2. Relevance: The answer directly addresses the question of how to identify a funnel spider with relevant details.
|
178 |
+
3. Conciseness: The answer is clear, with pertinent details, and avoids unnecessary information.
|
179 |
+
|
180 |
+
Final evaluation:")
|
181 |
+
```
|
182 |
+
|
183 |
+
<!--
|
184 |
+
### Downstream Use
|
185 |
+
|
186 |
+
*List how someone could finetune this model on their own dataset.*
|
187 |
+
-->
|
188 |
+
|
189 |
+
<!--
|
190 |
+
### Out-of-Scope Use
|
191 |
+
|
192 |
+
*List how the model may foreseeably be misused and address what users ought not to do with the model.*
|
193 |
+
-->
|
194 |
+
|
195 |
+
<!--
|
196 |
+
## Bias, Risks and Limitations
|
197 |
+
|
198 |
+
*What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
|
199 |
+
-->
|
200 |
+
|
201 |
+
<!--
|
202 |
+
### Recommendations
|
203 |
+
|
204 |
+
*What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
|
205 |
+
-->
|
206 |
+
|
207 |
+
## Training Details
|
208 |
+
|
209 |
+
### Training Set Metrics
|
210 |
+
| Training set | Min | Median | Max |
|
211 |
+
|:-------------|:----|:--------|:----|
|
212 |
+
| Word count | 11 | 76.2020 | 196 |
|
213 |
+
|
214 |
+
| Label | Training Sample Count |
|
215 |
+
|:------|:----------------------|
|
216 |
+
| 0 | 94 |
|
217 |
+
| 1 | 104 |
|
218 |
+
|
219 |
+
### Training Hyperparameters
|
220 |
+
- batch_size: (16, 16)
|
221 |
+
- num_epochs: (1, 1)
|
222 |
+
- max_steps: -1
|
223 |
+
- sampling_strategy: oversampling
|
224 |
+
- num_iterations: 20
|
225 |
+
- body_learning_rate: (2e-05, 2e-05)
|
226 |
+
- head_learning_rate: 2e-05
|
227 |
+
- loss: CosineSimilarityLoss
|
228 |
+
- distance_metric: cosine_distance
|
229 |
+
- margin: 0.25
|
230 |
+
- end_to_end: False
|
231 |
+
- use_amp: False
|
232 |
+
- warmup_proportion: 0.1
|
233 |
+
- l2_weight: 0.01
|
234 |
+
- seed: 42
|
235 |
+
- eval_max_steps: -1
|
236 |
+
- load_best_model_at_end: False
|
237 |
+
|
238 |
+
### Training Results
|
239 |
+
| Epoch | Step | Training Loss | Validation Loss |
|
240 |
+
|:------:|:----:|:-------------:|:---------------:|
|
241 |
+
| 0.0020 | 1 | 0.2119 | - |
|
242 |
+
| 0.1010 | 50 | 0.255 | - |
|
243 |
+
| 0.2020 | 100 | 0.1703 | - |
|
244 |
+
| 0.3030 | 150 | 0.0611 | - |
|
245 |
+
| 0.4040 | 200 | 0.0351 | - |
|
246 |
+
| 0.5051 | 250 | 0.0197 | - |
|
247 |
+
| 0.6061 | 300 | 0.0172 | - |
|
248 |
+
| 0.7071 | 350 | 0.0109 | - |
|
249 |
+
| 0.8081 | 400 | 0.0108 | - |
|
250 |
+
| 0.9091 | 450 | 0.0072 | - |
|
251 |
+
|
252 |
+
### Framework Versions
|
253 |
+
- Python: 3.10.14
|
254 |
+
- SetFit: 1.1.0
|
255 |
+
- Sentence Transformers: 3.1.1
|
256 |
+
- Transformers: 4.44.0
|
257 |
+
- PyTorch: 2.4.0+cu121
|
258 |
+
- Datasets: 3.0.0
|
259 |
+
- Tokenizers: 0.19.1
|
260 |
+
|
261 |
+
## Citation
|
262 |
+
|
263 |
+
### BibTeX
|
264 |
+
```bibtex
|
265 |
+
@article{https://doi.org/10.48550/arxiv.2209.11055,
|
266 |
+
doi = {10.48550/ARXIV.2209.11055},
|
267 |
+
url = {https://arxiv.org/abs/2209.11055},
|
268 |
+
author = {Tunstall, Lewis and Reimers, Nils and Jo, Unso Eun Seo and Bates, Luke and Korat, Daniel and Wasserblat, Moshe and Pereg, Oren},
|
269 |
+
keywords = {Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences},
|
270 |
+
title = {Efficient Few-Shot Learning Without Prompts},
|
271 |
+
publisher = {arXiv},
|
272 |
+
year = {2022},
|
273 |
+
copyright = {Creative Commons Attribution 4.0 International}
|
274 |
+
}
|
275 |
+
```
|
276 |
+
|
277 |
+
<!--
|
278 |
+
## Glossary
|
279 |
+
|
280 |
+
*Clearly define terms in order to be accessible across audiences.*
|
281 |
+
-->
|
282 |
+
|
283 |
+
<!--
|
284 |
+
## Model Card Authors
|
285 |
+
|
286 |
+
*Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
|
287 |
+
-->
|
288 |
+
|
289 |
+
<!--
|
290 |
+
## Model Card Contact
|
291 |
+
|
292 |
+
*Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
|
293 |
+
-->
|
config.json
ADDED
@@ -0,0 +1,32 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "BAAI/bge-base-en-v1.5",
|
3 |
+
"architectures": [
|
4 |
+
"BertModel"
|
5 |
+
],
|
6 |
+
"attention_probs_dropout_prob": 0.1,
|
7 |
+
"classifier_dropout": null,
|
8 |
+
"gradient_checkpointing": false,
|
9 |
+
"hidden_act": "gelu",
|
10 |
+
"hidden_dropout_prob": 0.1,
|
11 |
+
"hidden_size": 768,
|
12 |
+
"id2label": {
|
13 |
+
"0": "LABEL_0"
|
14 |
+
},
|
15 |
+
"initializer_range": 0.02,
|
16 |
+
"intermediate_size": 3072,
|
17 |
+
"label2id": {
|
18 |
+
"LABEL_0": 0
|
19 |
+
},
|
20 |
+
"layer_norm_eps": 1e-12,
|
21 |
+
"max_position_embeddings": 512,
|
22 |
+
"model_type": "bert",
|
23 |
+
"num_attention_heads": 12,
|
24 |
+
"num_hidden_layers": 12,
|
25 |
+
"pad_token_id": 0,
|
26 |
+
"position_embedding_type": "absolute",
|
27 |
+
"torch_dtype": "float32",
|
28 |
+
"transformers_version": "4.44.0",
|
29 |
+
"type_vocab_size": 2,
|
30 |
+
"use_cache": true,
|
31 |
+
"vocab_size": 30522
|
32 |
+
}
|
config_sentence_transformers.json
ADDED
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"__version__": {
|
3 |
+
"sentence_transformers": "3.1.1",
|
4 |
+
"transformers": "4.44.0",
|
5 |
+
"pytorch": "2.4.0+cu121"
|
6 |
+
},
|
7 |
+
"prompts": {},
|
8 |
+
"default_prompt_name": null,
|
9 |
+
"similarity_fn_name": null
|
10 |
+
}
|
config_setfit.json
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"normalize_embeddings": false,
|
3 |
+
"labels": null
|
4 |
+
}
|
model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4fc11758b8fdb64cf479832f3e59f5cb86764aa957e5c8560f706f9c9ca7eb8e
|
3 |
+
size 437951328
|
model_head.pkl
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a294344b9f3b56fd99a4984faf66fe39b0fb64afc86be02facb476f4320b6622
|
3 |
+
size 7007
|
modules.json
ADDED
@@ -0,0 +1,20 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
[
|
2 |
+
{
|
3 |
+
"idx": 0,
|
4 |
+
"name": "0",
|
5 |
+
"path": "",
|
6 |
+
"type": "sentence_transformers.models.Transformer"
|
7 |
+
},
|
8 |
+
{
|
9 |
+
"idx": 1,
|
10 |
+
"name": "1",
|
11 |
+
"path": "1_Pooling",
|
12 |
+
"type": "sentence_transformers.models.Pooling"
|
13 |
+
},
|
14 |
+
{
|
15 |
+
"idx": 2,
|
16 |
+
"name": "2",
|
17 |
+
"path": "2_Normalize",
|
18 |
+
"type": "sentence_transformers.models.Normalize"
|
19 |
+
}
|
20 |
+
]
|
sentence_bert_config.json
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"max_seq_length": 512,
|
3 |
+
"do_lower_case": true
|
4 |
+
}
|
special_tokens_map.json
ADDED
@@ -0,0 +1,37 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"cls_token": {
|
3 |
+
"content": "[CLS]",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": false,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"mask_token": {
|
10 |
+
"content": "[MASK]",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": false,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"pad_token": {
|
17 |
+
"content": "[PAD]",
|
18 |
+
"lstrip": false,
|
19 |
+
"normalized": false,
|
20 |
+
"rstrip": false,
|
21 |
+
"single_word": false
|
22 |
+
},
|
23 |
+
"sep_token": {
|
24 |
+
"content": "[SEP]",
|
25 |
+
"lstrip": false,
|
26 |
+
"normalized": false,
|
27 |
+
"rstrip": false,
|
28 |
+
"single_word": false
|
29 |
+
},
|
30 |
+
"unk_token": {
|
31 |
+
"content": "[UNK]",
|
32 |
+
"lstrip": false,
|
33 |
+
"normalized": false,
|
34 |
+
"rstrip": false,
|
35 |
+
"single_word": false
|
36 |
+
}
|
37 |
+
}
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
ADDED
@@ -0,0 +1,57 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"added_tokens_decoder": {
|
3 |
+
"0": {
|
4 |
+
"content": "[PAD]",
|
5 |
+
"lstrip": false,
|
6 |
+
"normalized": false,
|
7 |
+
"rstrip": false,
|
8 |
+
"single_word": false,
|
9 |
+
"special": true
|
10 |
+
},
|
11 |
+
"100": {
|
12 |
+
"content": "[UNK]",
|
13 |
+
"lstrip": false,
|
14 |
+
"normalized": false,
|
15 |
+
"rstrip": false,
|
16 |
+
"single_word": false,
|
17 |
+
"special": true
|
18 |
+
},
|
19 |
+
"101": {
|
20 |
+
"content": "[CLS]",
|
21 |
+
"lstrip": false,
|
22 |
+
"normalized": false,
|
23 |
+
"rstrip": false,
|
24 |
+
"single_word": false,
|
25 |
+
"special": true
|
26 |
+
},
|
27 |
+
"102": {
|
28 |
+
"content": "[SEP]",
|
29 |
+
"lstrip": false,
|
30 |
+
"normalized": false,
|
31 |
+
"rstrip": false,
|
32 |
+
"single_word": false,
|
33 |
+
"special": true
|
34 |
+
},
|
35 |
+
"103": {
|
36 |
+
"content": "[MASK]",
|
37 |
+
"lstrip": false,
|
38 |
+
"normalized": false,
|
39 |
+
"rstrip": false,
|
40 |
+
"single_word": false,
|
41 |
+
"special": true
|
42 |
+
}
|
43 |
+
},
|
44 |
+
"clean_up_tokenization_spaces": true,
|
45 |
+
"cls_token": "[CLS]",
|
46 |
+
"do_basic_tokenize": true,
|
47 |
+
"do_lower_case": true,
|
48 |
+
"mask_token": "[MASK]",
|
49 |
+
"model_max_length": 512,
|
50 |
+
"never_split": null,
|
51 |
+
"pad_token": "[PAD]",
|
52 |
+
"sep_token": "[SEP]",
|
53 |
+
"strip_accents": null,
|
54 |
+
"tokenize_chinese_chars": true,
|
55 |
+
"tokenizer_class": "BertTokenizer",
|
56 |
+
"unk_token": "[UNK]"
|
57 |
+
}
|
vocab.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|