Add new SentenceTransformer model.
Browse files- .gitattributes +2 -0
- 1_Pooling/config.json +7 -0
- README.md +125 -0
- config.json +29 -0
- config_sentence_transformers.json +7 -0
- eval/similarity_evaluation_results.csv +121 -0
- modules.json +14 -0
- pytorch_model.bin +3 -0
- sentence_bert_config.json +4 -0
- sentencepiece.bpe.model +3 -0
- similarity_evaluation_sts-test_results.csv +2 -0
- special_tokens_map.json +1 -0
- tokenizer.json +3 -0
- tokenizer_config.json +1 -0
.gitattributes
CHANGED
@@ -25,3 +25,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
25 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
26 |
*.zstandard filter=lfs diff=lfs merge=lfs -text
|
27 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
25 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
26 |
*.zstandard filter=lfs diff=lfs merge=lfs -text
|
27 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
28 |
+
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
29 |
+
pytorch_model.bin filter=lfs diff=lfs merge=lfs -text
|
1_Pooling/config.json
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"word_embedding_dimension": 768,
|
3 |
+
"pooling_mode_cls_token": false,
|
4 |
+
"pooling_mode_mean_tokens": true,
|
5 |
+
"pooling_mode_max_tokens": false,
|
6 |
+
"pooling_mode_mean_sqrt_len_tokens": false
|
7 |
+
}
|
README.md
ADDED
@@ -0,0 +1,125 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
pipeline_tag: sentence-similarity
|
3 |
+
tags:
|
4 |
+
- sentence-transformers
|
5 |
+
- feature-extraction
|
6 |
+
- sentence-similarity
|
7 |
+
- transformers
|
8 |
+
---
|
9 |
+
|
10 |
+
# {MODEL_NAME}
|
11 |
+
|
12 |
+
This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.
|
13 |
+
|
14 |
+
<!--- Describe your model here -->
|
15 |
+
|
16 |
+
## Usage (Sentence-Transformers)
|
17 |
+
|
18 |
+
Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed:
|
19 |
+
|
20 |
+
```
|
21 |
+
pip install -U sentence-transformers
|
22 |
+
```
|
23 |
+
|
24 |
+
Then you can use the model like this:
|
25 |
+
|
26 |
+
```python
|
27 |
+
from sentence_transformers import SentenceTransformer
|
28 |
+
sentences = ["This is an example sentence", "Each sentence is converted"]
|
29 |
+
|
30 |
+
model = SentenceTransformer('{MODEL_NAME}')
|
31 |
+
embeddings = model.encode(sentences)
|
32 |
+
print(embeddings)
|
33 |
+
```
|
34 |
+
|
35 |
+
|
36 |
+
|
37 |
+
## Usage (HuggingFace Transformers)
|
38 |
+
Without [sentence-transformers](https://www.SBERT.net), you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.
|
39 |
+
|
40 |
+
```python
|
41 |
+
from transformers import AutoTokenizer, AutoModel
|
42 |
+
import torch
|
43 |
+
|
44 |
+
|
45 |
+
#Mean Pooling - Take attention mask into account for correct averaging
|
46 |
+
def mean_pooling(model_output, attention_mask):
|
47 |
+
token_embeddings = model_output[0] #First element of model_output contains all token embeddings
|
48 |
+
input_mask_expanded = attention_mask.unsqueeze(-1).expand(token_embeddings.size()).float()
|
49 |
+
return torch.sum(token_embeddings * input_mask_expanded, 1) / torch.clamp(input_mask_expanded.sum(1), min=1e-9)
|
50 |
+
|
51 |
+
|
52 |
+
# Sentences we want sentence embeddings for
|
53 |
+
sentences = ['This is an example sentence', 'Each sentence is converted']
|
54 |
+
|
55 |
+
# Load model from HuggingFace Hub
|
56 |
+
tokenizer = AutoTokenizer.from_pretrained('{MODEL_NAME}')
|
57 |
+
model = AutoModel.from_pretrained('{MODEL_NAME}')
|
58 |
+
|
59 |
+
# Tokenize sentences
|
60 |
+
encoded_input = tokenizer(sentences, padding=True, truncation=True, return_tensors='pt')
|
61 |
+
|
62 |
+
# Compute token embeddings
|
63 |
+
with torch.no_grad():
|
64 |
+
model_output = model(**encoded_input)
|
65 |
+
|
66 |
+
# Perform pooling. In this case, mean pooling.
|
67 |
+
sentence_embeddings = mean_pooling(model_output, encoded_input['attention_mask'])
|
68 |
+
|
69 |
+
print("Sentence embeddings:")
|
70 |
+
print(sentence_embeddings)
|
71 |
+
```
|
72 |
+
|
73 |
+
|
74 |
+
|
75 |
+
## Evaluation Results
|
76 |
+
|
77 |
+
<!--- Describe how your model was evaluated -->
|
78 |
+
|
79 |
+
For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name={MODEL_NAME})
|
80 |
+
|
81 |
+
|
82 |
+
## Training
|
83 |
+
The model was trained with the parameters:
|
84 |
+
|
85 |
+
**DataLoader**:
|
86 |
+
|
87 |
+
`torch.utils.data.dataloader.DataLoader` of length 11 with parameters:
|
88 |
+
```
|
89 |
+
{'batch_size': 15, 'sampler': 'torch.utils.data.sampler.RandomSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'}
|
90 |
+
```
|
91 |
+
|
92 |
+
**Loss**:
|
93 |
+
|
94 |
+
`sentence_transformers.losses.CosineSimilarityLoss.CosineSimilarityLoss`
|
95 |
+
|
96 |
+
Parameters of the fit()-Method:
|
97 |
+
```
|
98 |
+
{
|
99 |
+
"epochs": 10,
|
100 |
+
"evaluation_steps": 1,
|
101 |
+
"evaluator": "sentence_transformers.evaluation.EmbeddingSimilarityEvaluator.EmbeddingSimilarityEvaluator",
|
102 |
+
"max_grad_norm": 1,
|
103 |
+
"optimizer_class": "<class 'transformers.optimization.AdamW'>",
|
104 |
+
"optimizer_params": {
|
105 |
+
"lr": 2e-05
|
106 |
+
},
|
107 |
+
"scheduler": "WarmupLinear",
|
108 |
+
"steps_per_epoch": null,
|
109 |
+
"warmup_steps": 11,
|
110 |
+
"weight_decay": 0.01
|
111 |
+
}
|
112 |
+
```
|
113 |
+
|
114 |
+
|
115 |
+
## Full Model Architecture
|
116 |
+
```
|
117 |
+
SentenceTransformer(
|
118 |
+
(0): Transformer({'max_seq_length': 128, 'do_lower_case': False}) with Transformer model: XLMRobertaModel
|
119 |
+
(1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False})
|
120 |
+
)
|
121 |
+
```
|
122 |
+
|
123 |
+
## Citing & Authors
|
124 |
+
|
125 |
+
<!--- Describe where people can find more information -->
|
config.json
ADDED
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "/root/.cache/torch/sentence_transformers/sentence-transformers_paraphrase-multilingual-mpnet-base-v2/",
|
3 |
+
"architectures": [
|
4 |
+
"XLMRobertaModel"
|
5 |
+
],
|
6 |
+
"attention_probs_dropout_prob": 0.1,
|
7 |
+
"bos_token_id": 0,
|
8 |
+
"classifier_dropout": null,
|
9 |
+
"eos_token_id": 2,
|
10 |
+
"gradient_checkpointing": false,
|
11 |
+
"hidden_act": "gelu",
|
12 |
+
"hidden_dropout_prob": 0.1,
|
13 |
+
"hidden_size": 768,
|
14 |
+
"initializer_range": 0.02,
|
15 |
+
"intermediate_size": 3072,
|
16 |
+
"layer_norm_eps": 1e-05,
|
17 |
+
"max_position_embeddings": 514,
|
18 |
+
"model_type": "xlm-roberta",
|
19 |
+
"num_attention_heads": 12,
|
20 |
+
"num_hidden_layers": 12,
|
21 |
+
"output_past": true,
|
22 |
+
"pad_token_id": 1,
|
23 |
+
"position_embedding_type": "absolute",
|
24 |
+
"torch_dtype": "float32",
|
25 |
+
"transformers_version": "4.16.2",
|
26 |
+
"type_vocab_size": 1,
|
27 |
+
"use_cache": true,
|
28 |
+
"vocab_size": 250002
|
29 |
+
}
|
config_sentence_transformers.json
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"__version__": {
|
3 |
+
"sentence_transformers": "2.0.0",
|
4 |
+
"transformers": "4.7.0",
|
5 |
+
"pytorch": "1.9.0+cu102"
|
6 |
+
}
|
7 |
+
}
|
eval/similarity_evaluation_results.csv
ADDED
@@ -0,0 +1,121 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
epoch,steps,cosine_pearson,cosine_spearman,euclidean_pearson,euclidean_spearman,manhattan_pearson,manhattan_spearman,dot_pearson,dot_spearman
|
2 |
+
0,1,0.24428484492337774,0.1856406898421688,0.2726824721314313,0.17419708567381592,0.20191195613958324,0.18182615511938452,-0.10491629431153585,-0.05976104399028721
|
3 |
+
0,2,0.22612378962507065,0.1474953426143259,0.2560357877483311,0.17928313197086163,0.17884221008459156,0.13732325002023446,-0.11666791021706803,-0.07374767130716295
|
4 |
+
0,3,0.19677200683261806,0.10934999538648299,0.23751600007738927,0.16529650465398593,0.15040892380050047,0.01398662731687573,-0.15956732474655524,-0.11570755325779014
|
5 |
+
0,4,0.11503452111144055,0.025430231485228605,0.15761318436439234,0.01907267361392145,0.058673887759007365,-0.022887208336705742,-0.18204305619588393,-0.21361394447592028
|
6 |
+
0,5,0.003985376234194979,-0.005086046297045721,0.03632262297764966,0.012715115742614302,-0.05827011779841074,-0.12460813427762014,-0.1913564390855504,-0.25048778012950174
|
7 |
+
0,6,-0.05103630332147053,-0.026701743059490034,-0.03439079582489007,-0.10807848381222157,-0.11438987777919786,-0.1296941805746659,-0.16467537031081167,-0.2492162685552403
|
8 |
+
0,7,-0.09921001875664615,-0.13986627316875733,-0.09075888010265239,-0.11952208798057443,-0.1596978337595956,-0.19199824771347596,-0.14141547945028185,-0.2835470810602989
|
9 |
+
0,8,-0.13983443384729216,-0.21234243290165886,-0.14450538879193142,-0.18182615511938452,-0.20808728093382015,-0.2492162685552403,-0.12541343825340248,-0.25048778012950174
|
10 |
+
0,9,-0.15921298240054033,-0.23777266438688743,-0.1797885889172838,-0.1907267361392145,-0.24392530175624805,-0.2492162685552403,-0.1048356212094887,-0.17038255095103164
|
11 |
+
0,10,-0.1609468704741812,-0.20471336345609026,-0.19744846790394677,-0.22124301392148887,-0.26585823378739704,-0.2657459190206389,-0.04824870832250316,-0.11443604168352871
|
12 |
+
0,11,-0.15410333770549933,-0.18436917826790739,-0.2093702035319172,-0.16021045835694017,-0.2808278669531017,-0.3038912662484818,0.046277154708522575,0.026701743059490034
|
13 |
+
0,-1,-0.15410333770549933,-0.18436917826790739,-0.2093702035319172,-0.16021045835694017,-0.2808278669531017,-0.3038912662484818,0.046277154708522575,0.026701743059490034
|
14 |
+
1,1,-0.12982786622499196,-0.15003836576284876,-0.21235892958369582,-0.1983558055847831,-0.2909451954930539,-0.25557382642654747,0.1675258309470891,0.24794475698097887
|
15 |
+
1,2,-0.19696595634631142,-0.16656801622824735,-0.25668213221583064,-0.2326866180898417,-0.3250005522733054,-0.3801819607041676,0.04540854622383069,-0.00381453472278429
|
16 |
+
1,3,-0.21781655597904578,-0.27591801161473034,-0.2747155636631352,-0.32296393986240324,-0.33711431182747686,-0.37509591440712187,-0.015301510197506163,-0.0673901134358558
|
17 |
+
1,4,-0.21040566408092895,-0.20725638660461312,-0.2787115989564245,-0.31406335884257325,-0.33960265883733304,-0.32804998615944897,0.05601493400121786,0.01525813889113716
|
18 |
+
1,5,-0.21075832691211716,-0.2492162685552403,-0.2826518195170755,-0.3356790556050176,-0.339197989506619,-0.35093719449615474,0.07529820685804627,0.057218020841764354
|
19 |
+
1,6,-0.20609520571963286,-0.27210347689194603,-0.27951092027179186,-0.35856626394172325,-0.33310429600778424,-0.37255289125859903,0.11402579008036735,0.11570755325779014
|
20 |
+
1,7,-0.169486532878668,-0.2708319653176846,-0.26423165753461786,-0.3344075440307561,-0.31980430200287796,-0.36110928709024614,0.20002633803728215,0.3115203356940504
|
21 |
+
1,8,-0.0901707317417178,-0.14113778474301875,-0.2177628930192162,-0.2810040579117761,-0.2751703064164917,-0.2988052199514361,0.2490250330297735,0.40179765746661195
|
22 |
+
1,9,-0.004556612508979109,0.16402499307972448,-0.14854868139699784,-0.22124301392148887,-0.20540087365227688,-0.23650115281262601,0.2707949105636119,0.4195988195062719
|
23 |
+
1,10,0.0677969000349118,0.32296393986240324,-0.05310832958181209,0.00381453472278429,-0.09639636112386697,-0.01652965046539859,0.2734802712416607,0.40179765746661195
|
24 |
+
1,11,0.11315696215310725,0.3064342893970047,0.029403827430142117,0.17419708567381592,0.0028628903329707502,0.16529650465398593,0.27353726982326,0.39035405329825906
|
25 |
+
1,-1,0.11315696215310725,0.3064342893970047,0.029403827430142117,0.17419708567381592,0.0028628903329707502,0.16529650465398593,0.27353726982326,0.39035405329825906
|
26 |
+
2,1,0.13940714819785371,0.3293214977337104,0.08566760205046141,0.25557382642654747,0.06824408200836007,0.21234243290165886,0.2712000137342313,0.39035405329825906
|
27 |
+
2,2,0.14699598094492095,0.37255289125859903,0.10174450966560532,0.30134824309995895,0.08663053449180397,0.2606598727235932,0.27337148870045525,0.39035405329825906
|
28 |
+
2,3,0.14656120764881206,0.32042091671388034,0.10977333616234204,0.29244766208012896,0.0968236329054813,0.20471336345609026,0.2728593350922896,0.39035405329825906
|
29 |
+
2,4,0.13675405809671398,0.4056121921893962,0.10241204205207184,0.29244766208012896,0.09005215314493258,0.2568453380008089,0.27303591622071344,0.3941685880210433
|
30 |
+
2,5,0.13799404823353287,0.3674668449615533,0.10605153984267622,0.2708319653176846,0.09628586334092931,0.23777266438688743,0.2774065715793497,0.381453472278429
|
31 |
+
2,6,0.11872275011544195,0.3318645208822333,0.07989894345468976,0.1907267361392145,0.06894963431264978,0.1907267361392145,0.2772175356157276,0.4208703310805334
|
32 |
+
2,7,0.10777041689179366,0.3038912662484818,0.06679490929721191,0.21742847919870456,0.05639645529687149,0.15893894678267878,0.27587894919073314,0.38526800700121333
|
33 |
+
2,8,0.08964774465066375,0.25048778012950174,0.04562989321901026,0.17419708567381592,0.03533869876279845,0.14495231946580303,0.2695904387278222,0.3700098681100762
|
34 |
+
2,9,0.07956450993625785,0.22887208336705742,0.03178353953441007,0.09790639121813012,0.020992268554907812,0.11825057640631301,0.2681886742472172,0.3827249838526905
|
35 |
+
2,10,0.06530632310909693,0.23904417596114885,0.01290320159856114,0.07883371760420867,-0.0008022100415914147,0.07883371760420867,0.2652623096945603,0.39925463431808905
|
36 |
+
2,11,0.0541526868004497,0.20598487503035168,-0.0014604552039510635,0.05340348611898007,-0.020671016000849286,0.045774416673411485,0.26474831609084454,0.3661953333872919
|
37 |
+
2,-1,0.0541526868004497,0.20598487503035168,-0.0014604552039510635,0.05340348611898007,-0.020671016000849286,0.045774416673411485,0.26474831609084454,0.3661953333872919
|
38 |
+
3,1,0.04082352652346983,0.1780116203966002,-0.016950879494709237,0.024158719910967172,-0.04256928272155714,-0.024158719910967172,0.26136168380031144,0.3064342893970047
|
39 |
+
3,2,0.039171457779169055,0.18309766669364594,-0.022019282369464054,0.005086046297045721,-0.0534613577807422,-0.06230406713881007,0.2626351974962769,0.33059300930797186
|
40 |
+
3,3,0.03962702632659584,0.21615696762444314,-0.021501770740796965,-0.01780116203966002,-0.05837792754185962,-0.024158719910967172,0.2610292488191349,0.35220870607041616
|
41 |
+
3,4,0.05138318751675695,0.19581278243626024,-0.009772229122192283,0.048317439821934344,-0.05131522224080271,0.006357557871307151,0.26451055406709656,0.36110928709024614
|
42 |
+
3,5,0.042919953964449316,0.18055464354512307,-0.017252621913390217,0.07374767130716295,-0.06272951255774184,-0.025430231485228605,0.25893214830143135,0.3293214977337104
|
43 |
+
3,6,0.01621406176307976,0.14113778474301875,-0.044121003338826015,-0.052131974544718636,-0.09192335750297764,-0.0673901134358558,0.24414022952783723,0.28609010420882175
|
44 |
+
3,7,0.01603013319187219,0.11952208798057443,-0.044933401380679173,-0.04450290509915005,-0.09369405138432947,-0.09663487964386869,0.2435001866086313,0.28609010420882175
|
45 |
+
3,8,0.041056573181203274,0.17419708567381592,-0.02521008601539706,0.01652965046539859,-0.07780417603142906,-0.012715115742614302,0.25960723616492987,0.2822755694860375
|
46 |
+
3,9,0.06668407275286904,0.20852789817887454,-0.004198856765651891,0.0928203449210844,-0.05824260675806149,0.026701743059490034,0.27463594492277404,0.354751729218939
|
47 |
+
3,10,0.09732084786674933,0.17165406252529306,0.024355250848309237,0.09917790279239155,-0.031437585110709654,0.09154883334682297,0.287547698402672,0.4412145162687163
|
48 |
+
3,11,0.09969461581415269,0.20471336345609026,0.02412155708892016,0.09154883334682297,-0.030725799309395584,0.12715115742614302,0.29083133024082536,0.49080346766491206
|
49 |
+
3,-1,0.09969461581415269,0.20471336345609026,0.02412155708892016,0.09154883334682297,-0.030725799309395584,0.12715115742614302,0.29083133024082536,0.49080346766491206
|
50 |
+
4,1,0.09688106671926407,0.20471336345609026,0.019849707506481476,0.15639592363415591,-0.033138832196187684,0.16402499307972448,0.29020825241438647,0.43739998154593196
|
51 |
+
4,2,0.10686210835084084,0.227600571792796,0.03527764918280338,0.21107092132739738,-0.014809215605804057,0.14495231946580303,0.2871024991131068,0.4437575394172391
|
52 |
+
4,3,0.11528029125188341,0.227600571792796,0.046012326958320846,0.16529650465398593,-0.0014913913693886001,0.21361394447592028,0.28614822805844137,0.43867149312019343
|
53 |
+
4,4,0.11327236066146648,0.2708319653176846,0.04816513612968369,0.21615696762444314,0.002628733543029182,0.21361394447592028,0.28244052688585697,0.43867149312019343
|
54 |
+
4,5,0.1097521443719211,0.2962621968029132,0.053605501185894744,0.23777266438688743,0.011076965089880491,0.2352296412383646,0.274889620003233,0.44248602784297775
|
55 |
+
4,6,0.085118512450099,0.24540173383245603,0.035052362001546225,0.20471336345609026,-0.004599692497119436,0.18055464354512307,0.2614755796846744,0.4094267269121805
|
56 |
+
4,7,0.06647371176348317,0.227600571792796,0.022546934649379773,0.20471336345609026,-0.01489167340998393,0.1678395278025088,0.24772311028462143,0.4310424236746248
|
57 |
+
4,8,0.058433927289096785,0.227600571792796,0.020394938037328333,0.18691220141643022,-0.01488648898505835,0.1347802268717116,0.238793669664602,0.4094267269121805
|
58 |
+
4,9,0.06218765255951492,0.2606598727235932,0.028254863204573506,0.18691220141643022,-0.0059663318280807195,0.1347802268717116,0.23662507196670224,0.4310424236746248
|
59 |
+
4,10,0.05946027313650021,0.19454127086199882,0.026231299009535372,0.18691220141643022,-0.007429255669431191,0.12333662270335873,0.23451281524353298,0.381453472278429
|
60 |
+
4,11,0.04488249947205441,0.23777266438688743,0.016194060108917257,0.19326975928773737,-0.014837134690265971,0.14622383104006448,0.22440272069256118,0.3661953333872919
|
61 |
+
4,-1,0.04488249947205441,0.23777266438688743,0.016194060108917257,0.19326975928773737,-0.014837134690265971,0.14622383104006448,0.22440272069256118,0.3661953333872919
|
62 |
+
5,1,0.053395547312709554,0.2326866180898417,0.025551009989655416,0.20471336345609026,-0.0030336681748221674,0.16021045835694017,0.23046363705984724,0.3878110301497362
|
63 |
+
5,2,0.0591569326701527,0.23777266438688743,0.03128838249262078,0.20471336345609026,0.0043441502556022605,0.17165406252529306,0.23440759776991563,0.3789104491299062
|
64 |
+
5,3,0.06450234649256706,0.23777266438688743,0.03449441250539357,0.18055464354512307,0.008766844563450683,0.20725638660461312,0.24108678497623875,0.3789104491299062
|
65 |
+
5,4,0.06214173452043999,0.23777266438688743,0.032097244328093966,0.16656801622824735,0.008197085492201326,0.18309766669364594,0.2426413857518142,0.3878110301497362
|
66 |
+
5,5,0.06578045671422762,0.23777266438688743,0.03542313897925867,0.15639592363415591,0.012700649092421475,0.18309766669364594,0.24607156573013772,0.4094267269121805
|
67 |
+
5,6,0.08462140876157086,0.2708319653176846,0.0522684701873643,0.1347802268717116,0.030843014529425592,0.1729255740995545,0.25759507497226136,0.3789104491299062
|
68 |
+
5,7,0.0990143040299816,0.23904417596114885,0.06462391531700609,0.18436917826790739,0.043981507888789885,0.21615696762444314,0.2652275602224404,0.38908254172399764
|
69 |
+
5,8,0.11663814351298049,0.23395812966410312,0.08161265028948292,0.20852789817887454,0.061440389630868536,0.19454127086199882,0.27160989904850547,0.4170557963577491
|
70 |
+
5,9,0.12327135444243054,0.20979940975313596,0.08851227558689899,0.19454127086199882,0.06924274661872722,0.227600571792796,0.2735691512978767,0.4348569583974091
|
71 |
+
5,10,0.1256530678741609,0.20979940975313596,0.09225233834877479,0.19454127086199882,0.07268984121640336,0.2708319653176846,0.2742497099134256,0.4348569583974091
|
72 |
+
5,11,0.12873361419741985,0.23141510651558028,0.09823515122296156,0.227600571792796,0.07889129768310928,0.2708319653176846,0.27363372133242203,0.4348569583974091
|
73 |
+
5,-1,0.12873361419741985,0.23141510651558028,0.09823515122296156,0.227600571792796,0.07889129768310928,0.2708319653176846,0.27363372133242203,0.4348569583974091
|
74 |
+
6,1,0.12567595694042172,0.23141510651558028,0.09857514532217378,0.24667324540671745,0.07986965314509253,0.2708319653176846,0.27056410374820217,0.4348569583974091
|
75 |
+
6,2,0.12149628573204947,0.25430231485228605,0.0961034528372844,0.27210347689194603,0.07726412770312542,0.2530308032780246,0.26964215394565244,0.4170557963577491
|
76 |
+
6,3,0.10559031426059957,0.26447440744637746,0.08239293894232613,0.24794475698097887,0.06377960458713212,0.27591801161473034,0.26486744741490464,0.4399430046944548
|
77 |
+
6,4,0.10259542528560504,0.26447440744637746,0.08061103179553537,0.24794475698097887,0.06204550170088915,0.2657459190206389,0.26434917766482163,0.4399430046944548
|
78 |
+
6,5,0.089628434872583,0.26447440744637746,0.06910921853351924,0.24158719910967172,0.050042422794654796,0.24158719910967172,0.26022687306020714,0.41196975006070335
|
79 |
+
6,6,0.06903104253323819,0.2657459190206389,0.051863084231036784,0.17546859724807737,0.03259363057545481,0.17928313197086163,0.2514680535201101,0.3878110301497362
|
80 |
+
6,7,0.06627359161868877,0.28863312735734464,0.04856471876248964,0.17546859724807737,0.0290375512498293,0.18055464354512307,0.2510346416505546,0.3878110301497362
|
81 |
+
6,8,0.06587585027778702,0.28863312735734464,0.04744858656278781,0.18691220141643022,0.027655219369644334,0.18055464354512307,0.25162183829556445,0.3916255648725205
|
82 |
+
6,9,0.07667645472428877,0.28609010420882175,0.05489932876112217,0.2199715023472274,0.03397228404864987,0.21361394447592028,0.25906650914738716,0.3916255648725205
|
83 |
+
6,10,0.08455113989210214,0.30134824309995895,0.06023682494544593,0.24158719910967172,0.03833933740430189,0.2568453380008089,0.26322096161083997,0.3916255648725205
|
84 |
+
6,11,0.06527064441009967,0.28609010420882175,0.04645573451990924,0.2199715023472274,0.023629072577399244,0.24667324540671745,0.25261854507949266,0.3916255648725205
|
85 |
+
6,-1,0.06527064441009967,0.28609010420882175,0.04645573451990924,0.2199715023472274,0.023629072577399244,0.24667324540671745,0.25261854507949266,0.3916255648725205
|
86 |
+
7,1,0.060974209679639874,0.2962621968029132,0.044182665431398774,0.2199715023472274,0.020208359608723723,0.2568453380008089,0.2500246080799369,0.3916255648725205
|
87 |
+
7,2,0.05889986791654363,0.2962621968029132,0.04367928785660007,0.20852789817887454,0.018990969899362664,0.2568453380008089,0.24852263732052335,0.3801819607041676
|
88 |
+
7,3,0.0586805419019325,0.2962621968029132,0.04510026160066266,0.20852789817887454,0.01947097540749484,0.24158719910967172,0.24833599501031361,0.3801819607041676
|
89 |
+
7,4,0.06086112074593676,0.3064342893970047,0.04777464187251144,0.24158719910967172,0.020697437016754666,0.24158719910967172,0.2494308562127234,0.3801819607041676
|
90 |
+
7,5,0.06280098215237057,0.3064342893970047,0.05051831916576996,0.24158719910967172,0.02218930339762014,0.24158719910967172,0.24971834944048754,0.39035405329825906
|
91 |
+
7,6,0.06158031126576389,0.3064342893970047,0.05075105683420316,0.24158719910967172,0.02159771518228107,0.20852789817887454,0.24800288493234027,0.39035405329825906
|
92 |
+
7,7,0.06397984201251046,0.3064342893970047,0.05426642849551435,0.24158719910967172,0.024405314267635883,0.21869999077296598,0.24762507347096327,0.39035405329825906
|
93 |
+
7,8,0.0659829841222002,0.30134824309995895,0.05731078982845222,0.2746465000404689,0.027236857467774822,0.21869999077296598,0.24696902815106028,0.41196975006070335
|
94 |
+
7,9,0.07164671048348455,0.30134824309995895,0.06349894299155719,0.25175929170376316,0.03336710644089848,0.27591801161473034,0.2474211727149824,0.41196975006070335
|
95 |
+
7,10,0.08963448736837631,0.3038912662484818,0.07951013636920258,0.27591801161473034,0.04793689428134984,0.2530308032780246,0.25395789779269695,0.40306916904087337
|
96 |
+
7,11,0.09347383398858716,0.2949906852286518,0.08502494874271449,0.27591801161473034,0.05304476405077205,0.2733749884662075,0.25375858054851336,0.40306916904087337
|
97 |
+
7,-1,0.09347383398858716,0.2949906852286518,0.08502494874271449,0.27591801161473034,0.05304476405077205,0.2733749884662075,0.25375858054851336,0.40306916904087337
|
98 |
+
8,1,0.09975180365559692,0.2733749884662075,0.09272251171908412,0.27591801161473034,0.06045198277887114,0.2733749884662075,0.25493848861839186,0.40306916904087337
|
99 |
+
8,2,0.10048681631029896,0.2949906852286518,0.09485771368118302,0.27591801161473034,0.061916887798136984,0.2733749884662075,0.2544032330293862,0.40306916904087337
|
100 |
+
8,3,0.0828267554578788,0.3038912662484818,0.08022802667414386,0.27591801161473034,0.046963894195038346,0.27591801161473034,0.2465196613596623,0.4056121921893962
|
101 |
+
8,4,0.059137661106999834,0.3064342893970047,0.06073868996694734,0.2988052199514361,0.02847229929037812,0.2657459190206389,0.2342259758391944,0.3661953333872919
|
102 |
+
8,5,0.037048084368087424,0.3089773125455275,0.04264002039285007,0.24158719910967172,0.011641961257956638,0.20852789817887454,0.22059703173457557,0.3661953333872919
|
103 |
+
8,6,0.013393957920069248,0.28863312735734464,0.023186005391405987,0.18309766669364594,-0.006231614744725678,0.15003836576284876,0.2034548517038958,0.33695056717927896
|
104 |
+
8,7,-0.004132710837137457,0.25557382642654747,0.008209258610480098,0.19962731715904453,-0.019377623384183553,0.16656801622824735,0.1884440463691251,0.3674668449615533
|
105 |
+
8,8,-0.016159192515121636,0.20852789817887454,-0.002296429673016746,0.15766743520841733,-0.028831385412794255,0.11443604168352871,0.1769550810769645,0.3674668449615533
|
106 |
+
8,9,-0.023883307089657038,0.18309766669364594,-0.009360229223159877,0.1474953426143259,-0.035199494385491366,0.11443604168352871,0.16903971135358642,0.3674668449615533
|
107 |
+
8,10,-0.022590568782368224,0.18309766669364594,-0.00861464573474373,0.15766743520841733,-0.03452284680631117,0.11443604168352871,0.17090633437618638,0.3674668449615533
|
108 |
+
8,11,-0.022205072593786467,0.18309766669364594,-0.00853361495784241,0.15766743520841733,-0.034443733200666994,0.11443604168352871,0.171610748993374,0.3674668449615533
|
109 |
+
8,-1,-0.022205072593786467,0.18309766669364594,-0.00853361495784241,0.15766743520841733,-0.034443733200666994,0.11443604168352871,0.171610748993374,0.3674668449615533
|
110 |
+
9,1,-0.019924152665853203,0.19708429401052166,-0.006927105965602912,0.15766743520841733,-0.03306191027379313,0.11443604168352871,0.17462414242576965,0.3674668449615533
|
111 |
+
9,2,-0.01310554134308347,0.25557382642654747,-0.0018651837179010037,0.15766743520841733,-0.028440416655785103,0.12460813427762014,0.182248413135009,0.3674668449615533
|
112 |
+
9,3,-0.006224843479141643,0.28863312735734464,0.003226885759312212,0.14622383104006448,-0.023843649331516222,0.11316453010926729,0.1894939173397539,0.33695056717927896
|
113 |
+
9,4,0.0015243244319323634,0.28863312735734464,0.00914542924075878,0.19962731715904453,-0.01850020127678038,0.13859476159449588,0.197196381669136,0.33695056717927896
|
114 |
+
9,5,0.006386296414661402,0.28863312735734464,0.012628857396022276,0.19962731715904453,-0.015287387789583351,0.15003836576284876,0.20199091464613517,0.33695056717927896
|
115 |
+
9,6,0.009383067148627529,0.28863312735734464,0.014698237108799568,0.19962731715904453,-0.013294489743420642,0.15003836576284876,0.2049775515320066,0.33695056717927896
|
116 |
+
9,7,0.010936210986226005,0.28863312735734464,0.01577361003181814,0.18309766669364594,-0.01222902667802249,0.15003836576284876,0.20654499656706846,0.33695056717927896
|
117 |
+
9,8,0.011462945729581623,0.28863312735734464,0.016048507472907553,0.18309766669364594,-0.011967145614619592,0.15003836576284876,0.20717147971628413,0.33695056717927896
|
118 |
+
9,9,0.012649786190150626,0.28863312735734464,0.016920545000646138,0.18309766669364594,-0.011154632276612609,0.15003836576284876,0.2083256074777095,0.33695056717927896
|
119 |
+
9,10,0.013565300652871653,0.28863312735734464,0.017650336624234146,0.18309766669364594,-0.010495535647356037,0.15003836576284876,0.20915339656481446,0.33695056717927896
|
120 |
+
9,11,0.014237166443878601,0.28863312735734464,0.01821190138046468,0.18309766669364594,-0.00997459346037672,0.15003836576284876,0.20963028248471155,0.35856626394172325
|
121 |
+
9,-1,0.014237166443878601,0.28863312735734464,0.01821190138046468,0.18309766669364594,-0.00997459346037672,0.15003836576284876,0.20963028248471155,0.35856626394172325
|
modules.json
ADDED
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
[
|
2 |
+
{
|
3 |
+
"idx": 0,
|
4 |
+
"name": "0",
|
5 |
+
"path": "",
|
6 |
+
"type": "sentence_transformers.models.Transformer"
|
7 |
+
},
|
8 |
+
{
|
9 |
+
"idx": 1,
|
10 |
+
"name": "1",
|
11 |
+
"path": "1_Pooling",
|
12 |
+
"type": "sentence_transformers.models.Pooling"
|
13 |
+
}
|
14 |
+
]
|
pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bd41ea7b13dee608d19d4f883abac33e9ac19090d7bcdb17cc5a02a18463c0f1
|
3 |
+
size 1112255985
|
sentence_bert_config.json
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"max_seq_length": 128,
|
3 |
+
"do_lower_case": false
|
4 |
+
}
|
sentencepiece.bpe.model
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cfc8146abe2a0488e9e2a0c56de7952f7c11ab059eca145a0a727afce0db2865
|
3 |
+
size 5069051
|
similarity_evaluation_sts-test_results.csv
ADDED
@@ -0,0 +1,2 @@
|
|
|
|
|
|
|
1 |
+
epoch,steps,cosine_pearson,cosine_spearman,euclidean_pearson,euclidean_spearman,manhattan_pearson,manhattan_spearman,dot_pearson,dot_spearman
|
2 |
+
-1,-1,0.8144962278898067,0.5888502610476809,0.8270483011118113,0.6193035502960349,0.8271866273573072,0.6196448091403334,0.7741324297034556,0.5041855668193329
|
special_tokens_map.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "sep_token": "</s>", "pad_token": "<pad>", "cls_token": "<s>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": false}}
|
tokenizer.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3a3313815c3d2e1b78b5182b09e66e6cd4cdd54df67a35c4a318c23d461821a4
|
3 |
+
size 17082913
|
tokenizer_config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"bos_token": "<s>", "eos_token": "</s>", "sep_token": "</s>", "cls_token": "<s>", "unk_token": "<unk>", "pad_token": "<pad>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "/root/.cache/torch/sentence_transformers/sentence-transformers_paraphrase-multilingual-mpnet-base-v2/", "tokenizer_class": "XLMRobertaTokenizer"}
|