jfarray commited on
Commit
c7cb5f1
1 Parent(s): 461fe20

Add new SentenceTransformer model.

Browse files
.gitattributes CHANGED
@@ -25,3 +25,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
25
  *.zip filter=lfs diff=lfs merge=lfs -text
26
  *.zstandard filter=lfs diff=lfs merge=lfs -text
27
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
25
  *.zip filter=lfs diff=lfs merge=lfs -text
26
  *.zstandard filter=lfs diff=lfs merge=lfs -text
27
  *tfevents* filter=lfs diff=lfs merge=lfs -text
28
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
29
+ pytorch_model.bin filter=lfs diff=lfs merge=lfs -text
1_Pooling/config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 768,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": true,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false
7
+ }
README.md ADDED
@@ -0,0 +1,125 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ pipeline_tag: sentence-similarity
3
+ tags:
4
+ - sentence-transformers
5
+ - feature-extraction
6
+ - sentence-similarity
7
+ - transformers
8
+ ---
9
+
10
+ # {MODEL_NAME}
11
+
12
+ This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.
13
+
14
+ <!--- Describe your model here -->
15
+
16
+ ## Usage (Sentence-Transformers)
17
+
18
+ Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed:
19
+
20
+ ```
21
+ pip install -U sentence-transformers
22
+ ```
23
+
24
+ Then you can use the model like this:
25
+
26
+ ```python
27
+ from sentence_transformers import SentenceTransformer
28
+ sentences = ["This is an example sentence", "Each sentence is converted"]
29
+
30
+ model = SentenceTransformer('{MODEL_NAME}')
31
+ embeddings = model.encode(sentences)
32
+ print(embeddings)
33
+ ```
34
+
35
+
36
+
37
+ ## Usage (HuggingFace Transformers)
38
+ Without [sentence-transformers](https://www.SBERT.net), you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings.
39
+
40
+ ```python
41
+ from transformers import AutoTokenizer, AutoModel
42
+ import torch
43
+
44
+
45
+ #Mean Pooling - Take attention mask into account for correct averaging
46
+ def mean_pooling(model_output, attention_mask):
47
+ token_embeddings = model_output[0] #First element of model_output contains all token embeddings
48
+ input_mask_expanded = attention_mask.unsqueeze(-1).expand(token_embeddings.size()).float()
49
+ return torch.sum(token_embeddings * input_mask_expanded, 1) / torch.clamp(input_mask_expanded.sum(1), min=1e-9)
50
+
51
+
52
+ # Sentences we want sentence embeddings for
53
+ sentences = ['This is an example sentence', 'Each sentence is converted']
54
+
55
+ # Load model from HuggingFace Hub
56
+ tokenizer = AutoTokenizer.from_pretrained('{MODEL_NAME}')
57
+ model = AutoModel.from_pretrained('{MODEL_NAME}')
58
+
59
+ # Tokenize sentences
60
+ encoded_input = tokenizer(sentences, padding=True, truncation=True, return_tensors='pt')
61
+
62
+ # Compute token embeddings
63
+ with torch.no_grad():
64
+ model_output = model(**encoded_input)
65
+
66
+ # Perform pooling. In this case, mean pooling.
67
+ sentence_embeddings = mean_pooling(model_output, encoded_input['attention_mask'])
68
+
69
+ print("Sentence embeddings:")
70
+ print(sentence_embeddings)
71
+ ```
72
+
73
+
74
+
75
+ ## Evaluation Results
76
+
77
+ <!--- Describe how your model was evaluated -->
78
+
79
+ For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name={MODEL_NAME})
80
+
81
+
82
+ ## Training
83
+ The model was trained with the parameters:
84
+
85
+ **DataLoader**:
86
+
87
+ `torch.utils.data.dataloader.DataLoader` of length 11 with parameters:
88
+ ```
89
+ {'batch_size': 15, 'sampler': 'torch.utils.data.sampler.RandomSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'}
90
+ ```
91
+
92
+ **Loss**:
93
+
94
+ `sentence_transformers.losses.CosineSimilarityLoss.CosineSimilarityLoss`
95
+
96
+ Parameters of the fit()-Method:
97
+ ```
98
+ {
99
+ "epochs": 10,
100
+ "evaluation_steps": 1,
101
+ "evaluator": "sentence_transformers.evaluation.EmbeddingSimilarityEvaluator.EmbeddingSimilarityEvaluator",
102
+ "max_grad_norm": 1,
103
+ "optimizer_class": "<class 'transformers.optimization.AdamW'>",
104
+ "optimizer_params": {
105
+ "lr": 2e-05
106
+ },
107
+ "scheduler": "WarmupLinear",
108
+ "steps_per_epoch": null,
109
+ "warmup_steps": 11,
110
+ "weight_decay": 0.01
111
+ }
112
+ ```
113
+
114
+
115
+ ## Full Model Architecture
116
+ ```
117
+ SentenceTransformer(
118
+ (0): Transformer({'max_seq_length': 128, 'do_lower_case': False}) with Transformer model: XLMRobertaModel
119
+ (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False})
120
+ )
121
+ ```
122
+
123
+ ## Citing & Authors
124
+
125
+ <!--- Describe where people can find more information -->
config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/root/.cache/torch/sentence_transformers/sentence-transformers_paraphrase-multilingual-mpnet-base-v2/",
3
+ "architectures": [
4
+ "XLMRobertaModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
+ "gradient_checkpointing": false,
11
+ "hidden_act": "gelu",
12
+ "hidden_dropout_prob": 0.1,
13
+ "hidden_size": 768,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 3072,
16
+ "layer_norm_eps": 1e-05,
17
+ "max_position_embeddings": 514,
18
+ "model_type": "xlm-roberta",
19
+ "num_attention_heads": 12,
20
+ "num_hidden_layers": 12,
21
+ "output_past": true,
22
+ "pad_token_id": 1,
23
+ "position_embedding_type": "absolute",
24
+ "torch_dtype": "float32",
25
+ "transformers_version": "4.16.2",
26
+ "type_vocab_size": 1,
27
+ "use_cache": true,
28
+ "vocab_size": 250002
29
+ }
config_sentence_transformers.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "2.0.0",
4
+ "transformers": "4.7.0",
5
+ "pytorch": "1.9.0+cu102"
6
+ }
7
+ }
eval/similarity_evaluation_results.csv ADDED
@@ -0,0 +1,121 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch,steps,cosine_pearson,cosine_spearman,euclidean_pearson,euclidean_spearman,manhattan_pearson,manhattan_spearman,dot_pearson,dot_spearman
2
+ 0,1,0.24428484492337774,0.1856406898421688,0.2726824721314313,0.17419708567381592,0.20191195613958324,0.18182615511938452,-0.10491629431153585,-0.05976104399028721
3
+ 0,2,0.22612378962507065,0.1474953426143259,0.2560357877483311,0.17928313197086163,0.17884221008459156,0.13732325002023446,-0.11666791021706803,-0.07374767130716295
4
+ 0,3,0.19677200683261806,0.10934999538648299,0.23751600007738927,0.16529650465398593,0.15040892380050047,0.01398662731687573,-0.15956732474655524,-0.11570755325779014
5
+ 0,4,0.11503452111144055,0.025430231485228605,0.15761318436439234,0.01907267361392145,0.058673887759007365,-0.022887208336705742,-0.18204305619588393,-0.21361394447592028
6
+ 0,5,0.003985376234194979,-0.005086046297045721,0.03632262297764966,0.012715115742614302,-0.05827011779841074,-0.12460813427762014,-0.1913564390855504,-0.25048778012950174
7
+ 0,6,-0.05103630332147053,-0.026701743059490034,-0.03439079582489007,-0.10807848381222157,-0.11438987777919786,-0.1296941805746659,-0.16467537031081167,-0.2492162685552403
8
+ 0,7,-0.09921001875664615,-0.13986627316875733,-0.09075888010265239,-0.11952208798057443,-0.1596978337595956,-0.19199824771347596,-0.14141547945028185,-0.2835470810602989
9
+ 0,8,-0.13983443384729216,-0.21234243290165886,-0.14450538879193142,-0.18182615511938452,-0.20808728093382015,-0.2492162685552403,-0.12541343825340248,-0.25048778012950174
10
+ 0,9,-0.15921298240054033,-0.23777266438688743,-0.1797885889172838,-0.1907267361392145,-0.24392530175624805,-0.2492162685552403,-0.1048356212094887,-0.17038255095103164
11
+ 0,10,-0.1609468704741812,-0.20471336345609026,-0.19744846790394677,-0.22124301392148887,-0.26585823378739704,-0.2657459190206389,-0.04824870832250316,-0.11443604168352871
12
+ 0,11,-0.15410333770549933,-0.18436917826790739,-0.2093702035319172,-0.16021045835694017,-0.2808278669531017,-0.3038912662484818,0.046277154708522575,0.026701743059490034
13
+ 0,-1,-0.15410333770549933,-0.18436917826790739,-0.2093702035319172,-0.16021045835694017,-0.2808278669531017,-0.3038912662484818,0.046277154708522575,0.026701743059490034
14
+ 1,1,-0.12982786622499196,-0.15003836576284876,-0.21235892958369582,-0.1983558055847831,-0.2909451954930539,-0.25557382642654747,0.1675258309470891,0.24794475698097887
15
+ 1,2,-0.19696595634631142,-0.16656801622824735,-0.25668213221583064,-0.2326866180898417,-0.3250005522733054,-0.3801819607041676,0.04540854622383069,-0.00381453472278429
16
+ 1,3,-0.21781655597904578,-0.27591801161473034,-0.2747155636631352,-0.32296393986240324,-0.33711431182747686,-0.37509591440712187,-0.015301510197506163,-0.0673901134358558
17
+ 1,4,-0.21040566408092895,-0.20725638660461312,-0.2787115989564245,-0.31406335884257325,-0.33960265883733304,-0.32804998615944897,0.05601493400121786,0.01525813889113716
18
+ 1,5,-0.21075832691211716,-0.2492162685552403,-0.2826518195170755,-0.3356790556050176,-0.339197989506619,-0.35093719449615474,0.07529820685804627,0.057218020841764354
19
+ 1,6,-0.20609520571963286,-0.27210347689194603,-0.27951092027179186,-0.35856626394172325,-0.33310429600778424,-0.37255289125859903,0.11402579008036735,0.11570755325779014
20
+ 1,7,-0.169486532878668,-0.2708319653176846,-0.26423165753461786,-0.3344075440307561,-0.31980430200287796,-0.36110928709024614,0.20002633803728215,0.3115203356940504
21
+ 1,8,-0.0901707317417178,-0.14113778474301875,-0.2177628930192162,-0.2810040579117761,-0.2751703064164917,-0.2988052199514361,0.2490250330297735,0.40179765746661195
22
+ 1,9,-0.004556612508979109,0.16402499307972448,-0.14854868139699784,-0.22124301392148887,-0.20540087365227688,-0.23650115281262601,0.2707949105636119,0.4195988195062719
23
+ 1,10,0.0677969000349118,0.32296393986240324,-0.05310832958181209,0.00381453472278429,-0.09639636112386697,-0.01652965046539859,0.2734802712416607,0.40179765746661195
24
+ 1,11,0.11315696215310725,0.3064342893970047,0.029403827430142117,0.17419708567381592,0.0028628903329707502,0.16529650465398593,0.27353726982326,0.39035405329825906
25
+ 1,-1,0.11315696215310725,0.3064342893970047,0.029403827430142117,0.17419708567381592,0.0028628903329707502,0.16529650465398593,0.27353726982326,0.39035405329825906
26
+ 2,1,0.13940714819785371,0.3293214977337104,0.08566760205046141,0.25557382642654747,0.06824408200836007,0.21234243290165886,0.2712000137342313,0.39035405329825906
27
+ 2,2,0.14699598094492095,0.37255289125859903,0.10174450966560532,0.30134824309995895,0.08663053449180397,0.2606598727235932,0.27337148870045525,0.39035405329825906
28
+ 2,3,0.14656120764881206,0.32042091671388034,0.10977333616234204,0.29244766208012896,0.0968236329054813,0.20471336345609026,0.2728593350922896,0.39035405329825906
29
+ 2,4,0.13675405809671398,0.4056121921893962,0.10241204205207184,0.29244766208012896,0.09005215314493258,0.2568453380008089,0.27303591622071344,0.3941685880210433
30
+ 2,5,0.13799404823353287,0.3674668449615533,0.10605153984267622,0.2708319653176846,0.09628586334092931,0.23777266438688743,0.2774065715793497,0.381453472278429
31
+ 2,6,0.11872275011544195,0.3318645208822333,0.07989894345468976,0.1907267361392145,0.06894963431264978,0.1907267361392145,0.2772175356157276,0.4208703310805334
32
+ 2,7,0.10777041689179366,0.3038912662484818,0.06679490929721191,0.21742847919870456,0.05639645529687149,0.15893894678267878,0.27587894919073314,0.38526800700121333
33
+ 2,8,0.08964774465066375,0.25048778012950174,0.04562989321901026,0.17419708567381592,0.03533869876279845,0.14495231946580303,0.2695904387278222,0.3700098681100762
34
+ 2,9,0.07956450993625785,0.22887208336705742,0.03178353953441007,0.09790639121813012,0.020992268554907812,0.11825057640631301,0.2681886742472172,0.3827249838526905
35
+ 2,10,0.06530632310909693,0.23904417596114885,0.01290320159856114,0.07883371760420867,-0.0008022100415914147,0.07883371760420867,0.2652623096945603,0.39925463431808905
36
+ 2,11,0.0541526868004497,0.20598487503035168,-0.0014604552039510635,0.05340348611898007,-0.020671016000849286,0.045774416673411485,0.26474831609084454,0.3661953333872919
37
+ 2,-1,0.0541526868004497,0.20598487503035168,-0.0014604552039510635,0.05340348611898007,-0.020671016000849286,0.045774416673411485,0.26474831609084454,0.3661953333872919
38
+ 3,1,0.04082352652346983,0.1780116203966002,-0.016950879494709237,0.024158719910967172,-0.04256928272155714,-0.024158719910967172,0.26136168380031144,0.3064342893970047
39
+ 3,2,0.039171457779169055,0.18309766669364594,-0.022019282369464054,0.005086046297045721,-0.0534613577807422,-0.06230406713881007,0.2626351974962769,0.33059300930797186
40
+ 3,3,0.03962702632659584,0.21615696762444314,-0.021501770740796965,-0.01780116203966002,-0.05837792754185962,-0.024158719910967172,0.2610292488191349,0.35220870607041616
41
+ 3,4,0.05138318751675695,0.19581278243626024,-0.009772229122192283,0.048317439821934344,-0.05131522224080271,0.006357557871307151,0.26451055406709656,0.36110928709024614
42
+ 3,5,0.042919953964449316,0.18055464354512307,-0.017252621913390217,0.07374767130716295,-0.06272951255774184,-0.025430231485228605,0.25893214830143135,0.3293214977337104
43
+ 3,6,0.01621406176307976,0.14113778474301875,-0.044121003338826015,-0.052131974544718636,-0.09192335750297764,-0.0673901134358558,0.24414022952783723,0.28609010420882175
44
+ 3,7,0.01603013319187219,0.11952208798057443,-0.044933401380679173,-0.04450290509915005,-0.09369405138432947,-0.09663487964386869,0.2435001866086313,0.28609010420882175
45
+ 3,8,0.041056573181203274,0.17419708567381592,-0.02521008601539706,0.01652965046539859,-0.07780417603142906,-0.012715115742614302,0.25960723616492987,0.2822755694860375
46
+ 3,9,0.06668407275286904,0.20852789817887454,-0.004198856765651891,0.0928203449210844,-0.05824260675806149,0.026701743059490034,0.27463594492277404,0.354751729218939
47
+ 3,10,0.09732084786674933,0.17165406252529306,0.024355250848309237,0.09917790279239155,-0.031437585110709654,0.09154883334682297,0.287547698402672,0.4412145162687163
48
+ 3,11,0.09969461581415269,0.20471336345609026,0.02412155708892016,0.09154883334682297,-0.030725799309395584,0.12715115742614302,0.29083133024082536,0.49080346766491206
49
+ 3,-1,0.09969461581415269,0.20471336345609026,0.02412155708892016,0.09154883334682297,-0.030725799309395584,0.12715115742614302,0.29083133024082536,0.49080346766491206
50
+ 4,1,0.09688106671926407,0.20471336345609026,0.019849707506481476,0.15639592363415591,-0.033138832196187684,0.16402499307972448,0.29020825241438647,0.43739998154593196
51
+ 4,2,0.10686210835084084,0.227600571792796,0.03527764918280338,0.21107092132739738,-0.014809215605804057,0.14495231946580303,0.2871024991131068,0.4437575394172391
52
+ 4,3,0.11528029125188341,0.227600571792796,0.046012326958320846,0.16529650465398593,-0.0014913913693886001,0.21361394447592028,0.28614822805844137,0.43867149312019343
53
+ 4,4,0.11327236066146648,0.2708319653176846,0.04816513612968369,0.21615696762444314,0.002628733543029182,0.21361394447592028,0.28244052688585697,0.43867149312019343
54
+ 4,5,0.1097521443719211,0.2962621968029132,0.053605501185894744,0.23777266438688743,0.011076965089880491,0.2352296412383646,0.274889620003233,0.44248602784297775
55
+ 4,6,0.085118512450099,0.24540173383245603,0.035052362001546225,0.20471336345609026,-0.004599692497119436,0.18055464354512307,0.2614755796846744,0.4094267269121805
56
+ 4,7,0.06647371176348317,0.227600571792796,0.022546934649379773,0.20471336345609026,-0.01489167340998393,0.1678395278025088,0.24772311028462143,0.4310424236746248
57
+ 4,8,0.058433927289096785,0.227600571792796,0.020394938037328333,0.18691220141643022,-0.01488648898505835,0.1347802268717116,0.238793669664602,0.4094267269121805
58
+ 4,9,0.06218765255951492,0.2606598727235932,0.028254863204573506,0.18691220141643022,-0.0059663318280807195,0.1347802268717116,0.23662507196670224,0.4310424236746248
59
+ 4,10,0.05946027313650021,0.19454127086199882,0.026231299009535372,0.18691220141643022,-0.007429255669431191,0.12333662270335873,0.23451281524353298,0.381453472278429
60
+ 4,11,0.04488249947205441,0.23777266438688743,0.016194060108917257,0.19326975928773737,-0.014837134690265971,0.14622383104006448,0.22440272069256118,0.3661953333872919
61
+ 4,-1,0.04488249947205441,0.23777266438688743,0.016194060108917257,0.19326975928773737,-0.014837134690265971,0.14622383104006448,0.22440272069256118,0.3661953333872919
62
+ 5,1,0.053395547312709554,0.2326866180898417,0.025551009989655416,0.20471336345609026,-0.0030336681748221674,0.16021045835694017,0.23046363705984724,0.3878110301497362
63
+ 5,2,0.0591569326701527,0.23777266438688743,0.03128838249262078,0.20471336345609026,0.0043441502556022605,0.17165406252529306,0.23440759776991563,0.3789104491299062
64
+ 5,3,0.06450234649256706,0.23777266438688743,0.03449441250539357,0.18055464354512307,0.008766844563450683,0.20725638660461312,0.24108678497623875,0.3789104491299062
65
+ 5,4,0.06214173452043999,0.23777266438688743,0.032097244328093966,0.16656801622824735,0.008197085492201326,0.18309766669364594,0.2426413857518142,0.3878110301497362
66
+ 5,5,0.06578045671422762,0.23777266438688743,0.03542313897925867,0.15639592363415591,0.012700649092421475,0.18309766669364594,0.24607156573013772,0.4094267269121805
67
+ 5,6,0.08462140876157086,0.2708319653176846,0.0522684701873643,0.1347802268717116,0.030843014529425592,0.1729255740995545,0.25759507497226136,0.3789104491299062
68
+ 5,7,0.0990143040299816,0.23904417596114885,0.06462391531700609,0.18436917826790739,0.043981507888789885,0.21615696762444314,0.2652275602224404,0.38908254172399764
69
+ 5,8,0.11663814351298049,0.23395812966410312,0.08161265028948292,0.20852789817887454,0.061440389630868536,0.19454127086199882,0.27160989904850547,0.4170557963577491
70
+ 5,9,0.12327135444243054,0.20979940975313596,0.08851227558689899,0.19454127086199882,0.06924274661872722,0.227600571792796,0.2735691512978767,0.4348569583974091
71
+ 5,10,0.1256530678741609,0.20979940975313596,0.09225233834877479,0.19454127086199882,0.07268984121640336,0.2708319653176846,0.2742497099134256,0.4348569583974091
72
+ 5,11,0.12873361419741985,0.23141510651558028,0.09823515122296156,0.227600571792796,0.07889129768310928,0.2708319653176846,0.27363372133242203,0.4348569583974091
73
+ 5,-1,0.12873361419741985,0.23141510651558028,0.09823515122296156,0.227600571792796,0.07889129768310928,0.2708319653176846,0.27363372133242203,0.4348569583974091
74
+ 6,1,0.12567595694042172,0.23141510651558028,0.09857514532217378,0.24667324540671745,0.07986965314509253,0.2708319653176846,0.27056410374820217,0.4348569583974091
75
+ 6,2,0.12149628573204947,0.25430231485228605,0.0961034528372844,0.27210347689194603,0.07726412770312542,0.2530308032780246,0.26964215394565244,0.4170557963577491
76
+ 6,3,0.10559031426059957,0.26447440744637746,0.08239293894232613,0.24794475698097887,0.06377960458713212,0.27591801161473034,0.26486744741490464,0.4399430046944548
77
+ 6,4,0.10259542528560504,0.26447440744637746,0.08061103179553537,0.24794475698097887,0.06204550170088915,0.2657459190206389,0.26434917766482163,0.4399430046944548
78
+ 6,5,0.089628434872583,0.26447440744637746,0.06910921853351924,0.24158719910967172,0.050042422794654796,0.24158719910967172,0.26022687306020714,0.41196975006070335
79
+ 6,6,0.06903104253323819,0.2657459190206389,0.051863084231036784,0.17546859724807737,0.03259363057545481,0.17928313197086163,0.2514680535201101,0.3878110301497362
80
+ 6,7,0.06627359161868877,0.28863312735734464,0.04856471876248964,0.17546859724807737,0.0290375512498293,0.18055464354512307,0.2510346416505546,0.3878110301497362
81
+ 6,8,0.06587585027778702,0.28863312735734464,0.04744858656278781,0.18691220141643022,0.027655219369644334,0.18055464354512307,0.25162183829556445,0.3916255648725205
82
+ 6,9,0.07667645472428877,0.28609010420882175,0.05489932876112217,0.2199715023472274,0.03397228404864987,0.21361394447592028,0.25906650914738716,0.3916255648725205
83
+ 6,10,0.08455113989210214,0.30134824309995895,0.06023682494544593,0.24158719910967172,0.03833933740430189,0.2568453380008089,0.26322096161083997,0.3916255648725205
84
+ 6,11,0.06527064441009967,0.28609010420882175,0.04645573451990924,0.2199715023472274,0.023629072577399244,0.24667324540671745,0.25261854507949266,0.3916255648725205
85
+ 6,-1,0.06527064441009967,0.28609010420882175,0.04645573451990924,0.2199715023472274,0.023629072577399244,0.24667324540671745,0.25261854507949266,0.3916255648725205
86
+ 7,1,0.060974209679639874,0.2962621968029132,0.044182665431398774,0.2199715023472274,0.020208359608723723,0.2568453380008089,0.2500246080799369,0.3916255648725205
87
+ 7,2,0.05889986791654363,0.2962621968029132,0.04367928785660007,0.20852789817887454,0.018990969899362664,0.2568453380008089,0.24852263732052335,0.3801819607041676
88
+ 7,3,0.0586805419019325,0.2962621968029132,0.04510026160066266,0.20852789817887454,0.01947097540749484,0.24158719910967172,0.24833599501031361,0.3801819607041676
89
+ 7,4,0.06086112074593676,0.3064342893970047,0.04777464187251144,0.24158719910967172,0.020697437016754666,0.24158719910967172,0.2494308562127234,0.3801819607041676
90
+ 7,5,0.06280098215237057,0.3064342893970047,0.05051831916576996,0.24158719910967172,0.02218930339762014,0.24158719910967172,0.24971834944048754,0.39035405329825906
91
+ 7,6,0.06158031126576389,0.3064342893970047,0.05075105683420316,0.24158719910967172,0.02159771518228107,0.20852789817887454,0.24800288493234027,0.39035405329825906
92
+ 7,7,0.06397984201251046,0.3064342893970047,0.05426642849551435,0.24158719910967172,0.024405314267635883,0.21869999077296598,0.24762507347096327,0.39035405329825906
93
+ 7,8,0.0659829841222002,0.30134824309995895,0.05731078982845222,0.2746465000404689,0.027236857467774822,0.21869999077296598,0.24696902815106028,0.41196975006070335
94
+ 7,9,0.07164671048348455,0.30134824309995895,0.06349894299155719,0.25175929170376316,0.03336710644089848,0.27591801161473034,0.2474211727149824,0.41196975006070335
95
+ 7,10,0.08963448736837631,0.3038912662484818,0.07951013636920258,0.27591801161473034,0.04793689428134984,0.2530308032780246,0.25395789779269695,0.40306916904087337
96
+ 7,11,0.09347383398858716,0.2949906852286518,0.08502494874271449,0.27591801161473034,0.05304476405077205,0.2733749884662075,0.25375858054851336,0.40306916904087337
97
+ 7,-1,0.09347383398858716,0.2949906852286518,0.08502494874271449,0.27591801161473034,0.05304476405077205,0.2733749884662075,0.25375858054851336,0.40306916904087337
98
+ 8,1,0.09975180365559692,0.2733749884662075,0.09272251171908412,0.27591801161473034,0.06045198277887114,0.2733749884662075,0.25493848861839186,0.40306916904087337
99
+ 8,2,0.10048681631029896,0.2949906852286518,0.09485771368118302,0.27591801161473034,0.061916887798136984,0.2733749884662075,0.2544032330293862,0.40306916904087337
100
+ 8,3,0.0828267554578788,0.3038912662484818,0.08022802667414386,0.27591801161473034,0.046963894195038346,0.27591801161473034,0.2465196613596623,0.4056121921893962
101
+ 8,4,0.059137661106999834,0.3064342893970047,0.06073868996694734,0.2988052199514361,0.02847229929037812,0.2657459190206389,0.2342259758391944,0.3661953333872919
102
+ 8,5,0.037048084368087424,0.3089773125455275,0.04264002039285007,0.24158719910967172,0.011641961257956638,0.20852789817887454,0.22059703173457557,0.3661953333872919
103
+ 8,6,0.013393957920069248,0.28863312735734464,0.023186005391405987,0.18309766669364594,-0.006231614744725678,0.15003836576284876,0.2034548517038958,0.33695056717927896
104
+ 8,7,-0.004132710837137457,0.25557382642654747,0.008209258610480098,0.19962731715904453,-0.019377623384183553,0.16656801622824735,0.1884440463691251,0.3674668449615533
105
+ 8,8,-0.016159192515121636,0.20852789817887454,-0.002296429673016746,0.15766743520841733,-0.028831385412794255,0.11443604168352871,0.1769550810769645,0.3674668449615533
106
+ 8,9,-0.023883307089657038,0.18309766669364594,-0.009360229223159877,0.1474953426143259,-0.035199494385491366,0.11443604168352871,0.16903971135358642,0.3674668449615533
107
+ 8,10,-0.022590568782368224,0.18309766669364594,-0.00861464573474373,0.15766743520841733,-0.03452284680631117,0.11443604168352871,0.17090633437618638,0.3674668449615533
108
+ 8,11,-0.022205072593786467,0.18309766669364594,-0.00853361495784241,0.15766743520841733,-0.034443733200666994,0.11443604168352871,0.171610748993374,0.3674668449615533
109
+ 8,-1,-0.022205072593786467,0.18309766669364594,-0.00853361495784241,0.15766743520841733,-0.034443733200666994,0.11443604168352871,0.171610748993374,0.3674668449615533
110
+ 9,1,-0.019924152665853203,0.19708429401052166,-0.006927105965602912,0.15766743520841733,-0.03306191027379313,0.11443604168352871,0.17462414242576965,0.3674668449615533
111
+ 9,2,-0.01310554134308347,0.25557382642654747,-0.0018651837179010037,0.15766743520841733,-0.028440416655785103,0.12460813427762014,0.182248413135009,0.3674668449615533
112
+ 9,3,-0.006224843479141643,0.28863312735734464,0.003226885759312212,0.14622383104006448,-0.023843649331516222,0.11316453010926729,0.1894939173397539,0.33695056717927896
113
+ 9,4,0.0015243244319323634,0.28863312735734464,0.00914542924075878,0.19962731715904453,-0.01850020127678038,0.13859476159449588,0.197196381669136,0.33695056717927896
114
+ 9,5,0.006386296414661402,0.28863312735734464,0.012628857396022276,0.19962731715904453,-0.015287387789583351,0.15003836576284876,0.20199091464613517,0.33695056717927896
115
+ 9,6,0.009383067148627529,0.28863312735734464,0.014698237108799568,0.19962731715904453,-0.013294489743420642,0.15003836576284876,0.2049775515320066,0.33695056717927896
116
+ 9,7,0.010936210986226005,0.28863312735734464,0.01577361003181814,0.18309766669364594,-0.01222902667802249,0.15003836576284876,0.20654499656706846,0.33695056717927896
117
+ 9,8,0.011462945729581623,0.28863312735734464,0.016048507472907553,0.18309766669364594,-0.011967145614619592,0.15003836576284876,0.20717147971628413,0.33695056717927896
118
+ 9,9,0.012649786190150626,0.28863312735734464,0.016920545000646138,0.18309766669364594,-0.011154632276612609,0.15003836576284876,0.2083256074777095,0.33695056717927896
119
+ 9,10,0.013565300652871653,0.28863312735734464,0.017650336624234146,0.18309766669364594,-0.010495535647356037,0.15003836576284876,0.20915339656481446,0.33695056717927896
120
+ 9,11,0.014237166443878601,0.28863312735734464,0.01821190138046468,0.18309766669364594,-0.00997459346037672,0.15003836576284876,0.20963028248471155,0.35856626394172325
121
+ 9,-1,0.014237166443878601,0.28863312735734464,0.01821190138046468,0.18309766669364594,-0.00997459346037672,0.15003836576284876,0.20963028248471155,0.35856626394172325
modules.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ }
14
+ ]
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd41ea7b13dee608d19d4f883abac33e9ac19090d7bcdb17cc5a02a18463c0f1
3
+ size 1112255985
sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 128,
3
+ "do_lower_case": false
4
+ }
sentencepiece.bpe.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfc8146abe2a0488e9e2a0c56de7952f7c11ab059eca145a0a727afce0db2865
3
+ size 5069051
similarity_evaluation_sts-test_results.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ epoch,steps,cosine_pearson,cosine_spearman,euclidean_pearson,euclidean_spearman,manhattan_pearson,manhattan_spearman,dot_pearson,dot_spearman
2
+ -1,-1,0.8144962278898067,0.5888502610476809,0.8270483011118113,0.6193035502960349,0.8271866273573072,0.6196448091403334,0.7741324297034556,0.5041855668193329
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "sep_token": "</s>", "pad_token": "<pad>", "cls_token": "<s>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": false}}
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a3313815c3d2e1b78b5182b09e66e6cd4cdd54df67a35c4a318c23d461821a4
3
+ size 17082913
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "sep_token": "</s>", "cls_token": "<s>", "unk_token": "<unk>", "pad_token": "<pad>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "/root/.cache/torch/sentence_transformers/sentence-transformers_paraphrase-multilingual-mpnet-base-v2/", "tokenizer_class": "XLMRobertaTokenizer"}