jfarray commited on
Commit
5e0ee9f
1 Parent(s): 138910f

Add new SentenceTransformer model.

Browse files
.gitattributes CHANGED
@@ -25,3 +25,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
25
  *.zip filter=lfs diff=lfs merge=lfs -text
26
  *.zstandard filter=lfs diff=lfs merge=lfs -text
27
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
25
  *.zip filter=lfs diff=lfs merge=lfs -text
26
  *.zstandard filter=lfs diff=lfs merge=lfs -text
27
  *tfevents* filter=lfs diff=lfs merge=lfs -text
28
+ pytorch_model.bin filter=lfs diff=lfs merge=lfs -text
1_Pooling/config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 768,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": true,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false
7
+ }
2_Dense/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"in_features": 768, "out_features": 256, "bias": true, "activation_function": "torch.nn.modules.activation.Tanh"}
2_Dense/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:afae1d3936856cfb47ac9809e6c8ae677187248b7d4b8dadd7d8577b9266f1ba
3
+ size 788519
README.md ADDED
@@ -0,0 +1,87 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ pipeline_tag: sentence-similarity
3
+ tags:
4
+ - sentence-transformers
5
+ - feature-extraction
6
+ - sentence-similarity
7
+ ---
8
+
9
+ # {MODEL_NAME}
10
+
11
+ This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 256 dimensional dense vector space and can be used for tasks like clustering or semantic search.
12
+
13
+ <!--- Describe your model here -->
14
+
15
+ ## Usage (Sentence-Transformers)
16
+
17
+ Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed:
18
+
19
+ ```
20
+ pip install -U sentence-transformers
21
+ ```
22
+
23
+ Then you can use the model like this:
24
+
25
+ ```python
26
+ from sentence_transformers import SentenceTransformer
27
+ sentences = ["This is an example sentence", "Each sentence is converted"]
28
+
29
+ model = SentenceTransformer('{MODEL_NAME}')
30
+ embeddings = model.encode(sentences)
31
+ print(embeddings)
32
+ ```
33
+
34
+
35
+
36
+ ## Evaluation Results
37
+
38
+ <!--- Describe how your model was evaluated -->
39
+
40
+ For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name={MODEL_NAME})
41
+
42
+
43
+ ## Training
44
+ The model was trained with the parameters:
45
+
46
+ **DataLoader**:
47
+
48
+ `torch.utils.data.dataloader.DataLoader` of length 11 with parameters:
49
+ ```
50
+ {'batch_size': 15, 'sampler': 'torch.utils.data.sampler.RandomSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'}
51
+ ```
52
+
53
+ **Loss**:
54
+
55
+ `sentence_transformers.losses.CosineSimilarityLoss.CosineSimilarityLoss`
56
+
57
+ Parameters of the fit()-Method:
58
+ ```
59
+ {
60
+ "epochs": 10,
61
+ "evaluation_steps": 1,
62
+ "evaluator": "sentence_transformers.evaluation.EmbeddingSimilarityEvaluator.EmbeddingSimilarityEvaluator",
63
+ "max_grad_norm": 1,
64
+ "optimizer_class": "<class 'transformers.optimization.AdamW'>",
65
+ "optimizer_params": {
66
+ "lr": 2e-05
67
+ },
68
+ "scheduler": "WarmupLinear",
69
+ "steps_per_epoch": null,
70
+ "warmup_steps": 11,
71
+ "weight_decay": 0.01
72
+ }
73
+ ```
74
+
75
+
76
+ ## Full Model Architecture
77
+ ```
78
+ SentenceTransformer(
79
+ (0): Transformer({'max_seq_length': 256, 'do_lower_case': False}) with Transformer model: BertModel
80
+ (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False})
81
+ (2): Dense({'in_features': 768, 'out_features': 256, 'bias': True, 'activation_function': 'torch.nn.modules.activation.Tanh'})
82
+ )
83
+ ```
84
+
85
+ ## Citing & Authors
86
+
87
+ <!--- Describe where people can find more information -->
config.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "dccuchile/bert-base-spanish-wwm-uncased",
3
+ "architectures": [
4
+ "BertModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "output_past": true,
20
+ "pad_token_id": 1,
21
+ "position_embedding_type": "absolute",
22
+ "torch_dtype": "float32",
23
+ "transformers_version": "4.16.2",
24
+ "type_vocab_size": 2,
25
+ "use_cache": true,
26
+ "vocab_size": 31002
27
+ }
config_sentence_transformers.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "2.2.0",
4
+ "transformers": "4.16.2",
5
+ "pytorch": "1.10.0+cu111"
6
+ }
7
+ }
eval/similarity_evaluation_results.csv ADDED
@@ -0,0 +1,121 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch,steps,cosine_pearson,cosine_spearman,euclidean_pearson,euclidean_spearman,manhattan_pearson,manhattan_spearman,dot_pearson,dot_spearman
2
+ 0,1,0.044028602135798976,0.0928203449210844,0.05400669510206391,0.05340348611898007,0.020463610329045097,0.01398662731687573,0.11680914704485608,0.2708319653176846
3
+ 0,2,0.05275131996206153,0.12333662270335873,0.05971412654727069,0.05340348611898007,0.0242926267299544,0.01398662731687573,0.12370470715988531,0.2568453380008089
4
+ 0,3,0.07680316143401236,0.12333662270335873,0.07346971601133402,0.07756220602994723,0.03181493791572689,0.01398662731687573,0.14140138015874898,0.2568453380008089
5
+ 0,4,0.11140220416150023,0.12333662270335873,0.0917990480554461,0.10553546066369869,0.04149898898373833,0.0419598819506272,0.1636063284028412,0.2708319653176846
6
+ 0,5,0.1671151023762335,0.16529650465398593,0.12017059118986438,0.12715115742614302,0.061033363174034086,0.06230406713881007,0.18857080819963543,0.3331360324564947
7
+ 0,6,0.24065719302740873,0.2606598727235932,0.15923973779833794,0.12715115742614302,0.091345477077191,0.09790639121813012,0.21739848759987018,0.27718952318899176
8
+ 0,7,0.2816739324975813,0.32804998615944897,0.17347996520021278,0.16021045835694017,0.10617365118128576,0.06866162501011723,0.2538977658839316,0.36873835653581477
9
+ 0,8,0.3548892617553565,0.4068837037636577,0.2099147683069106,0.19962731715904453,0.13550452822742595,0.0381453472278429,0.28343206763849466,0.4348569583974091
10
+ 0,9,0.40975573088122297,0.46283021303116056,0.26772634870906475,0.20979940975313596,0.1901571851435937,0.1767401088223388,0.31568987307998364,0.42849940052610197
11
+ 0,10,0.4175238640507224,0.5073331181303107,0.3228185058118779,0.27591801161473034,0.23448566979794136,0.21107092132739738,0.3129039251571303,0.4170557963577491
12
+ 0,11,0.41920038907567053,0.5505645116551992,0.4138092583042801,0.38526800700121333,0.33191454870071174,0.37763893755564476,0.31902148093836985,0.4590156783083763
13
+ 0,-1,0.41920038907567053,0.5505645116551992,0.4138092583042801,0.38526800700121333,0.33191454870071174,0.37763893755564476,0.31902148093836985,0.4590156783083763
14
+ 1,1,0.38939160941261597,0.5111476528530949,0.45913435377501977,0.5340348611898006,0.38074919448079236,0.4348569583974091,0.30806342309372875,0.4590156783083763
15
+ 1,2,0.37184391394725136,0.5111476528530949,0.4471608917786044,0.5709086968433822,0.37528639867845875,0.44630056256576195,0.30572159382257685,0.5149621875758792
16
+ 1,3,0.3912993762637881,0.4895319560906506,0.45216492417953713,0.5798092778632121,0.3792540463033001,0.44884358571428484,0.31956487506377795,0.5454784653581536
17
+ 1,4,0.4228734660154477,0.5594650926750293,0.46368984206001684,0.6090540440712251,0.3781872600692586,0.46155870145689915,0.34118154685295127,0.5607366042492907
18
+ 1,5,0.43186671302194346,0.537849395912585,0.46510637243342895,0.6090540440712251,0.37538233064248977,0.4539296320113305,0.3537843939447062,0.5848953241602579
19
+ 1,6,0.4484090822047207,0.5403924190611078,0.47163997417404274,0.564551138972075,0.37603682422957047,0.44884358571428484,0.37289664876364675,0.5785377662889507
20
+ 1,7,0.4455175486364017,0.5225912570214478,0.4688190803157346,0.564551138972075,0.3756189558674257,0.44757207414002337,0.3787145301679507,0.5556505579522449
21
+ 1,8,0.453060278309398,0.5442069537838921,0.47206219653914,0.5887098588830421,0.38062966298436696,0.455201143585592,0.3928671385545982,0.5556505579522449
22
+ 1,9,0.46145642351805094,0.5556505579522449,0.4747676541173092,0.5594650926750293,0.38763345657318465,0.44757207414002337,0.40771855212737745,0.5556505579522449
23
+ 1,10,0.45749221557591563,0.5353063727640621,0.47064460796147034,0.5594650926750293,0.3922219263718788,0.4437575394172391,0.4095562141644134,0.5772662547146893
24
+ 1,11,0.4347541707123933,0.5403924190611078,0.4551135459893348,0.5607366042492907,0.38390310267647726,0.5086046297045721,0.3958241647117268,0.5416639306353692
25
+ 1,-1,0.4347541707123933,0.5403924190611078,0.4551135459893348,0.5607366042492907,0.38390310267647726,0.5086046297045721,0.3958241647117268,0.5416639306353692
26
+ 2,1,0.4059294206710656,0.5607366042492907,0.4315399786476907,0.5683656736948592,0.368515662598141,0.48444590979360486,0.3766168651200732,0.5111476528530949
27
+ 2,2,0.38130669695095903,0.5251342801699707,0.40915564373087976,0.5442069537838921,0.35426191276486857,0.481902886645082,0.3590649732014715,0.4895319560906506
28
+ 2,3,0.3739297080247543,0.5251342801699707,0.4002734843291297,0.5442069537838921,0.34880256321495257,0.48444590979360486,0.3562462201915426,0.4895319560906506
29
+ 2,4,0.36903808444060743,0.5340348611898006,0.39433300004424227,0.5442069537838921,0.3474310007462018,0.46283021303116056,0.3529159874498117,0.5086046297045721
30
+ 2,5,0.3689746418150539,0.5391209074868464,0.3928459818882813,0.5607366042492907,0.34833302328426086,0.48444590979360486,0.35672836200373836,0.4793598634965592
31
+ 2,6,0.3644825982238747,0.5340348611898006,0.38835767953935046,0.5442069537838921,0.3441595620205853,0.48444590979360486,0.35612515599686356,0.4793598634965592
32
+ 2,7,0.3680134685809739,0.5340348611898006,0.3897359699573702,0.5607366042492907,0.3428678288768968,0.46028718988263767,0.36286341124260063,0.4984325371104806
33
+ 2,8,0.37124534600314923,0.5391209074868464,0.39100675279982794,0.5607366042492907,0.3427617226566216,0.46028718988263767,0.3688273103216684,0.4984325371104806
34
+ 2,9,0.38324295679222353,0.5340348611898006,0.3969209910856405,0.5391209074868464,0.34226357145638214,0.4793598634965592,0.39176714087684517,0.5454784653581536
35
+ 2,10,0.3842442014968694,0.5047900949817878,0.3969264445585573,0.5454784653581536,0.33482066162409746,0.4653732361796834,0.4121296936265192,0.5594650926750293
36
+ 2,11,0.3632118937103562,0.39544009959530474,0.3870333552939945,0.4742738171995134,0.31799754141648634,0.41069823848644194,0.4186666226846874,0.5314918380412778
37
+ 2,-1,0.3632118937103562,0.39544009959530474,0.3870333552939945,0.4742738171995134,0.31799754141648634,0.41069823848644194,0.4186666226846874,0.5314918380412778
38
+ 3,1,0.3111888298433141,0.2199715023472274,0.35954385597428384,0.3916255648725205,0.2870738460172908,0.2682889421691617,0.35774010671758777,0.4043406806151348
39
+ 3,2,0.2419540364615831,0.0928203449210844,0.31542575280589946,0.24158719910967172,0.24516495628244855,0.18818371299069164,0.2309284827112575,0.09790639121813012
40
+ 3,3,0.21533914449446653,-0.01780116203966002,0.29714506149712255,0.20598487503035168,0.22957109399928316,0.11952208798057443,0.1802027621660931,0.005086046297045721
41
+ 3,4,0.21894101498250465,0.010172092594091442,0.2987756785579045,0.22378603707001168,0.23042313127675204,0.16148196993120162,0.1968777098802367,0.01907267361392145
42
+ 3,5,0.23318403521640468,0.0839197639012544,0.30530181511609183,0.24794475698097887,0.23648168896947866,0.16148196993120162,0.24061750950283073,0.15766743520841733
43
+ 3,6,0.25500866149505647,0.21742847919870456,0.314675011273474,0.3420366134763247,0.24612435271465996,0.24285871068393314,0.3016798381608026,0.31024882411978894
44
+ 3,7,0.28409792739925566,0.31660638199109614,0.3272050686737481,0.38908254172399764,0.25876871132233514,0.34839417134763184,0.3613573234124852,0.45774416673411483
45
+ 3,8,0.3093707912713613,0.4259563773775791,0.335941483665028,0.47300230562525203,0.27111008617957155,0.40052614589235047,0.3777382951716596,0.5569220695265065
46
+ 3,9,0.3215661093019967,0.5086046297045721,0.3397643404268252,0.5213197454471864,0.2790042446494514,0.5022470718332649,0.3605442691383628,0.5340348611898006
47
+ 3,10,0.32262543184581577,0.5086046297045721,0.3397201151555009,0.5086046297045721,0.28095861545145306,0.48317439821934344,0.3426404718744297,0.464101724605422
48
+ 3,11,0.32232495332808236,0.5086046297045721,0.3388176100863287,0.5086046297045721,0.27940496836348067,0.48317439821934344,0.33805397815457966,0.464101724605422
49
+ 3,-1,0.32232495332808236,0.5086046297045721,0.3388176100863287,0.5086046297045721,0.27940496836348067,0.48317439821934344,0.33805397815457966,0.464101724605422
50
+ 4,1,0.32121130265314746,0.5086046297045721,0.3387425248355813,0.5086046297045721,0.27797273569792375,0.48317439821934344,0.3355417406792243,0.464101724605422
51
+ 4,2,0.3150314335859403,0.5086046297045721,0.333839421649669,0.5086046297045721,0.2697850872748557,0.48698893294212775,0.3336188245578467,0.4450290509915006
52
+ 4,3,0.30981710710651356,0.5060616065560491,0.32965665356935386,0.4946180023876964,0.26153731952855197,0.4539296320113305,0.3335013438800938,0.5060616065560491
53
+ 4,4,0.3039611123047342,0.5060616065560491,0.32514140604922087,0.46155870145689915,0.253307526802248,0.42849940052610197,0.33494194779490677,0.5060616065560491
54
+ 4,5,0.3001453103011365,0.4208703310805334,0.32385429812293504,0.43358544682314765,0.24942931745973085,0.4310424236746248,0.3395370452138924,0.5060616065560491
55
+ 4,6,0.297724849905515,0.44248602784297775,0.324028951639797,0.455201143585592,0.24863769302406089,0.4170557963577491,0.3403293383867696,0.47554532877377487
56
+ 4,7,0.29743370160479327,0.44248602784297775,0.3266187885596849,0.455201143585592,0.2511816233567781,0.4170557963577491,0.3408391825931216,0.47554532877377487
57
+ 4,8,0.2997862095644924,0.44248602784297775,0.3307142055154821,0.455201143585592,0.25581264878508203,0.4310424236746248,0.3400143115278199,0.47554532877377487
58
+ 4,9,0.3076778629319174,0.44248602784297775,0.3383757332801397,0.455201143585592,0.26459317021905576,0.45647265515985347,0.3428054664517925,0.5060616065560491
59
+ 4,10,0.31187775578710814,0.5086046297045721,0.34253855501926966,0.464101724605422,0.2699934453225055,0.4539296320113305,0.3345121000324598,0.5060616065560491
60
+ 4,11,0.31416307370270197,0.47300230562525203,0.3451948397052399,0.48698893294212775,0.275247778811942,0.4920749792391734,0.3228218029092112,0.4450290509915006
61
+ 4,-1,0.31416307370270197,0.47300230562525203,0.3451948397052399,0.48698893294212775,0.275247778811942,0.4920749792391734,0.3228218029092112,0.4450290509915006
62
+ 5,1,0.3102839559506583,0.4946180023876964,0.34278302618290774,0.48698893294212775,0.2758412046841528,0.4704592824767292,0.3126085296311749,0.4246848658033177
63
+ 5,2,0.30785679545153727,0.5162336991501406,0.34176580398617595,0.4920749792391734,0.2779343165643041,0.4691877709024677,0.306441802097247,0.4132412616349648
64
+ 5,3,0.30788792556249367,0.5403924190611078,0.342979019862447,0.4920749792391734,0.2816436294912663,0.4691877709024677,0.3046498287662398,0.4132412616349648
65
+ 5,4,0.30764645417782843,0.5213197454471864,0.34389763801140844,0.5022470718332649,0.2839629847447862,0.5009755602590035,0.30430402180175276,0.4437575394172391
66
+ 5,5,0.3068650196825221,0.5264057917442321,0.34458184923389445,0.48317439821934344,0.2840154468464772,0.4793598634965592,0.3069239845788843,0.4437575394172391
67
+ 5,6,0.30375134805779846,0.4793598634965592,0.34252170885500144,0.4806313750708206,0.280140767518872,0.4806313750708206,0.3107008361117029,0.4437575394172391
68
+ 5,7,0.30172518529388564,0.4806313750708206,0.34212853718753344,0.4971610255362192,0.27792099853238755,0.5022470718332649,0.3166386769871796,0.5047900949817878
69
+ 5,8,0.3013732765676238,0.5187767222986636,0.3431951718359202,0.4984325371104806,0.27898600340268476,0.4857174213678663,0.32299271902064913,0.5047900949817878
70
+ 5,9,0.301903030108604,0.4984325371104806,0.34481957946617575,0.4984325371104806,0.28211497654202067,0.4857174213678663,0.3245655526620992,0.5047900949817878
71
+ 5,10,0.29784728781874126,0.455201143585592,0.3421494542049185,0.4984325371104806,0.28030196619244124,0.464101724605422,0.3214684218507993,0.5136906760016178
72
+ 5,11,0.2924067173725062,0.455201143585592,0.33818424391033375,0.45774416673411483,0.27645599828422546,0.47681684034803634,0.31718204452014387,0.5136906760016178
73
+ 5,-1,0.2924067173725062,0.455201143585592,0.33818424391033375,0.45774416673411483,0.27645599828422546,0.47681684034803634,0.31718204452014387,0.5136906760016178
74
+ 6,1,0.2885272628115487,0.45774416673411483,0.33572074256475165,0.45774416673411483,0.27435471918898446,0.455201143585592,0.3133510811896758,0.4895319560906506
75
+ 6,2,0.2839215576248625,0.4361284699716705,0.3327563173403644,0.45774416673411483,0.2720666876443617,0.455201143585592,0.30887226022966174,0.4895319560906506
76
+ 6,3,0.2789118587917517,0.43358544682314765,0.32942078266888664,0.45774416673411483,0.26911163663813015,0.455201143585592,0.3042959785944026,0.4895319560906506
77
+ 6,4,0.2713794342602401,0.455201143585592,0.3236492930158097,0.45774416673411483,0.26382760016946843,0.455201143585592,0.2978919687469969,0.4895319560906506
78
+ 6,5,0.2653344593401437,0.455201143585592,0.3185785370837683,0.455201143585592,0.2588977754203019,0.455201143585592,0.29337578393697,0.4895319560906506
79
+ 6,6,0.25860096044329306,0.42977091210036333,0.3126960552598408,0.455201143585592,0.25323105924647943,0.455201143585592,0.28841827647780904,0.4895319560906506
80
+ 6,7,0.25494369272426254,0.42977091210036333,0.31002942346253,0.455201143585592,0.25094996210788423,0.455201143585592,0.2858483927654416,0.5136906760016178
81
+ 6,8,0.25278258006004317,0.4513866088628077,0.3091028066237646,0.47681684034803634,0.25086173127450123,0.464101724605422,0.2842456554623047,0.5136906760016178
82
+ 6,9,0.2519209219536299,0.43867149312019343,0.30923074571401116,0.4984325371104806,0.25192798733599175,0.4857174213678663,0.28342479535392673,0.5136906760016178
83
+ 6,10,0.25161480991859025,0.5187767222986636,0.3100105749727334,0.4984325371104806,0.2539677883419662,0.4691877709024677,0.2822594927705823,0.5264057917442321
84
+ 6,11,0.2520542219905743,0.5022470718332649,0.311358796906959,0.4857174213678663,0.256618010116366,0.5022470718332649,0.28190673498746627,0.5264057917442321
85
+ 6,-1,0.2520542219905743,0.5022470718332649,0.311358796906959,0.4857174213678663,0.256618010116366,0.5022470718332649,0.28190673498746627,0.5264057917442321
86
+ 7,1,0.252853737471713,0.5022470718332649,0.3130995454750759,0.5187767222986636,0.25962273610779857,0.5022470718332649,0.28174910338406794,0.5480214885066763
87
+ 7,2,0.2525904867321626,0.5022470718332649,0.31360745394344347,0.5022470718332649,0.2613820590234466,0.5022470718332649,0.28073470579951476,0.517505210724402
88
+ 7,3,0.2533319312962192,0.5022470718332649,0.3147722924247869,0.5022470718332649,0.26366501674656795,0.5022470718332649,0.2806905949272515,0.48698893294212775
89
+ 7,4,0.2528235539472975,0.5047900949817878,0.3146303537598034,0.5022470718332649,0.26441104691811695,0.5022470718332649,0.2799372009722368,0.45647265515985347
90
+ 7,5,0.25177495483365187,0.5047900949817878,0.3140981144164873,0.5022470718332649,0.2645461466343892,0.5022470718332649,0.2789157036119476,0.45647265515985347
91
+ 7,6,0.25082796918212125,0.5047900949817878,0.31379295959200415,0.5022470718332649,0.2647127529372475,0.5022470718332649,0.27817460522235077,0.45647265515985347
92
+ 7,7,0.2516625485028545,0.5047900949817878,0.31490064303698495,0.5022470718332649,0.26631251506820286,0.5022470718332649,0.2784941627377695,0.45647265515985347
93
+ 7,8,0.2516683884980386,0.5047900949817878,0.3150549230320982,0.5238627685957092,0.26659219932258077,0.5022470718332649,0.27815999467448366,0.45647265515985347
94
+ 7,9,0.25155198887856295,0.5047900949817878,0.3150867499736677,0.5238627685957092,0.2668214386239413,0.5022470718332649,0.27773351942635166,0.45647265515985347
95
+ 7,10,0.2506801712735185,0.5047900949817878,0.3144852967721268,0.5238627685957092,0.26642053310779606,0.5022470718332649,0.27671958501945826,0.45647265515985347
96
+ 7,11,0.2526886486676884,0.5047900949817878,0.31667795354885314,0.5022470718332649,0.26835647615605085,0.5022470718332649,0.27931100054136254,0.45647265515985347
97
+ 7,-1,0.2526886486676884,0.5047900949817878,0.31667795354885314,0.5022470718332649,0.26835647615605085,0.5022470718332649,0.27931100054136254,0.45647265515985347
98
+ 8,1,0.25336299082208047,0.5047900949817878,0.3175068286474958,0.5022470718332649,0.2689290668764881,0.5022470718332649,0.2807662199584674,0.45647265515985347
99
+ 8,2,0.25392667453315215,0.5022470718332649,0.3181358576326492,0.5022470718332649,0.26921636414376376,0.5022470718332649,0.28196707623751416,0.48698893294212775
100
+ 8,3,0.2549366265566119,0.5022470718332649,0.319212784539419,0.5022470718332649,0.2700095196507663,0.5022470718332649,0.2834836358933535,0.517505210724402
101
+ 8,4,0.25655646871462545,0.5022470718332649,0.32075349557848165,0.5022470718332649,0.27120337892895047,0.5022470718332649,0.2852831603821673,0.5556505579522449
102
+ 8,5,0.2576118825801414,0.5022470718332649,0.32173478175340164,0.5187767222986636,0.271878908949732,0.5022470718332649,0.2866275044044628,0.5556505579522449
103
+ 8,6,0.2578220190686094,0.5022470718332649,0.32183276070068806,0.5187767222986636,0.2716865371372322,0.5022470718332649,0.2873792323026814,0.5556505579522449
104
+ 8,7,0.2577034078580955,0.5022470718332649,0.32170828637273885,0.5187767222986636,0.27134303464940707,0.5022470718332649,0.2877537546038792,0.5556505579522449
105
+ 8,8,0.2577978571182213,0.5022470718332649,0.3218107639639885,0.5187767222986636,0.2713240858744541,0.5022470718332649,0.28822584027234266,0.5556505579522449
106
+ 8,9,0.25796235006692486,0.5022470718332649,0.32179829548742406,0.5187767222986636,0.27108433820159644,0.5022470718332649,0.28872055711154043,0.5556505579522449
107
+ 8,10,0.25830342372212095,0.5022470718332649,0.32204121536732944,0.5187767222986636,0.27132017848236295,0.5022470718332649,0.2891789074582076,0.5556505579522449
108
+ 8,11,0.25861831672005564,0.5187767222986636,0.32224737855047736,0.5187767222986636,0.27153304948941553,0.5022470718332649,0.289579519022339,0.5556505579522449
109
+ 8,-1,0.25861831672005564,0.5187767222986636,0.32224737855047736,0.5187767222986636,0.27153304948941553,0.5022470718332649,0.289579519022339,0.5556505579522449
110
+ 9,1,0.25917467417266016,0.5187767222986636,0.3226657561959536,0.5187767222986636,0.27198344346750136,0.5022470718332649,0.2901332626089178,0.5556505579522449
111
+ 9,2,0.2593984926333763,0.5187767222986636,0.32280646240246624,0.5187767222986636,0.272146897334996,0.5022470718332649,0.2903804534182478,0.5556505579522449
112
+ 9,3,0.2602795994392014,0.4971610255362192,0.3234911177024182,0.5187767222986636,0.2728868627389731,0.5022470718332649,0.29109565302509655,0.5556505579522449
113
+ 9,4,0.2610263806240683,0.4971610255362192,0.3240651044845065,0.5187767222986636,0.2735578735200072,0.5022470718332649,0.29166175368723535,0.5556505579522449
114
+ 9,5,0.2616140029270722,0.4971610255362192,0.32449540339662253,0.5187767222986636,0.27402993328530306,0.5022470718332649,0.2920729689892911,0.5556505579522449
115
+ 9,6,0.2616990794503368,0.4971610255362192,0.3244993922965195,0.5187767222986636,0.27411172948473056,0.5022470718332649,0.29207657691253347,0.5556505579522449
116
+ 9,7,0.261612501781844,0.4971610255362192,0.32437973602156184,0.5187767222986636,0.27408277630134514,0.5022470718332649,0.29193268164884345,0.5556505579522449
117
+ 9,8,0.26158913004769074,0.4971610255362192,0.32434716537211994,0.5187767222986636,0.2741056313091124,0.5022470718332649,0.2918122770300895,0.5556505579522449
118
+ 9,9,0.26139726466100244,0.4971610255362192,0.32417246829045276,0.5187767222986636,0.27397146957750906,0.5022470718332649,0.2916065033835955,0.5556505579522449
119
+ 9,10,0.2612411945254866,0.4971610255362192,0.3240027432086403,0.5187767222986636,0.2738070994474979,0.5022470718332649,0.2914661927834061,0.5556505579522449
120
+ 9,11,0.261143560239595,0.4971610255362192,0.32391792559873595,0.5187767222986636,0.27373785138725176,0.5022470718332649,0.2913784653460315,0.5556505579522449
121
+ 9,-1,0.261143560239595,0.4971610255362192,0.32391792559873595,0.5187767222986636,0.27373785138725176,0.5022470718332649,0.2913784653460315,0.5556505579522449
modules.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ },
14
+ {
15
+ "idx": 2,
16
+ "name": "2",
17
+ "path": "2_Dense",
18
+ "type": "sentence_transformers.models.Dense"
19
+ }
20
+ ]
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bdb41d776ed6ee983b7f8cdbc7b7e6f636303f8f49df7bd606d296faba5adf54
3
+ size 439484849
sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 256,
3
+ "do_lower_case": false
4
+ }
similarity_evaluation_sts-test_results.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ epoch,steps,cosine_pearson,cosine_spearman,euclidean_pearson,euclidean_spearman,manhattan_pearson,manhattan_spearman,dot_pearson,dot_spearman
2
+ -1,-1,0.7840062516092543,0.4338212432092088,0.8007057948669679,0.509369451168439,0.8039624144333989,0.5113845033919161,0.7155110886983155,0.3179557403592815
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"do_lower_case": true, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": false, "do_basic_tokenize": true, "never_split": null, "model_max_length": 512, "special_tokens_map_file": "/root/.cache/huggingface/transformers/78141ed1e8dcc5ff370950397ca0d1c5c9da478f54ec14544187d8a93eff1a26.f982506b52498d4adb4bd491f593dc92b2ef6be61bfdbe9d30f53f963f9f5b66", "name_or_path": "dccuchile/bert-base-spanish-wwm-uncased", "tokenizer_class": "BertTokenizer"}
vocab.txt ADDED
The diff for this file is too large to render. See raw diff