xiomarablanco commited on
Commit
5716ce5
1 Parent(s): 54d9e93

ModelosPreentrenados Bert

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +10 -0
  2. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/100_Epochs/1_Pooling/config.json +7 -0
  3. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/100_Epochs/README.md +87 -0
  4. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/100_Epochs/config.json +28 -0
  5. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/100_Epochs/config_sentence_transformers.json +7 -0
  6. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/100_Epochs/eval/similarity_evaluation_results.csv +0 -0
  7. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/100_Epochs/merges.txt +0 -0
  8. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/100_Epochs/modules.json +20 -0
  9. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/100_Epochs/pytorch_model.bin +3 -0
  10. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/100_Epochs/sentence_bert_config.json +4 -0
  11. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/100_Epochs/special_tokens_map.json +1 -0
  12. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/100_Epochs/tokenizer.json +0 -0
  13. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/100_Epochs/tokenizer_config.json +1 -0
  14. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/100_Epochs/vocab.json +0 -0
  15. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/10_Epochs/1_Pooling/config.json +7 -0
  16. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/10_Epochs/README.md +87 -0
  17. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/10_Epochs/config.json +28 -0
  18. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/10_Epochs/config_sentence_transformers.json +7 -0
  19. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/10_Epochs/eval/similarity_evaluation_results.csv +121 -0
  20. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/10_Epochs/merges.txt +0 -0
  21. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/10_Epochs/modules.json +20 -0
  22. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/10_Epochs/pytorch_model.bin +3 -0
  23. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/10_Epochs/sentence_bert_config.json +4 -0
  24. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/10_Epochs/special_tokens_map.json +1 -0
  25. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/10_Epochs/tokenizer.json +0 -0
  26. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/10_Epochs/tokenizer_config.json +1 -0
  27. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/10_Epochs/vocab.json +0 -0
  28. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/1_Epochs/1_Pooling/config.json +7 -0
  29. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/1_Epochs/README.md +87 -0
  30. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/1_Epochs/config.json +28 -0
  31. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/1_Epochs/config_sentence_transformers.json +7 -0
  32. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/1_Epochs/eval/similarity_evaluation_results.csv +13 -0
  33. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/1_Epochs/merges.txt +0 -0
  34. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/1_Epochs/modules.json +20 -0
  35. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/1_Epochs/pytorch_model.bin +3 -0
  36. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/1_Epochs/sentence_bert_config.json +4 -0
  37. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/1_Epochs/special_tokens_map.json +1 -0
  38. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/1_Epochs/tokenizer.json +0 -0
  39. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/1_Epochs/tokenizer_config.json +1 -0
  40. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/1_Epochs/vocab.json +0 -0
  41. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/30_Epochs/1_Pooling/config.json +7 -0
  42. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/30_Epochs/README.md +87 -0
  43. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/30_Epochs/config.json +28 -0
  44. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/30_Epochs/config_sentence_transformers.json +7 -0
  45. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/30_Epochs/eval/similarity_evaluation_results.csv +361 -0
  46. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/30_Epochs/merges.txt +0 -0
  47. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/30_Epochs/modules.json +20 -0
  48. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/30_Epochs/pytorch_model.bin +3 -0
  49. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/30_Epochs/sentence_bert_config.json +4 -0
  50. codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/30_Epochs/special_tokens_map.json +1 -0
.gitattributes CHANGED
@@ -32,3 +32,13 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
35
+ codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_paraphrase-multilingual-MiniLM-L12-v2/1_Epochs/tokenizer.json filter=lfs diff=lfs merge=lfs -text
36
+ codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_paraphrase-multilingual-MiniLM-L12-v2/1_Epochs/unigram.json filter=lfs diff=lfs merge=lfs -text
37
+ codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_paraphrase-multilingual-MiniLM-L12-v2/10_Epochs/tokenizer.json filter=lfs diff=lfs merge=lfs -text
38
+ codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_paraphrase-multilingual-MiniLM-L12-v2/10_Epochs/unigram.json filter=lfs diff=lfs merge=lfs -text
39
+ codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_paraphrase-multilingual-MiniLM-L12-v2/100_Epochs/tokenizer.json filter=lfs diff=lfs merge=lfs -text
40
+ codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_paraphrase-multilingual-MiniLM-L12-v2/100_Epochs/unigram.json filter=lfs diff=lfs merge=lfs -text
41
+ codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_paraphrase-multilingual-MiniLM-L12-v2/30_Epochs/tokenizer.json filter=lfs diff=lfs merge=lfs -text
42
+ codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_paraphrase-multilingual-MiniLM-L12-v2/30_Epochs/unigram.json filter=lfs diff=lfs merge=lfs -text
43
+ codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_paraphrase-multilingual-MiniLM-L12-v2/5_Epochs/tokenizer.json filter=lfs diff=lfs merge=lfs -text
44
+ codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_paraphrase-multilingual-MiniLM-L12-v2/5_Epochs/unigram.json filter=lfs diff=lfs merge=lfs -text
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/100_Epochs/1_Pooling/config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 768,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": true,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false
7
+ }
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/100_Epochs/README.md ADDED
@@ -0,0 +1,87 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ pipeline_tag: sentence-similarity
3
+ tags:
4
+ - sentence-transformers
5
+ - feature-extraction
6
+ - sentence-similarity
7
+ ---
8
+
9
+ # {MODEL_NAME}
10
+
11
+ This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.
12
+
13
+ <!--- Describe your model here -->
14
+
15
+ ## Usage (Sentence-Transformers)
16
+
17
+ Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed:
18
+
19
+ ```
20
+ pip install -U sentence-transformers
21
+ ```
22
+
23
+ Then you can use the model like this:
24
+
25
+ ```python
26
+ from sentence_transformers import SentenceTransformer
27
+ sentences = ["This is an example sentence", "Each sentence is converted"]
28
+
29
+ model = SentenceTransformer('{MODEL_NAME}')
30
+ embeddings = model.encode(sentences)
31
+ print(embeddings)
32
+ ```
33
+
34
+
35
+
36
+ ## Evaluation Results
37
+
38
+ <!--- Describe how your model was evaluated -->
39
+
40
+ For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name={MODEL_NAME})
41
+
42
+
43
+ ## Training
44
+ The model was trained with the parameters:
45
+
46
+ **DataLoader**:
47
+
48
+ `torch.utils.data.dataloader.DataLoader` of length 11 with parameters:
49
+ ```
50
+ {'batch_size': 15, 'sampler': 'torch.utils.data.sampler.RandomSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'}
51
+ ```
52
+
53
+ **Loss**:
54
+
55
+ `sentence_transformers.losses.CosineSimilarityLoss.CosineSimilarityLoss`
56
+
57
+ Parameters of the fit()-Method:
58
+ ```
59
+ {
60
+ "epochs": 100,
61
+ "evaluation_steps": 1,
62
+ "evaluator": "sentence_transformers.evaluation.EmbeddingSimilarityEvaluator.EmbeddingSimilarityEvaluator",
63
+ "max_grad_norm": 1,
64
+ "optimizer_class": "<class 'transformers.optimization.AdamW'>",
65
+ "optimizer_params": {
66
+ "lr": 2e-05
67
+ },
68
+ "scheduler": "WarmupLinear",
69
+ "steps_per_epoch": null,
70
+ "warmup_steps": 110,
71
+ "weight_decay": 0.01
72
+ }
73
+ ```
74
+
75
+
76
+ ## Full Model Architecture
77
+ ```
78
+ SentenceTransformer(
79
+ (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: RobertaModel
80
+ (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False})
81
+ (2): Normalize()
82
+ )
83
+ ```
84
+
85
+ ## Citing & Authors
86
+
87
+ <!--- Describe where people can find more information -->
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/100_Epochs/config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "C:\\Users\\javier.sanz/.cache\\torch\\sentence_transformers\\sentence-transformers_all-distilroberta-v1\\",
3
+ "architectures": [
4
+ "RobertaModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
+ "gradient_checkpointing": false,
11
+ "hidden_act": "gelu",
12
+ "hidden_dropout_prob": 0.1,
13
+ "hidden_size": 768,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 3072,
16
+ "layer_norm_eps": 1e-05,
17
+ "max_position_embeddings": 514,
18
+ "model_type": "roberta",
19
+ "num_attention_heads": 12,
20
+ "num_hidden_layers": 6,
21
+ "pad_token_id": 1,
22
+ "position_embedding_type": "absolute",
23
+ "torch_dtype": "float32",
24
+ "transformers_version": "4.18.0",
25
+ "type_vocab_size": 1,
26
+ "use_cache": true,
27
+ "vocab_size": 50265
28
+ }
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/100_Epochs/config_sentence_transformers.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "2.0.0",
4
+ "transformers": "4.6.1",
5
+ "pytorch": "1.8.1"
6
+ }
7
+ }
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/100_Epochs/eval/similarity_evaluation_results.csv ADDED
The diff for this file is too large to render. See raw diff
 
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/100_Epochs/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/100_Epochs/modules.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ },
14
+ {
15
+ "idx": 2,
16
+ "name": "2",
17
+ "path": "2_Normalize",
18
+ "type": "sentence_transformers.models.Normalize"
19
+ }
20
+ ]
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/100_Epochs/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06c353a1a8f288de7838253bd1079d342c288c8d945b093d7f2aa24b9e85052d
3
+ size 328509745
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/100_Epochs/sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 512,
3
+ "do_lower_case": false
4
+ }
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/100_Epochs/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "sep_token": "</s>", "pad_token": "<pad>", "cls_token": "<s>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": false}}
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/100_Epochs/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/100_Epochs/tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"errors": "replace", "bos_token": "<s>", "eos_token": "</s>", "sep_token": "</s>", "cls_token": "<s>", "unk_token": "<unk>", "pad_token": "<pad>", "mask_token": "<mask>", "add_prefix_space": false, "trim_offsets": true, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "C:\\Users\\javier.sanz/.cache\\torch\\sentence_transformers\\sentence-transformers_all-distilroberta-v1\\", "tokenizer_class": "RobertaTokenizer"}
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/100_Epochs/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/10_Epochs/1_Pooling/config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 768,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": true,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false
7
+ }
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/10_Epochs/README.md ADDED
@@ -0,0 +1,87 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ pipeline_tag: sentence-similarity
3
+ tags:
4
+ - sentence-transformers
5
+ - feature-extraction
6
+ - sentence-similarity
7
+ ---
8
+
9
+ # {MODEL_NAME}
10
+
11
+ This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.
12
+
13
+ <!--- Describe your model here -->
14
+
15
+ ## Usage (Sentence-Transformers)
16
+
17
+ Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed:
18
+
19
+ ```
20
+ pip install -U sentence-transformers
21
+ ```
22
+
23
+ Then you can use the model like this:
24
+
25
+ ```python
26
+ from sentence_transformers import SentenceTransformer
27
+ sentences = ["This is an example sentence", "Each sentence is converted"]
28
+
29
+ model = SentenceTransformer('{MODEL_NAME}')
30
+ embeddings = model.encode(sentences)
31
+ print(embeddings)
32
+ ```
33
+
34
+
35
+
36
+ ## Evaluation Results
37
+
38
+ <!--- Describe how your model was evaluated -->
39
+
40
+ For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name={MODEL_NAME})
41
+
42
+
43
+ ## Training
44
+ The model was trained with the parameters:
45
+
46
+ **DataLoader**:
47
+
48
+ `torch.utils.data.dataloader.DataLoader` of length 11 with parameters:
49
+ ```
50
+ {'batch_size': 15, 'sampler': 'torch.utils.data.sampler.RandomSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'}
51
+ ```
52
+
53
+ **Loss**:
54
+
55
+ `sentence_transformers.losses.CosineSimilarityLoss.CosineSimilarityLoss`
56
+
57
+ Parameters of the fit()-Method:
58
+ ```
59
+ {
60
+ "epochs": 10,
61
+ "evaluation_steps": 1,
62
+ "evaluator": "sentence_transformers.evaluation.EmbeddingSimilarityEvaluator.EmbeddingSimilarityEvaluator",
63
+ "max_grad_norm": 1,
64
+ "optimizer_class": "<class 'transformers.optimization.AdamW'>",
65
+ "optimizer_params": {
66
+ "lr": 2e-05
67
+ },
68
+ "scheduler": "WarmupLinear",
69
+ "steps_per_epoch": null,
70
+ "warmup_steps": 11,
71
+ "weight_decay": 0.01
72
+ }
73
+ ```
74
+
75
+
76
+ ## Full Model Architecture
77
+ ```
78
+ SentenceTransformer(
79
+ (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: RobertaModel
80
+ (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False})
81
+ (2): Normalize()
82
+ )
83
+ ```
84
+
85
+ ## Citing & Authors
86
+
87
+ <!--- Describe where people can find more information -->
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/10_Epochs/config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "C:\\Users\\javier.sanz/.cache\\torch\\sentence_transformers\\sentence-transformers_all-distilroberta-v1\\",
3
+ "architectures": [
4
+ "RobertaModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
+ "gradient_checkpointing": false,
11
+ "hidden_act": "gelu",
12
+ "hidden_dropout_prob": 0.1,
13
+ "hidden_size": 768,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 3072,
16
+ "layer_norm_eps": 1e-05,
17
+ "max_position_embeddings": 514,
18
+ "model_type": "roberta",
19
+ "num_attention_heads": 12,
20
+ "num_hidden_layers": 6,
21
+ "pad_token_id": 1,
22
+ "position_embedding_type": "absolute",
23
+ "torch_dtype": "float32",
24
+ "transformers_version": "4.18.0",
25
+ "type_vocab_size": 1,
26
+ "use_cache": true,
27
+ "vocab_size": 50265
28
+ }
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/10_Epochs/config_sentence_transformers.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "2.0.0",
4
+ "transformers": "4.6.1",
5
+ "pytorch": "1.8.1"
6
+ }
7
+ }
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/10_Epochs/eval/similarity_evaluation_results.csv ADDED
@@ -0,0 +1,121 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch,steps,cosine_pearson,cosine_spearman,euclidean_pearson,euclidean_spearman,manhattan_pearson,manhattan_spearman,dot_pearson,dot_spearman
2
+ 0,1,0.3079180908463921,0.2821978077379946,0.2862899581249616,0.2821978077379946,0.282337709193886,0.2821978077379946,0.3079178129162619,0.2821978077379946
3
+ 0,2,0.31131479303919996,0.2945208561108328,0.2897132342977788,0.2945208561108328,0.2865662619784499,0.2821978077379946,0.3113147970308761,0.2945208561108328
4
+ 0,3,0.3284127944014854,0.3216315625310768,0.306575736940227,0.3216315625310768,0.30614511241493003,0.26741014969058874,0.3284133418623011,0.3216315625310768
5
+ 0,4,0.352597505115391,0.2600163206668858,0.3295261315787658,0.2600163206668858,0.3325837949223793,0.25508710131775053,0.35259806564092944,0.2600163206668858
6
+ 0,5,0.3551597157583929,0.25878401582960203,0.33152392741063885,0.25878401582960203,0.33429619743211614,0.2661778448533049,0.3551596753533005,0.25878401582960203
7
+ 0,6,0.35043772124942996,0.2538547964804667,0.3261282942159988,0.2538547964804667,0.328972015248802,0.24153174810762854,0.35043773697792024,0.2538547964804667
8
+ 0,7,0.3531437394117739,0.2267440900602227,0.32868948525340524,0.2267440900602227,0.33580512934451684,0.2439963577821962,0.3531442858569088,0.2267440900602227
9
+ 0,8,0.3447260758971231,0.23660252875849325,0.32081523051112093,0.23660252875849325,0.3297170549202325,0.25878401582960203,0.34472556072014954,0.23660252875849325
10
+ 0,9,0.3400983337168123,0.2242794803856551,0.3153422447001065,0.2242794803856551,0.3229377274838364,0.25139018680589914,0.3400975818903731,0.2242794803856551
11
+ 0,10,0.3657602332490278,0.2255117852229389,0.3369792626648629,0.2255117852229389,0.3492117265431015,0.2057949078263978,0.36575903801861387,0.2255117852229389
12
+ 0,11,0.3709162092808649,0.20333029815183015,0.33930640130306977,0.20333029815183015,0.347661614646439,0.19593646912812723,0.37091667459884575,0.20333029815183015
13
+ 0,-1,0.3709162092808649,0.20333029815183015,0.33930640130306977,0.20333029815183015,0.347661614646439,0.19593646912812723,0.37091667459884575,0.20333029815183015
14
+ 1,1,0.3846267052099176,0.2057949078263978,0.3488563629531395,0.2057949078263978,0.35375435610605543,0.25015788196861527,0.38462619476770255,0.2057949078263978
15
+ 1,2,0.37073446015334616,0.2242794803856551,0.33306260007825195,0.2242794803856551,0.33403133354680953,0.2760362835515755,0.3707342145965276,0.2242794803856551
16
+ 1,3,0.3536039195552287,0.24769327229404767,0.3173715461099598,0.24769327229404767,0.31989608585750245,0.30068238029725186,0.35360355357258305,0.24769327229404767
17
+ 1,4,0.33182872393638296,0.24153174810762854,0.29845257206420167,0.24153174810762854,0.30119560386902877,0.2908239415989813,0.3318291639287195,0.24153174810762854
18
+ 1,5,0.32684956923117303,0.2563194061550344,0.294157393732839,0.2563194061550344,0.2981503980720115,0.2834301125752784,0.32684895030135946,0.2563194061550344
19
+ 1,6,0.3068446442240446,0.23044100457207417,0.2768351518249034,0.23044100457207417,0.28132523992510866,0.23413791908392564,0.30684456178720926,0.23044100457207417
20
+ 1,7,0.2964842076711062,0.20456260298911397,0.26797612013510297,0.20456260298911397,0.2738707736137515,0.22920869973479036,0.29648242602080066,0.20456260298911397
21
+ 1,8,0.3115156071982348,0.20456260298911397,0.2809008759950055,0.20456260298911397,0.2921234856991086,0.23660252875849325,0.311516730638835,0.20456260298911397
22
+ 1,9,0.32960365818103643,0.21811795619923596,0.29743691028586855,0.21811795619923596,0.31437864845225805,0.22304717554837125,0.32960341546599725,0.21811795619923596
23
+ 1,10,0.3400217906437536,0.21688565136195218,0.30755301319767436,0.21688565136195218,0.32560397915956985,0.20086568847726252,0.34002177585732496,0.21688565136195218
24
+ 1,11,0.34493583660534327,0.13801814177578775,0.3143195835384497,0.13801814177578775,0.3273562231308115,0.22181487071108744,0.344935205256093,0.13801814177578775
25
+ 1,-1,0.34493583660534327,0.13801814177578775,0.3143195835384497,0.13801814177578775,0.3273562231308115,0.22181487071108744,0.344935205256093,0.13801814177578775
26
+ 2,1,0.3290941382331395,0.16512884819603177,0.3015199550719144,0.16512884819603177,0.30612623906652475,0.1786842014061538,0.3290948794943809,0.16512884819603177
27
+ 2,2,0.30874677060809075,0.1614319336841803,0.2805953017542121,0.1614319336841803,0.2795174384301422,0.1367858369385039,0.3087465498652837,0.1614319336841803
28
+ 2,3,0.27634225202052937,0.11583665470467899,0.2473810144843208,0.11583665470467899,0.2421328935267727,0.08872594828443497,0.2763420804560257,0.11583665470467899
29
+ 2,4,0.23042344150597516,0.07393829023702915,0.20294272180750733,0.07393829023702915,0.19938188308403962,0.06900907088789388,0.23042326307869213,0.07393829023702915
30
+ 2,5,0.2027165474387368,0.08256442409801588,0.17407724185706097,0.08256442409801588,0.17603853708715492,0.04559527897950131,0.20271680736177852,0.08256442409801588
31
+ 2,6,0.1833949866921096,0.050524498328636584,0.15322669538243508,0.050524498328636584,0.16033050019445041,0.04559527897950131,0.1833954990077173,0.050524498328636584
32
+ 2,7,0.18349338193069412,0.029575316094811656,0.1505843180237409,0.029575316094811656,0.16164008065743868,0.040666059630366035,0.18349392758265198,0.029575316094811656
33
+ 2,8,0.1811588346984214,-0.019716877396541106,0.14558797949192115,-0.019716877396541106,0.16063119767392472,0.01725226772197347,0.18115917973506274,-0.019716877396541106
34
+ 2,9,0.18129587568711095,-0.0024646096745676383,0.14246304346810723,-0.0024646096745676383,0.16159974424601578,0.0024646096745676383,0.18129590586839714,-0.0024646096745676383
35
+ 2,10,0.1843968927395075,-0.0024646096745676383,0.14212025332903755,-0.0024646096745676383,0.1631775341294583,-0.0024646096745676383,0.18439677392437726,-0.0024646096745676383
36
+ 2,11,0.196213652490966,0.008626133860986734,0.14880423490022607,0.008626133860986734,0.17130013080206055,0.03943375479308221,0.19621341197323858,0.008626133860986734
37
+ 2,-1,0.196213652490966,0.008626133860986734,0.14880423490022607,0.008626133860986734,0.17130013080206055,0.03943375479308221,0.19621341197323858,0.008626133860986734
38
+ 3,1,0.20315060596476525,0.03450453544394694,0.15300531325225597,0.03450453544394694,0.1759004248316642,0.06161524186419096,0.20315045271115703,0.03450453544394694
39
+ 3,2,0.20487683177107932,0.007393829023702914,0.15408833654609616,0.007393829023702914,0.1763450965452999,0.022181487071108744,0.20487774504379508,0.007393829023702914
40
+ 3,3,0.2030525883591865,0.019716877396541106,0.15276177845740274,0.019716877396541106,0.1736070056321622,0.033272230606663115,0.20305454553931945,0.019716877396541106
41
+ 3,4,0.1943801808705284,0.008626133860986734,0.1452041945304844,0.008626133860986734,0.16421607599146043,0.03820144995579839,0.19438061255905734,0.008626133860986734
42
+ 3,5,0.18053823250258139,-0.007393829023702914,0.13344439792817692,-0.007393829023702914,0.14923017187172433,-0.024646096745676385,0.18053946827932113,-0.007393829023702914
43
+ 3,6,0.16781535629960348,-0.051756803165920406,0.12327986193445309,-0.051756803165920406,0.1353568148490294,-0.04929219349135277,0.16781581981480093,-0.051756803165920406
44
+ 3,7,0.1571019518243144,-0.06900907088789388,0.11576185453176037,-0.06900907088789388,0.12511108462876153,-0.040666059630366035,0.1571016908407938,-0.06900907088789388
45
+ 3,8,0.1495594515204336,-0.04189836446764985,0.11122467170220865,-0.04189836446764985,0.11826749711560043,-0.023413791908392562,0.14955883270506837,-0.04189836446764985
46
+ 3,9,0.13708347414177413,-0.013555353210122011,0.10394442585743288,-0.013555353210122011,0.10723454932601731,-0.03080762093209548,0.1370844845959746,-0.013555353210122011
47
+ 3,10,0.1265556346875377,0.011090743535554372,0.09791655074792965,0.011090743535554372,0.0971255293751095,-0.011090743535554372,0.12655487903661417,0.011090743535554372
48
+ 3,11,0.1227801929926005,0.009858438698270553,0.09795447809989949,0.009858438698270553,0.09284981845260526,-0.040666059630366035,0.12278063644585543,0.009858438698270553
49
+ 3,-1,0.1227801929926005,0.009858438698270553,0.09795447809989949,0.009858438698270553,0.09284981845260526,-0.040666059630366035,0.12278063644585543,0.009858438698270553
50
+ 4,1,0.12364276328034653,-0.04559527897950131,0.09918162255481067,-0.04559527897950131,0.09296931148752147,-0.04559527897950131,0.12364220735619064,-0.04559527897950131
51
+ 4,2,0.11941501189834736,-0.040666059630366035,0.09570144614125387,-0.040666059630366035,0.0876425353403244,-0.03080762093209548,0.11941555038687245,-0.040666059630366035
52
+ 4,3,0.11492765873193739,-0.014787658047405828,0.09465793004056601,-0.014787658047405828,0.08608407413377396,-0.0320399257693793,0.11492873594465024,-0.014787658047405828
53
+ 4,4,0.10545533396300369,-0.009858438698270553,0.0872183770567984,-0.009858438698270553,0.07878389632995043,0.014787658047405828,0.10545452914105798,-0.009858438698270553
54
+ 4,5,0.09179123153965794,-0.009858438698270553,0.07551889899051509,-0.009858438698270553,0.06816621422100311,0.022181487071108744,0.09179113862504511,-0.009858438698270553
55
+ 4,6,0.0860644611804787,-0.020949182233824925,0.07011831698808839,-0.020949182233824925,0.06335544684579478,-0.009858438698270553,0.086063981251136,-0.020949182233824925
56
+ 4,7,0.07782152866762292,0.025878401582960203,0.06509185560079991,0.025878401582960203,0.05708525977259367,0.033272230606663115,0.07782195261633024,0.025878401582960203
57
+ 4,8,0.058455877925577715,0.06531215637604242,0.05051292040157318,0.06531215637604242,0.04126965025122563,0.06777676605061006,0.058456959747032966,0.06531215637604242
58
+ 4,9,0.04148139667261161,0.08749364344715116,0.03866595138624518,0.08749364344715116,0.030296496661174472,0.08256442409801588,0.041480881535243104,0.08749364344715116
59
+ 4,10,0.03392312729102964,0.08749364344715116,0.03156810017644657,0.08749364344715116,0.02185591107915039,0.06284754670147479,0.03392342137931422,0.08749364344715116
60
+ 4,11,0.02803361011507645,0.050524498328636584,0.022533154726306806,0.050524498328636584,0.011508602267209261,0.04189836446764985,0.028033303285293333,0.050524498328636584
61
+ 4,-1,0.02803361011507645,0.050524498328636584,0.022533154726306806,0.050524498328636584,0.011508602267209261,0.04189836446764985,0.028033303285293333,0.050524498328636584
62
+ 5,1,0.023055748504659954,0.04313066930493367,0.01819583380169934,0.04313066930493367,0.008013570914071046,0.022181487071108744,0.023055088490607258,0.04313066930493367
63
+ 5,2,0.01534190070618971,0.057918327352339496,0.013383465040610013,0.057918327352339496,0.003918272805114538,0.018484572559257288,0.015341892487985051,0.057918327352339496
64
+ 5,3,0.010381043833006337,0.08995825312171879,0.011764873763863128,0.08995825312171879,0.0032552430073913424,0.09981669181998934,0.010380767759000854,0.08995825312171879
65
+ 5,4,0.00926577040467444,0.12569509340294957,0.012201169398417672,0.12569509340294957,0.004673062648834422,0.11583665470467899,0.009265269973960016,0.12569509340294957
66
+ 5,5,0.009204227349769514,0.09858438698270554,0.012912673051373905,0.09858438698270554,0.006408234534243051,0.12446278856566573,0.009204887586664191,0.09858438698270554
67
+ 5,6,0.010316661694947847,0.09488747247085406,0.012488166344883679,0.09488747247085406,0.007526448974959639,0.06777676605061006,0.010317090393961767,0.09488747247085406
68
+ 5,7,0.006818329045141069,0.09488747247085406,0.00893638507272667,0.09488747247085406,0.005311970557346262,0.020949182233824925,0.006818721155666688,0.09488747247085406
69
+ 5,8,0.0008942761449032277,0.046827583816785125,0.0014714379517025353,0.046827583816785125,-0.0016031991989895952,0.007393829023702914,0.0008939216208441944,0.046827583816785125
70
+ 5,9,-0.007105730971267286,-0.025878401582960203,-0.008161827986496464,-0.025878401582960203,-0.01044707116767142,-0.04805988865406895,-0.007105470617644612,-0.025878401582960203
71
+ 5,10,-0.01222512822351949,-0.025878401582960203,-0.014262976687423004,-0.025878401582960203,-0.01544744394491792,-0.07024137572517769,-0.012224655239239955,-0.025878401582960203
72
+ 5,11,-0.01701962150153817,-0.04559527897950131,-0.021507106096183545,-0.04559527897950131,-0.018353601266599566,-0.06777676605061006,-0.017018282925094545,-0.04559527897950131
73
+ 5,-1,-0.01701962150153817,-0.04559527897950131,-0.021507106096183545,-0.04559527897950131,-0.018353601266599566,-0.06777676605061006,-0.017018282925094545,-0.04559527897950131
74
+ 6,1,-0.017251266510128392,-0.08009981442344825,-0.021697761431910254,-0.08009981442344825,-0.01700121635636173,-0.08256442409801588,-0.01725192096686521,-0.08009981442344825
75
+ 6,2,-0.02383416776735081,-0.08502903377258351,-0.02691817076806672,-0.08502903377258351,-0.022330737111589415,-0.08256442409801588,-0.023834899394285722,-0.08502903377258351
76
+ 6,3,-0.03256359607226502,-0.08256442409801588,-0.03397870655466261,-0.08256442409801588,-0.029957330206890433,-0.08256442409801588,-0.03256521563119029,-0.08256442409801588
77
+ 6,4,-0.03941771066196562,-0.07270598539974532,-0.03809339144854593,-0.07270598539974532,-0.03654928484405671,-0.051756803165920406,-0.03941726189300836,-0.07270598539974532
78
+ 6,5,-0.0439877588389697,-0.08502903377258351,-0.04023894262706567,-0.08502903377258351,-0.04100207512641553,-0.06777676605061006,-0.043988405240065045,-0.08502903377258351
79
+ 6,6,-0.04994438828730061,-0.08995825312171879,-0.044339298450672644,-0.08995825312171879,-0.046419958237563015,-0.06777676605061006,-0.049944964133513015,-0.08995825312171879
80
+ 6,7,-0.058499494294420695,-0.13062431275208483,-0.051002434877578856,-0.13062431275208483,-0.05498879213848377,-0.019716877396541106,-0.05849955514313695,-0.13062431275208483
81
+ 6,8,-0.07252656571801218,-0.08133211926073207,-0.0635280735403859,-0.08133211926073207,-0.06912017610112842,-0.06654446121332623,-0.07252611720922753,-0.08133211926073207
82
+ 6,9,-0.07710227037789122,-0.08133211926073207,-0.06795969052387736,-0.08133211926073207,-0.07414488357246932,-0.06654446121332623,-0.0771021912566973,-0.08133211926073207
83
+ 6,10,-0.07532221251124444,-0.025878401582960203,-0.06625254818309723,-0.025878401582960203,-0.07278235303101953,-0.03943375479308221,-0.07532273185017827,-0.025878401582960203
84
+ 6,11,-0.07272747570609535,-0.028343011257527837,-0.06409129552071485,-0.028343011257527837,-0.0704830819766307,-0.0320399257693793,-0.07272813567814992,-0.028343011257527837
85
+ 6,-1,-0.07272747570609535,-0.028343011257527837,-0.06409129552071485,-0.028343011257527837,-0.0704830819766307,-0.0320399257693793,-0.07272813567814992,-0.028343011257527837
86
+ 7,1,-0.06906745293615767,-0.036969145118514575,-0.0611117719496892,-0.036969145118514575,-0.0671352924778276,-0.056686022515055674,-0.06906702982900767,-0.036969145118514575
87
+ 7,2,-0.06325351214249353,-0.036969145118514575,-0.05579843053693883,-0.036969145118514575,-0.06307608497914301,-0.036969145118514575,-0.06325306922482998,-0.036969145118514575
88
+ 7,3,-0.05959302486626158,-0.046827583816785125,-0.05256425961445022,-0.046827583816785125,-0.06014822805370701,-0.007393829023702914,-0.059593354244760255,-0.046827583816785125
89
+ 7,4,-0.0609156027585929,-0.022181487071108744,-0.05419237486288025,-0.022181487071108744,-0.06228117272256674,-0.007393829023702914,-0.0609151119830821,-0.022181487071108744
90
+ 7,5,-0.0610318706402702,-0.07393829023702915,-0.054794395538168346,-0.07393829023702915,-0.0630364316085634,-0.0714736805624615,-0.06103215460417999,-0.07393829023702915
91
+ 7,6,-0.05836504424940291,-0.0714736805624615,-0.05287759886469777,-0.0714736805624615,-0.06056004871661357,-0.08626133860986734,-0.058365678918866404,-0.0714736805624615
92
+ 7,7,-0.05651888876673806,-0.04189836446764985,-0.05075173283319276,-0.04189836446764985,-0.05910456202292421,-0.046827583816785125,-0.05651845764206764,-0.04189836446764985
93
+ 7,8,-0.05521164148203063,-0.05915063218962331,-0.04913487419852178,-0.05915063218962331,-0.05769524932190345,-0.046827583816785125,-0.055212456371525595,-0.05915063218962331
94
+ 7,9,-0.05717913979344355,-0.07393829023702915,-0.050972786868436336,-0.07393829023702915,-0.05948318938556484,-0.04313066930493367,-0.05717971843886933,-0.07393829023702915
95
+ 7,10,-0.05655601802449423,-0.04929219349135277,-0.04971316488865022,-0.04929219349135277,-0.057116421104800634,-0.027110706420244022,-0.05655589507600833,-0.04929219349135277
96
+ 7,11,-0.05982640088745892,-0.06654446121332623,-0.053115130224186353,-0.06654446121332623,-0.059114440737582846,-0.0837967289352997,-0.0598260266230491,-0.06654446121332623
97
+ 7,-1,-0.05982640088745892,-0.06654446121332623,-0.053115130224186353,-0.06654446121332623,-0.059114440737582846,-0.0837967289352997,-0.0598260266230491,-0.06654446121332623
98
+ 8,1,-0.06472141215289594,-0.10104899665727317,-0.057720454516140524,-0.10104899665727317,-0.06331446104317344,-0.0837967289352997,-0.06472133490947363,-0.10104899665727317
99
+ 8,2,-0.06894659935533098,-0.10104899665727317,-0.061537143117773196,-0.10104899665727317,-0.06643052207443315,-0.09119055795900262,-0.0689469090954155,-0.10104899665727317
100
+ 8,3,-0.07429121746666983,-0.10844282568097609,-0.06624852870490482,-0.10844282568097609,-0.07090669698545685,-0.09119055795900262,-0.07429084216742525,-0.10844282568097609
101
+ 8,4,-0.08097896487794796,-0.07640289991159678,-0.07194075708278905,-0.07640289991159678,-0.07655801478260654,-0.09735208214542172,-0.08097924021312805,-0.07640289991159678
102
+ 8,5,-0.08485156803473846,-0.09365516763357025,-0.0748233957497541,-0.09365516763357025,-0.07877432459291583,-0.09242286279628643,-0.08485116382786372,-0.09365516763357025
103
+ 8,6,-0.08929910482256719,-0.06654446121332623,-0.07843737411123444,-0.06654446121332623,-0.08256988978896772,-0.04559527897950131,-0.08929900020261056,-0.06654446121332623
104
+ 8,7,-0.09221708875576931,-0.04929219349135277,-0.0808690949044732,-0.04929219349135277,-0.0858335733392962,-0.03820144995579839,-0.09221729809379861,-0.04929219349135277
105
+ 8,8,-0.0943327355563108,-0.0640798515387586,-0.08298992974157252,-0.0640798515387586,-0.08938011092948131,-0.0714736805624615,-0.09433288083213939,-0.0640798515387586
106
+ 8,9,-0.09369933535771614,-0.04436297414221749,-0.08236245452347823,-0.04436297414221749,-0.09019596467104057,-0.04436297414221749,-0.0936988269188532,-0.04436297414221749
107
+ 8,10,-0.09456707267095033,-0.009858438698270553,-0.08391255039451029,-0.009858438698270553,-0.09335407250996029,-0.04559527897950131,-0.0945669792990522,-0.009858438698270553
108
+ 8,11,-0.09328747906587528,-0.009858438698270553,-0.08299495458798656,-0.009858438698270553,-0.09237940689501219,-0.013555353210122011,-0.09328680345373135,-0.009858438698270553
109
+ 8,-1,-0.09328747906587528,-0.009858438698270553,-0.08299495458798656,-0.009858438698270553,-0.09237940689501219,-0.013555353210122011,-0.09328680345373135,-0.009858438698270553
110
+ 9,1,-0.09077082475085795,-0.013555353210122011,-0.0807523267852622,-0.013555353210122011,-0.08991927259084748,-0.013555353210122011,-0.09077021849760272,-0.013555353210122011
111
+ 9,2,-0.08730505046668151,-0.013555353210122011,-0.07722881576286894,-0.013555353210122011,-0.08615296423019589,-0.013555353210122011,-0.08730514731293662,-0.013555353210122011
112
+ 9,3,-0.084425330711635,-0.013555353210122011,-0.07425471323720745,-0.013555353210122011,-0.0829314063927245,0.0024646096745676383,-0.08442578087060619,-0.013555353210122011
113
+ 9,4,-0.08163720295781689,-0.013555353210122011,-0.07110824380299627,-0.013555353210122011,-0.07967243712697947,0.0024646096745676383,-0.0816380786484051,-0.013555353210122011
114
+ 9,5,-0.07979522690674855,-0.013555353210122011,-0.06894896352653047,-0.013555353210122011,-0.07742458974137986,0.0024646096745676383,-0.07979546789941216,-0.013555353210122011
115
+ 9,6,-0.0787299060015612,-0.013555353210122011,-0.06749723193789928,-0.013555353210122011,-0.07581524826828964,0.0024646096745676383,-0.07872970479040174,-0.013555353210122011
116
+ 9,7,-0.07846240939901694,-0.013555353210122011,-0.06704975497642751,-0.013555353210122011,-0.07534177849824533,0.0024646096745676383,-0.07846196888774402,-0.013555353210122011
117
+ 9,8,-0.07804669058566185,-0.013555353210122011,-0.06638411371973299,-0.013555353210122011,-0.07457284784422641,0.0024646096745676383,-0.07804648969273284,-0.013555353210122011
118
+ 9,9,-0.07738672697683906,-0.013555353210122011,-0.06556844725141212,-0.013555353210122011,-0.07365304187403025,-0.013555353210122011,-0.07738671344907981,-0.013555353210122011
119
+ 9,10,-0.07710890163505166,-0.013555353210122011,-0.06518325128325161,-0.013555353210122011,-0.07323256074310489,-0.013555353210122011,-0.07710913356538113,-0.013555353210122011
120
+ 9,11,-0.07721819942052116,-0.013555353210122011,-0.06529429750981788,-0.013555353210122011,-0.07335260291845239,-0.013555353210122011,-0.07721728742227899,-0.013555353210122011
121
+ 9,-1,-0.07721819942052116,-0.013555353210122011,-0.06529429750981788,-0.013555353210122011,-0.07335260291845239,-0.013555353210122011,-0.07721728742227899,-0.013555353210122011
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/10_Epochs/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/10_Epochs/modules.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ },
14
+ {
15
+ "idx": 2,
16
+ "name": "2",
17
+ "path": "2_Normalize",
18
+ "type": "sentence_transformers.models.Normalize"
19
+ }
20
+ ]
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/10_Epochs/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b728b89265b83499c320c6f9d8f3c60dae3496696b3e5dfce47c20e61f3899bc
3
+ size 328509745
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/10_Epochs/sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 512,
3
+ "do_lower_case": false
4
+ }
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/10_Epochs/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "sep_token": "</s>", "pad_token": "<pad>", "cls_token": "<s>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": false}}
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/10_Epochs/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/10_Epochs/tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"errors": "replace", "bos_token": "<s>", "eos_token": "</s>", "sep_token": "</s>", "cls_token": "<s>", "unk_token": "<unk>", "pad_token": "<pad>", "mask_token": "<mask>", "add_prefix_space": false, "trim_offsets": true, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "C:\\Users\\javier.sanz/.cache\\torch\\sentence_transformers\\sentence-transformers_all-distilroberta-v1\\", "tokenizer_class": "RobertaTokenizer"}
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/10_Epochs/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/1_Epochs/1_Pooling/config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 768,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": true,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false
7
+ }
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/1_Epochs/README.md ADDED
@@ -0,0 +1,87 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ pipeline_tag: sentence-similarity
3
+ tags:
4
+ - sentence-transformers
5
+ - feature-extraction
6
+ - sentence-similarity
7
+ ---
8
+
9
+ # {MODEL_NAME}
10
+
11
+ This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.
12
+
13
+ <!--- Describe your model here -->
14
+
15
+ ## Usage (Sentence-Transformers)
16
+
17
+ Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed:
18
+
19
+ ```
20
+ pip install -U sentence-transformers
21
+ ```
22
+
23
+ Then you can use the model like this:
24
+
25
+ ```python
26
+ from sentence_transformers import SentenceTransformer
27
+ sentences = ["This is an example sentence", "Each sentence is converted"]
28
+
29
+ model = SentenceTransformer('{MODEL_NAME}')
30
+ embeddings = model.encode(sentences)
31
+ print(embeddings)
32
+ ```
33
+
34
+
35
+
36
+ ## Evaluation Results
37
+
38
+ <!--- Describe how your model was evaluated -->
39
+
40
+ For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name={MODEL_NAME})
41
+
42
+
43
+ ## Training
44
+ The model was trained with the parameters:
45
+
46
+ **DataLoader**:
47
+
48
+ `torch.utils.data.dataloader.DataLoader` of length 11 with parameters:
49
+ ```
50
+ {'batch_size': 15, 'sampler': 'torch.utils.data.sampler.RandomSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'}
51
+ ```
52
+
53
+ **Loss**:
54
+
55
+ `sentence_transformers.losses.CosineSimilarityLoss.CosineSimilarityLoss`
56
+
57
+ Parameters of the fit()-Method:
58
+ ```
59
+ {
60
+ "epochs": 1,
61
+ "evaluation_steps": 1,
62
+ "evaluator": "sentence_transformers.evaluation.EmbeddingSimilarityEvaluator.EmbeddingSimilarityEvaluator",
63
+ "max_grad_norm": 1,
64
+ "optimizer_class": "<class 'transformers.optimization.AdamW'>",
65
+ "optimizer_params": {
66
+ "lr": 2e-05
67
+ },
68
+ "scheduler": "WarmupLinear",
69
+ "steps_per_epoch": null,
70
+ "warmup_steps": 2,
71
+ "weight_decay": 0.01
72
+ }
73
+ ```
74
+
75
+
76
+ ## Full Model Architecture
77
+ ```
78
+ SentenceTransformer(
79
+ (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: RobertaModel
80
+ (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False})
81
+ (2): Normalize()
82
+ )
83
+ ```
84
+
85
+ ## Citing & Authors
86
+
87
+ <!--- Describe where people can find more information -->
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/1_Epochs/config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "C:\\Users\\javier.sanz/.cache\\torch\\sentence_transformers\\sentence-transformers_all-distilroberta-v1\\",
3
+ "architectures": [
4
+ "RobertaModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
+ "gradient_checkpointing": false,
11
+ "hidden_act": "gelu",
12
+ "hidden_dropout_prob": 0.1,
13
+ "hidden_size": 768,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 3072,
16
+ "layer_norm_eps": 1e-05,
17
+ "max_position_embeddings": 514,
18
+ "model_type": "roberta",
19
+ "num_attention_heads": 12,
20
+ "num_hidden_layers": 6,
21
+ "pad_token_id": 1,
22
+ "position_embedding_type": "absolute",
23
+ "torch_dtype": "float32",
24
+ "transformers_version": "4.18.0",
25
+ "type_vocab_size": 1,
26
+ "use_cache": true,
27
+ "vocab_size": 50265
28
+ }
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/1_Epochs/config_sentence_transformers.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "2.0.0",
4
+ "transformers": "4.6.1",
5
+ "pytorch": "1.8.1"
6
+ }
7
+ }
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/1_Epochs/eval/similarity_evaluation_results.csv ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch,steps,cosine_pearson,cosine_spearman,euclidean_pearson,euclidean_spearman,manhattan_pearson,manhattan_spearman,dot_pearson,dot_spearman
2
+ 0,1,0.3079180908463921,0.2821978077379946,0.2862899581249616,0.2821978077379946,0.282337709193886,0.2821978077379946,0.3079178129162619,0.2821978077379946
3
+ 0,2,0.32548766782904776,0.2772685883888593,0.3035337247059212,0.2772685883888593,0.2995923411031205,0.26741014969058874,0.32548816780684997,0.2772685883888593
4
+ 0,3,0.3490661711081152,0.2846624174125622,0.3291103990868327,0.2846624174125622,0.3296520871936004,0.29328855127354897,0.3490658951873121,0.2846624174125622
5
+ 0,4,0.3426125570112317,0.3228638673683606,0.3243943353053205,0.3228638673683606,0.3211900550055936,0.3105408189955224,0.3426126072273302,0.3228638673683606
6
+ 0,5,0.38868154682462697,0.23413791908392564,0.3722658678376025,0.23413791908392564,0.375585351532577,0.23413791908392564,0.3886818109165836,0.23413791908392564
7
+ 0,6,0.44325481904054914,0.3265607818802121,0.42515482341909583,0.3265607818802121,0.4394627454096105,0.30930851415823857,0.4432547533835597,0.3265607818802121
8
+ 0,7,0.4541711960251608,0.3536714883004561,0.4356915496070872,0.3536714883004561,0.4518866575618384,0.3733883656969972,0.4541705976778214,0.3536714883004561
9
+ 0,8,0.46205329141355655,0.33149000122934735,0.44200902406845455,0.33149000122934735,0.4597854978468851,0.4115898156527956,0.4620547774687095,0.33149000122934735
10
+ 0,9,0.4620970067322604,0.3154700383446577,0.44039892825404725,0.3154700383446577,0.458477825684674,0.4165190350019308,0.4620966232023245,0.3154700383446577
11
+ 0,10,0.45961863369996,0.3080762093209548,0.43727990086026014,0.3080762093209548,0.45559589582181775,0.42144825435106614,0.4596178677812056,0.3080762093209548
12
+ 0,11,0.45846851343352313,0.31300542867009007,0.43580055484019814,0.31300542867009007,0.4542892389733235,0.40049907211724123,0.4584688539997073,0.31300542867009007
13
+ 0,-1,0.45846851343352313,0.31300542867009007,0.43580055484019814,0.31300542867009007,0.4542892389733235,0.40049907211724123,0.4584688539997073,0.31300542867009007
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/1_Epochs/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/1_Epochs/modules.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ },
14
+ {
15
+ "idx": 2,
16
+ "name": "2",
17
+ "path": "2_Normalize",
18
+ "type": "sentence_transformers.models.Normalize"
19
+ }
20
+ ]
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/1_Epochs/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43b70fc7266904d4a2188d1118a17dda49912814c3e6029683bad73721a72da0
3
+ size 328509745
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/1_Epochs/sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 512,
3
+ "do_lower_case": false
4
+ }
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/1_Epochs/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "sep_token": "</s>", "pad_token": "<pad>", "cls_token": "<s>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": false}}
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/1_Epochs/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/1_Epochs/tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"errors": "replace", "bos_token": "<s>", "eos_token": "</s>", "sep_token": "</s>", "cls_token": "<s>", "unk_token": "<unk>", "pad_token": "<pad>", "mask_token": "<mask>", "add_prefix_space": false, "trim_offsets": true, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "C:\\Users\\javier.sanz/.cache\\torch\\sentence_transformers\\sentence-transformers_all-distilroberta-v1\\", "tokenizer_class": "RobertaTokenizer"}
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/1_Epochs/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/30_Epochs/1_Pooling/config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 768,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": true,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false
7
+ }
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/30_Epochs/README.md ADDED
@@ -0,0 +1,87 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ pipeline_tag: sentence-similarity
3
+ tags:
4
+ - sentence-transformers
5
+ - feature-extraction
6
+ - sentence-similarity
7
+ ---
8
+
9
+ # {MODEL_NAME}
10
+
11
+ This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.
12
+
13
+ <!--- Describe your model here -->
14
+
15
+ ## Usage (Sentence-Transformers)
16
+
17
+ Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed:
18
+
19
+ ```
20
+ pip install -U sentence-transformers
21
+ ```
22
+
23
+ Then you can use the model like this:
24
+
25
+ ```python
26
+ from sentence_transformers import SentenceTransformer
27
+ sentences = ["This is an example sentence", "Each sentence is converted"]
28
+
29
+ model = SentenceTransformer('{MODEL_NAME}')
30
+ embeddings = model.encode(sentences)
31
+ print(embeddings)
32
+ ```
33
+
34
+
35
+
36
+ ## Evaluation Results
37
+
38
+ <!--- Describe how your model was evaluated -->
39
+
40
+ For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name={MODEL_NAME})
41
+
42
+
43
+ ## Training
44
+ The model was trained with the parameters:
45
+
46
+ **DataLoader**:
47
+
48
+ `torch.utils.data.dataloader.DataLoader` of length 11 with parameters:
49
+ ```
50
+ {'batch_size': 15, 'sampler': 'torch.utils.data.sampler.RandomSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'}
51
+ ```
52
+
53
+ **Loss**:
54
+
55
+ `sentence_transformers.losses.CosineSimilarityLoss.CosineSimilarityLoss`
56
+
57
+ Parameters of the fit()-Method:
58
+ ```
59
+ {
60
+ "epochs": 30,
61
+ "evaluation_steps": 1,
62
+ "evaluator": "sentence_transformers.evaluation.EmbeddingSimilarityEvaluator.EmbeddingSimilarityEvaluator",
63
+ "max_grad_norm": 1,
64
+ "optimizer_class": "<class 'transformers.optimization.AdamW'>",
65
+ "optimizer_params": {
66
+ "lr": 2e-05
67
+ },
68
+ "scheduler": "WarmupLinear",
69
+ "steps_per_epoch": null,
70
+ "warmup_steps": 33,
71
+ "weight_decay": 0.01
72
+ }
73
+ ```
74
+
75
+
76
+ ## Full Model Architecture
77
+ ```
78
+ SentenceTransformer(
79
+ (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: RobertaModel
80
+ (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False})
81
+ (2): Normalize()
82
+ )
83
+ ```
84
+
85
+ ## Citing & Authors
86
+
87
+ <!--- Describe where people can find more information -->
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/30_Epochs/config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "C:\\Users\\javier.sanz/.cache\\torch\\sentence_transformers\\sentence-transformers_all-distilroberta-v1\\",
3
+ "architectures": [
4
+ "RobertaModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
+ "gradient_checkpointing": false,
11
+ "hidden_act": "gelu",
12
+ "hidden_dropout_prob": 0.1,
13
+ "hidden_size": 768,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 3072,
16
+ "layer_norm_eps": 1e-05,
17
+ "max_position_embeddings": 514,
18
+ "model_type": "roberta",
19
+ "num_attention_heads": 12,
20
+ "num_hidden_layers": 6,
21
+ "pad_token_id": 1,
22
+ "position_embedding_type": "absolute",
23
+ "torch_dtype": "float32",
24
+ "transformers_version": "4.18.0",
25
+ "type_vocab_size": 1,
26
+ "use_cache": true,
27
+ "vocab_size": 50265
28
+ }
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/30_Epochs/config_sentence_transformers.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "2.0.0",
4
+ "transformers": "4.6.1",
5
+ "pytorch": "1.8.1"
6
+ }
7
+ }
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/30_Epochs/eval/similarity_evaluation_results.csv ADDED
@@ -0,0 +1,361 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch,steps,cosine_pearson,cosine_spearman,euclidean_pearson,euclidean_spearman,manhattan_pearson,manhattan_spearman,dot_pearson,dot_spearman
2
+ 0,1,0.3079180908463921,0.2821978077379946,0.2862899581249616,0.2821978077379946,0.282337709193886,0.2821978077379946,0.3079178129162619,0.2821978077379946
3
+ 0,2,0.3102918414707915,0.2945208561108328,0.2885662438392038,0.2945208561108328,0.28534301152878905,0.2821978077379946,0.31029325825435755,0.2945208561108328
4
+ 0,3,0.3161761998347872,0.29945007545996805,0.2943682441888069,0.29945007545996805,0.2926365352307254,0.2821978077379946,0.3161766436600215,0.29945007545996805
5
+ 0,4,0.3233389854082247,0.29945007545996805,0.3015640779202423,0.29945007545996805,0.30207604702438295,0.2821978077379946,0.32333848122301145,0.29945007545996805
6
+ 0,5,0.3324403126797135,0.25015788196861527,0.3104693160829496,0.25015788196861527,0.3136097773289431,0.26741014969058874,0.33244074694292103,0.25015788196861527
7
+ 0,6,0.34379961772474665,0.27110706420244024,0.3213046971930523,0.27110706420244024,0.326944209884841,0.23413791908392564,0.34380017876388236,0.27110706420244024
8
+ 0,7,0.35867039288622243,0.21072412717553307,0.3351883488596057,0.21072412717553307,0.3442847723302974,0.2057949078263978,0.358670438503069,0.21072412717553307
9
+ 0,8,0.36312378272828943,0.21072412717553307,0.3390918599405152,0.21072412717553307,0.3496461026930249,0.22058256587380362,0.3631253174089124,0.21072412717553307
10
+ 0,9,0.3623971608176939,0.23660252875849325,0.3378571580830686,0.23660252875849325,0.3484022687142426,0.21565334652466836,0.3623972408383269,0.23660252875849325
11
+ 0,10,0.3630907121125635,0.24153174810762854,0.33762150145176517,0.24153174810762854,0.34911644684735077,0.25139018680589914,0.3630911961807046,0.24153174810762854
12
+ 0,11,0.3702149795183414,0.3154700383446577,0.3431712975287372,0.3154700383446577,0.3570512418393157,0.2834301125752784,0.37021410072893335,0.3154700383446577
13
+ 0,-1,0.3702149795183414,0.3154700383446577,0.3431712975287372,0.3154700383446577,0.3570512418393157,0.2834301125752784,0.37021410072893335,0.3154700383446577
14
+ 1,1,0.36936778820067784,0.3216315625310768,0.3407088950266476,0.3216315625310768,0.35584057599140906,0.29821777062268423,0.3693674578814742,0.3216315625310768
15
+ 1,2,0.36768222080789326,0.3105408189955224,0.3371903016569286,0.3105408189955224,0.3528744722141206,0.3043792948091033,0.3676827214468988,0.3105408189955224
16
+ 1,3,0.36461598348985585,0.2649455400160211,0.33240546680040134,0.2649455400160211,0.34850067000696605,0.30930851415823857,0.36461546947573703,0.2649455400160211
17
+ 1,4,0.35656651117560034,0.34011613509033406,0.3232689351266155,0.34011613509033406,0.3388121535214809,0.3351869157411988,0.3565664907671555,0.34011613509033406
18
+ 1,5,0.3384344896248657,0.30191468513453573,0.30607349343437407,0.30191468513453573,0.32096637173475795,0.3265607818802121,0.3384345826127576,0.30191468513453573
19
+ 1,6,0.31713403172628785,0.3142377335073739,0.28649680167299163,0.3142377335073739,0.3016127941451553,0.3265607818802121,0.3171325899595321,0.3142377335073739
20
+ 1,7,0.28598634687349983,0.2735716738770078,0.25799813395637633,0.2735716738770078,0.2714122484236446,0.2735716738770078,0.28598544876049753,0.2735716738770078
21
+ 1,8,0.24626106162354613,0.23290561424664183,0.22083789157734657,0.23290561424664183,0.22973273898893742,0.24769327229404767,0.24626081329364755,0.23290561424664183
22
+ 1,9,0.2360758336355505,0.23290561424664183,0.2109093800734074,0.23290561424664183,0.21883075805733274,0.25508710131775053,0.23607575082055132,0.23290561424664183
23
+ 1,10,0.22333821228738465,0.20333029815183015,0.19847168715267255,0.20333029815183015,0.20594146317294346,0.22058256587380362,0.22333758881365312,0.20333029815183015
24
+ 1,11,0.2278109320527154,0.19840107880269486,0.20242685390231951,0.19840107880269486,0.2087123261350129,0.23537022392120946,0.22780982394215188,0.19840107880269486
25
+ 1,-1,0.2278109320527154,0.19840107880269486,0.20242685390231951,0.19840107880269486,0.2087123261350129,0.23537022392120946,0.22780982394215188,0.19840107880269486
26
+ 2,1,0.2452375812805108,0.1786842014061538,0.21811164713593595,0.1786842014061538,0.22736844906872428,0.22304717554837125,0.24523835250161485,0.1786842014061538
27
+ 2,2,0.24910048649496594,0.1478765804740583,0.22160676171593852,0.1478765804740583,0.24041588743509032,0.2119564320128169,0.24909825557099002,0.1478765804740583
28
+ 2,3,0.24442553804587566,0.21442104168738454,0.21795167739562765,0.21442104168738454,0.24026884379428937,0.2649455400160211,0.24442667410877264,0.21442104168738454
29
+ 2,4,0.24861177857123834,0.18114881108072142,0.22177953007135542,0.18114881108072142,0.24377447295609456,0.23290561424664183,0.24861142113416804,0.18114881108072142
30
+ 2,5,0.2426977135372219,0.18114881108072142,0.21726308011452183,0.18114881108072142,0.23712836909225202,0.2255117852229389,0.2427002051765677,0.18114881108072142
31
+ 2,6,0.2543445480281932,0.18114881108072142,0.22824096490289994,0.18114881108072142,0.24878348054534039,0.2255117852229389,0.25434296002724227,0.18114881108072142
32
+ 2,7,0.2761606811448173,0.18114881108072142,0.2476746610760068,0.18114881108072142,0.2678091501814641,0.2131887368501007,0.27616032406767893,0.18114881108072142
33
+ 2,8,0.2789878051692154,0.18114881108072142,0.2497969747221947,0.18114881108072142,0.2694962624491478,0.2131887368501007,0.2789888699598022,0.18114881108072142
34
+ 2,9,0.2785838910671605,0.1614319336841803,0.2495754184962633,0.1614319336841803,0.26759218032048054,0.1540381046604774,0.2785828784421418,0.1614319336841803
35
+ 2,10,0.275965601655409,0.1540381046604774,0.2469051424586284,0.1540381046604774,0.2624257866379952,0.1540381046604774,0.27596423274260085,0.1540381046604774
36
+ 2,11,0.27070762170422247,0.1614319336841803,0.241719173510697,0.1614319336841803,0.2507321866363784,0.15157349498590975,0.2707066901738455,0.1614319336841803
37
+ 2,-1,0.27070762170422247,0.1614319336841803,0.241719173510697,0.1614319336841803,0.2507321866363784,0.15157349498590975,0.2707066901738455,0.1614319336841803
38
+ 3,1,0.2437973593244876,0.15157349498590975,0.21907237693860065,0.15157349498590975,0.22168718187727826,0.15773501917232885,0.24379751667841504,0.15157349498590975
39
+ 3,2,0.21796477487404953,0.15773501917232885,0.19766691681972304,0.15773501917232885,0.19386063074734478,0.13432122726393628,0.21796403653985932,0.15773501917232885
40
+ 3,3,0.1834494830844665,0.13555353210122012,0.17071314716565228,0.13555353210122012,0.16182034380096386,0.13432122726393628,0.18344735971516526,0.13555353210122012
41
+ 3,4,0.17075147800698326,0.13801814177578775,0.16211056872913632,0.13801814177578775,0.14571809255021456,0.09858438698270554,0.1707517624909497,0.13801814177578775
42
+ 3,5,0.20690663879946858,0.13555353210122012,0.19693510721926444,0.13555353210122012,0.17804707784676918,0.15280579982319356,0.20690797425981683,0.13555353210122012
43
+ 3,6,0.23069438973903192,0.20086568847726252,0.21897398255720593,0.20086568847726252,0.19445914050734497,0.17375498205701848,0.23069638197881362,0.20086568847726252
44
+ 3,7,0.26090969496252225,0.2883593319244137,0.2464435458917295,0.2883593319244137,0.22369682220041753,0.25139018680589914,0.26090895113976764,0.2883593319244137
45
+ 3,8,0.2835241553523494,0.2846624174125622,0.2674751688857458,0.2846624174125622,0.24561134005315094,0.25015788196861527,0.28352247810627823,0.2846624174125622
46
+ 3,9,0.28510572749594726,0.31916695285650915,0.2693558375099329,0.31916695285650915,0.24906232030210884,0.2748039787142917,0.28510761714430455,0.31916695285650915
47
+ 3,10,0.27047260372963555,0.2698747593651564,0.25825212384402013,0.2698747593651564,0.24127201553873745,0.24769327229404767,0.27047138614566735,0.2698747593651564
48
+ 3,11,0.25751087639860293,0.25878401582960203,0.24764691337075634,0.25878401582960203,0.23197535329430435,0.25878401582960203,0.25750971109046744,0.25878401582960203
49
+ 3,-1,0.25751087639860293,0.25878401582960203,0.24764691337075634,0.25878401582960203,0.23197535329430435,0.25878401582960203,0.25750971109046744,0.25878401582960203
50
+ 4,1,0.2278670502545877,0.26864245452787255,0.22290850619745292,0.26864245452787255,0.20552518305389245,0.21565334652466836,0.22786712458859046,0.26864245452787255
51
+ 4,2,0.19926071178878946,0.20209799331454634,0.19533962615729916,0.20209799331454634,0.1771307216606581,0.16882576270788321,0.19926219628648723,0.20209799331454634
52
+ 4,3,0.1973029053989826,0.20086568847726252,0.19386011774184558,0.20086568847726252,0.17623260522558198,0.16389654335874795,0.19730287499305138,0.20086568847726252
53
+ 4,4,0.2230881038072967,0.22920869973479036,0.21937960187644678,0.22920869973479036,0.2016594564738479,0.22920869973479036,0.2230882332701184,0.22920869973479036
54
+ 4,5,0.23115769746619758,0.2131887368501007,0.22706675805744264,0.2131887368501007,0.21019065446888904,0.19470416429084345,0.2311585156431114,0.2131887368501007
55
+ 4,6,0.2229857545264993,0.2131887368501007,0.21998760585690053,0.2131887368501007,0.20670036071309333,0.2131887368501007,0.22298547765447901,0.2131887368501007
56
+ 4,7,0.19534951429850922,0.1367858369385039,0.19435220763427444,0.1367858369385039,0.1912216167637827,0.15157349498590975,0.19534813958690134,0.1367858369385039
57
+ 4,8,0.18422606929050345,0.12692739824023336,0.1840801864603841,0.12692739824023336,0.18523978846252467,0.15157349498590975,0.1842251217924526,0.12692739824023336
58
+ 4,9,0.2028696954116062,0.21565334652466836,0.20341266088536275,0.21565334652466836,0.20657203770600566,0.2846624174125622,0.20286970809565552,0.21565334652466836
59
+ 4,10,0.22289243874739087,0.24769327229404767,0.2189719189059325,0.24769327229404767,0.21636044914979266,0.2563194061550344,0.22289156289023526,0.24769327229404767
60
+ 4,11,0.22959752795862418,0.17621959173158613,0.2184824620659441,0.17621959173158613,0.21229969659521308,0.18361342075528905,0.22959767477057674,0.17621959173158613
61
+ 4,-1,0.22959752795862418,0.17621959173158613,0.2184824620659441,0.17621959173158613,0.21229969659521308,0.18361342075528905,0.22959767477057674,0.17621959173158613
62
+ 5,1,0.23523994790835162,0.10967513051825992,0.21675488732982515,0.10967513051825992,0.2100777758361948,0.09488747247085406,0.23524056324074932,0.10967513051825992
63
+ 5,2,0.23353353247166062,0.10351360633184081,0.20811215419682907,0.10351360633184081,0.20268189966675357,0.04929219349135277,0.2335354925860216,0.10351360633184081
64
+ 5,3,0.22609895711862238,0.01601996288468965,0.19557509552803484,0.01601996288468965,0.19038694126482678,0.03820144995579839,0.22610026181500517,0.01601996288468965
65
+ 5,4,0.22914020709304672,0.03080762093209548,0.19506451901272862,0.03080762093209548,0.18972157714248708,0.060382937026907134,0.2291400221602664,0.03080762093209548
66
+ 5,5,0.22883580431208225,0.018484572559257288,0.19224524327342024,0.018484572559257288,0.18616512090672024,0.04805988865406895,0.22883546036040142,0.018484572559257288
67
+ 5,6,0.22227767376689445,0.018484572559257288,0.1856708617738887,0.018484572559257288,0.18034853842558077,0.05298910800320422,0.22227857468114678,0.018484572559257288
68
+ 5,7,0.20636218340869375,-0.020949182233824925,0.17072968612431322,-0.020949182233824925,0.1658648792834651,0.013555353210122011,0.2063616703207703,-0.020949182233824925
69
+ 5,8,0.19125932636065704,-0.01601996288468965,0.15713040294221126,-0.01601996288468965,0.15343259299718084,-0.0012323048372838191,0.19125995487351846,-0.01601996288468965
70
+ 5,9,0.15123526401557832,0.01601996288468965,0.12007491324401073,0.01601996288468965,0.1159999998923083,-0.025878401582960203,0.15123597378895715,0.01601996288468965
71
+ 5,10,0.13959539725930198,0.028343011257527837,0.11205075895679045,0.028343011257527837,0.1092339245768948,-0.013555353210122011,0.1395957950566103,0.028343011257527837
72
+ 5,11,0.12749169903386123,-0.012323048372838192,0.10343482039963746,-0.012323048372838192,0.10274520955072222,-0.027110706420244022,0.12749099718309265,-0.012323048372838192
73
+ 5,-1,0.12749169903386123,-0.012323048372838192,0.10343482039963746,-0.012323048372838192,0.10274520955072222,-0.027110706420244022,0.12749099718309265,-0.012323048372838192
74
+ 6,1,0.1261081595226637,-0.04559527897950131,0.10429813524181773,-0.04559527897950131,0.10372322125464434,-0.051756803165920406,0.12610775499263915,-0.04559527897950131
75
+ 6,2,0.1342302505098415,-0.03573684028123075,0.11481512227475459,-0.03573684028123075,0.11605804665166095,-0.060382937026907134,0.13422990725516132,-0.03573684028123075
76
+ 6,3,0.1536211632032343,0.04559527897950131,0.13613838824889685,0.04559527897950131,0.1389362676521559,0.06777676605061006,0.15362065827030882,0.04559527897950131
77
+ 6,4,0.17041267667460852,0.04805988865406895,0.1544960873189723,0.04805988865406895,0.15776662987693735,0.07270598539974532,0.1704120809332786,0.04805988865406895
78
+ 6,5,0.17610656176820008,0.046827583816785125,0.16218688234868373,0.046827583816785125,0.16466895310459156,0.07270598539974532,0.17610662358963025,0.046827583816785125
79
+ 6,6,0.18014225720490223,0.033272230606663115,0.16811200138968818,0.033272230606663115,0.1695696316154338,0.11337204503011135,0.18014255926920728,0.033272230606663115
80
+ 6,7,0.1539755220810478,-0.018484572559257288,0.14577249825251046,-0.018484572559257288,0.145603393516457,0.018484572559257288,0.15397490305104725,-0.018484572559257288
81
+ 6,8,0.1287087362181055,0.020949182233824925,0.12473165325687158,0.020949182233824925,0.12329951597786068,0.020949182233824925,0.12871010542023187,0.020949182233824925
82
+ 6,9,0.1088793355316271,-0.022181487071108744,0.10816025656751246,-0.022181487071108744,0.10623586823707254,0.05298910800320422,0.10887907317370252,-0.022181487071108744
83
+ 6,10,0.09618946433875819,0.025878401582960203,0.09997021952836033,0.025878401582960203,0.0978007315552987,0.013555353210122011,0.09618890803335327,0.025878401582960203
84
+ 6,11,0.08762134889280004,0.0714736805624615,0.0946139217289927,0.0714736805624615,0.09224623119769274,0.0714736805624615,0.08762222692915621,0.0714736805624615
85
+ 6,-1,0.08762134889280004,0.0714736805624615,0.0946139217289927,0.0714736805624615,0.09224623119769274,0.0714736805624615,0.08762222692915621,0.0714736805624615
86
+ 7,1,0.05302998283509866,0.06161524186419096,0.06136425557629192,0.06161524186419096,0.056644697068278016,0.06161524186419096,0.053030914401850204,0.06161524186419096
87
+ 7,2,0.03785636831048968,0.05298910800320422,0.04523746297207182,0.05298910800320422,0.0401499292868948,0.07517059507431297,0.03785680515003324,0.05298910800320422
88
+ 7,3,0.020577668313321598,0.05298910800320422,0.026758969599422464,0.05298910800320422,0.02097232482704524,0.08009981442344825,0.02057869524794277,0.05298910800320422
89
+ 7,4,0.023887491455300905,-0.012323048372838192,0.029553038586638374,-0.012323048372838192,0.02155187753305872,-0.028343011257527837,0.023886955637562268,-0.012323048372838192
90
+ 7,5,0.05160207226114591,0.0714736805624615,0.05742506546720527,0.0714736805624615,0.04483841109759608,0.0714736805624615,0.051603651264640546,0.0714736805624615
91
+ 7,6,0.06688848257071714,-0.014787658047405828,0.07275994955404096,-0.014787658047405828,0.059516039364205464,-0.006161524186419096,0.06688855981965684,-0.014787658047405828
92
+ 7,7,0.08324382067794542,0.01725226772197347,0.08893132811058546,0.01725226772197347,0.07874831379432522,-0.0012323048372838191,0.08324485090298461,0.01725226772197347
93
+ 7,8,0.10425421174383559,0.0776352047488806,0.10946549134677688,0.0776352047488806,0.10004939826861622,0.020949182233824925,0.10425399561672169,0.0776352047488806
94
+ 7,9,0.12002936172746556,0.10721052084369227,0.1259932184079851,0.10721052084369227,0.1175848089422584,0.09488747247085406,0.120028567931227,0.10721052084369227
95
+ 7,10,0.12202521637612537,0.09858438698270554,0.13030893132905652,0.09858438698270554,0.12379133592683478,0.09242286279628643,0.12202522207705817,0.09858438698270554
96
+ 7,11,0.12047795343128159,0.14910888531134212,0.13114938749202507,0.14910888531134212,0.1255516919803514,0.11830126437924662,0.12047891779158075,0.14910888531134212
97
+ 7,-1,0.12047795343128159,0.14910888531134212,0.13114938749202507,0.14910888531134212,0.1255516919803514,0.11830126437924662,0.12047891779158075,0.14910888531134212
98
+ 8,1,0.10367783964628656,0.18484572559257287,0.11753504941945435,0.18484572559257287,0.11404100295677604,0.17991650624343758,0.10367876883970423,0.18484572559257287
99
+ 8,2,0.08628622606381792,0.0776352047488806,0.09927023582440425,0.0776352047488806,0.09676885222675317,0.10967513051825992,0.0862859375654445,0.0776352047488806
100
+ 8,3,0.07398052668539788,0.050524498328636584,0.08364461631082981,0.050524498328636584,0.0823174421976038,0.08502903377258351,0.07398115339116747,0.050524498328636584
101
+ 8,4,0.07968043094267019,0.07886750958616442,0.08868105279039504,0.07886750958616442,0.08772915384005481,0.08872594828443497,0.0796807989188768,0.07886750958616442
102
+ 8,5,0.0913134011077035,0.11953356921653045,0.09717854898591424,0.11953356921653045,0.09426758672530473,0.09242286279628643,0.09131349993366358,0.11953356921653045
103
+ 8,6,0.11226917834209388,0.12692739824023336,0.11752655962347935,0.12692739824023336,0.11213157429233853,0.14417966596220685,0.11226876670392537,0.12692739824023336
104
+ 8,7,0.1427176780440666,0.1478765804740583,0.14586563982507086,0.1478765804740583,0.13953091633145914,0.1219981788910981,0.14271720726433024,0.1478765804740583
105
+ 8,8,0.15340475714228985,0.1281597030775172,0.15735321456090764,0.1281597030775172,0.15342174095023392,0.1478765804740583,0.15340422482537086,0.1281597030775172
106
+ 8,9,0.14126907589458543,0.14664427563677448,0.14589442612852246,0.14664427563677448,0.13995450674011722,0.16266423852146414,0.14126823374008252,0.14664427563677448
107
+ 8,10,0.13160711890367868,0.13308892242665246,0.13743605924508384,0.13308892242665246,0.12972283951006197,0.12692739824023336,0.13160686019594414,0.13308892242665246
108
+ 8,11,0.1486289255727679,0.1614319336841803,0.15336355156009565,0.1614319336841803,0.14504746856619405,0.13925044661307157,0.1486291025228969,0.1614319336841803
109
+ 8,-1,0.1486289255727679,0.1614319336841803,0.15336355156009565,0.1614319336841803,0.14504746856619405,0.13925044661307157,0.1486291025228969,0.1614319336841803
110
+ 9,1,0.16176801458974083,0.16636115303331558,0.16704986977572134,0.16636115303331558,0.15803467790326112,0.13925044661307157,0.1617670474469548,0.16636115303331558
111
+ 9,2,0.1702687854856832,0.16636115303331558,0.1761295529442214,0.16636115303331558,0.16654792668498003,0.14417966596220685,0.17027101574299341,0.16636115303331558
112
+ 9,3,0.1794540904861532,0.142947361124923,0.1847900190120435,0.142947361124923,0.1752653421061391,0.15034119014862593,0.1794537258040722,0.142947361124923
113
+ 9,4,0.1901740694195568,0.12939200791480102,0.19482493595089073,0.12939200791480102,0.1846303385987285,0.09119055795900262,0.19017341339828453,0.12939200791480102
114
+ 9,5,0.1962413206423465,0.12446278856566573,0.200519362522369,0.12446278856566573,0.1901819202231166,0.17498728689430232,0.1962410167298008,0.12446278856566573
115
+ 9,6,0.2020142973880297,0.1786842014061538,0.20639669669690724,0.1786842014061538,0.19671307963805904,0.1786842014061538,0.20201343313354153,0.1786842014061538
116
+ 9,7,0.19644494974782128,0.15896732400961266,0.20195558572465794,0.15896732400961266,0.19018289568301625,0.16882576270788321,0.1964450000326114,0.15896732400961266
117
+ 9,8,0.18900147222695818,0.15896732400961266,0.19533567302329669,0.15896732400961266,0.181104694728567,0.1478765804740583,0.1890011584034093,0.15896732400961266
118
+ 9,9,0.16375756327435137,0.16882576270788321,0.171494887641397,0.16882576270788321,0.15312112810732106,0.16636115303331558,0.1637582872154399,0.16882576270788321
119
+ 9,10,0.1566142652074977,0.16882576270788321,0.1645565863880661,0.16882576270788321,0.1455493800727883,0.15650271433504503,0.15661467131946302,0.16882576270788321
120
+ 9,11,0.16049734139102423,0.16389654335874795,0.1682590315722275,0.16389654335874795,0.14856595252324867,0.1873103352671405,0.16049811550238802,0.16389654335874795
121
+ 9,-1,0.16049734139102423,0.16389654335874795,0.1682590315722275,0.16389654335874795,0.14856595252324867,0.1873103352671405,0.16049811550238802,0.16389654335874795
122
+ 10,1,0.1824486313057398,0.16389654335874795,0.189034526334402,0.16389654335874795,0.1707407705751082,0.15280579982319356,0.1824494619490654,0.16389654335874795
123
+ 10,2,0.21621952187395538,0.19593646912812723,0.22068233927319242,0.19593646912812723,0.2041818702205512,0.1873103352671405,0.21622109601495396,0.19593646912812723
124
+ 10,3,0.22606735476092432,0.23290561424664183,0.2291162493436356,0.23290561424664183,0.21448613426683993,0.20702721266368163,0.22606804840460695,0.23290561424664183
125
+ 10,4,0.23351630919142732,0.21811795619923596,0.23465057344410478,0.21811795619923596,0.22146206417904457,0.20702721266368163,0.2335170155862974,0.21811795619923596
126
+ 10,5,0.22617471727387253,0.21688565136195218,0.22637501117393433,0.21688565136195218,0.21526230518610256,0.2131887368501007,0.22617490893392628,0.21688565136195218
127
+ 10,6,0.2008739593456581,0.19593646912812723,0.2019304673943434,0.19593646912812723,0.19303419852486028,0.19593646912812723,0.20087215019876897,0.19593646912812723
128
+ 10,7,0.17212580706994873,0.15896732400961266,0.17392395278304662,0.15896732400961266,0.1670324327690546,0.13925044661307157,0.17212478463890662,0.15896732400961266
129
+ 10,8,0.1453856110791112,0.13925044661307157,0.14748072821378141,0.13925044661307157,0.14220932054689195,0.13925044661307157,0.14538692759428473,0.13925044661307157
130
+ 10,9,0.12328016072786185,0.12446278856566573,0.12518162700698982,0.12446278856566573,0.1215358490809334,0.13308892242665246,0.12327954344590616,0.12446278856566573
131
+ 10,10,0.10750218149514255,0.16512884819603177,0.10904103331389231,0.16512884819603177,0.10661810912495118,0.13308892242665246,0.10750272055189065,0.16512884819603177
132
+ 10,11,0.08875623784007493,0.12446278856566573,0.09077502921617099,0.12446278856566573,0.0884594404681523,0.16512884819603177,0.08875667412760314,0.12446278856566573
133
+ 10,-1,0.08875623784007493,0.12446278856566573,0.09077502921617099,0.12446278856566573,0.0884594404681523,0.16512884819603177,0.08875667412760314,0.12446278856566573
134
+ 11,1,0.08302672609924468,0.13308892242665246,0.08730000689945511,0.13308892242665246,0.08223070071776882,0.09242286279628643,0.08302653355793888,0.13308892242665246
135
+ 11,2,0.07366395699159475,0.10104899665727317,0.07919586854629751,0.10104899665727317,0.07147978938948946,0.10104899665727317,0.07366326156425232,0.10104899665727317
136
+ 11,3,0.06986613333100213,0.10104899665727317,0.07750789532854709,0.10104899665727317,0.06621880829683793,0.10104899665727317,0.06986710597416936,0.10104899665727317
137
+ 11,4,0.0620393512549261,0.15280579982319356,0.07101617374068495,0.15280579982319356,0.05687224336208165,0.12076587405381427,0.062039676955718354,0.15280579982319356
138
+ 11,5,0.059178092416221934,0.14910888531134212,0.06913337707314174,0.14910888531134212,0.05153680861109978,0.10474591116912464,0.05917829495024203,0.14910888531134212
139
+ 11,6,0.0757693650570112,0.17621959173158613,0.08616145679979383,0.17621959173158613,0.06661159844866991,0.07393829023702915,0.07576932282585136,0.17621959173158613
140
+ 11,7,0.09237910456208068,0.2131887368501007,0.10237452326739427,0.2131887368501007,0.08283872926166679,0.1540381046604774,0.09238005523327417,0.2131887368501007
141
+ 11,8,0.10610299404593891,0.22058256587380362,0.11564815160847008,0.22058256587380362,0.09720667749718995,0.16266423852146414,0.10610253544997686,0.22058256587380362
142
+ 11,9,0.1062885553119478,0.18854264010442434,0.11588264867245238,0.18854264010442434,0.09878699798139842,0.16266423852146414,0.10628876593160744,0.18854264010442434
143
+ 11,10,0.10758671989328628,0.18854264010442434,0.11771792719966735,0.18854264010442434,0.10312543836311686,0.16266423852146414,0.1075867471584677,0.18854264010442434
144
+ 11,11,0.11342157819066764,0.2255117852229389,0.1231631894484963,0.2255117852229389,0.10794132803591885,0.16882576270788321,0.11342148790206283,0.2255117852229389
145
+ 11,-1,0.11342157819066764,0.2255117852229389,0.1231631894484963,0.2255117852229389,0.10794132803591885,0.16882576270788321,0.11342148790206283,0.2255117852229389
146
+ 12,1,0.11703012507343775,0.1873103352671405,0.12597078855115926,0.1873103352671405,0.11099589507006638,0.15650271433504503,0.11703025431870738,0.1873103352671405
147
+ 12,2,0.1275857062811969,0.17498728689430232,0.13497670304586074,0.17498728689430232,0.12016541434895277,0.12076587405381427,0.12758595806269857,0.17498728689430232
148
+ 12,3,0.13719825857647316,0.18484572559257287,0.1435565871963705,0.18484572559257287,0.12877139765131837,0.11213974019282755,0.1371979708444724,0.18484572559257287
149
+ 12,4,0.13994631058794926,0.142947361124923,0.14542901861362845,0.142947361124923,0.13132318178437957,0.07886750958616442,0.13994579815495423,0.142947361124923
150
+ 12,5,0.14351231056150765,0.13062431275208483,0.14844969024508617,0.13062431275208483,0.13508724773864786,0.13801814177578775,0.14351230217898953,0.13062431275208483
151
+ 12,6,0.1576546915880126,0.1922395546162758,0.16156898719844426,0.1922395546162758,0.14844084383527095,0.1540381046604774,0.15765474497878304,0.1922395546162758
152
+ 12,7,0.17652798272622774,0.2082595175009654,0.17937389909418314,0.2082595175009654,0.16665714881543542,0.1540381046604774,0.1765276282058402,0.2082595175009654
153
+ 12,8,0.1872741287184585,0.23290561424664183,0.1902043040035371,0.23290561424664183,0.17765296136944886,0.18607803042985668,0.18727495856285062,0.23290561424664183
154
+ 12,9,0.197687503783829,0.2649455400160211,0.2006386107495033,0.2649455400160211,0.18839241488872227,0.21811795619923596,0.19768694578873233,0.2649455400160211
155
+ 12,10,0.19910105432692923,0.24029944327034472,0.20283557368079516,0.24029944327034472,0.19177488771370277,0.24029944327034472,0.19910184795697025,0.24029944327034472
156
+ 12,11,0.20083899422484947,0.21688565136195218,0.2052976303577805,0.21688565136195218,0.19502590601668135,0.2624809303414535,0.20083957037866013,0.21688565136195218
157
+ 12,-1,0.20083899422484947,0.21688565136195218,0.2052976303577805,0.21688565136195218,0.19502590601668135,0.2624809303414535,0.20083957037866013,0.21688565136195218
158
+ 13,1,0.20904072407773916,0.21688565136195218,0.21376429428988866,0.21688565136195218,0.2040324806474459,0.24646096745676385,0.2090394594707577,0.21688565136195218
159
+ 13,2,0.22539313505125397,0.21688565136195218,0.2295705306223735,0.21688565136195218,0.22078868836078128,0.22920869973479036,0.22539184863863024,0.21688565136195218
160
+ 13,3,0.24281879264063705,0.2661778448533049,0.24662854000250467,0.2661778448533049,0.23983697223048037,0.22920869973479036,0.24281926437517665,0.2661778448533049
161
+ 13,4,0.26223221827132137,0.2661778448533049,0.2654198747599035,0.2661778448533049,0.26128726364913224,0.2649455400160211,0.26223277931918654,0.2661778448533049
162
+ 13,5,0.27182716417232566,0.24522866261948,0.2749743561755036,0.24522866261948,0.2727262476441382,0.28712702708712984,0.27182629491317334,0.24522866261948
163
+ 13,6,0.2739013370896408,0.22920869973479036,0.2777014367155881,0.22920869973479036,0.2761447762274925,0.2748039787142917,0.27390162417683217,0.22920869973479036
164
+ 13,7,0.2853772661164533,0.29328855127354897,0.2896311073953694,0.29328855127354897,0.29021924670225363,0.3277930867174959,0.2853770070626842,0.29328855127354897
165
+ 13,8,0.29312461374934345,0.29328855127354897,0.2980162405456133,0.29328855127354897,0.3001583515988536,0.3277930867174959,0.29312424085762967,0.29328855127354897
166
+ 13,9,0.2952002458270657,0.31793464801922533,0.3008731931802958,0.31793464801922533,0.30539148844642783,0.35243918346317227,0.29520134419014,0.31793464801922533
167
+ 13,10,0.2976309966799597,0.31793464801922533,0.3039291172102294,0.31793464801922533,0.31010914281679663,0.3684591463478619,0.29763131203964915,0.31793464801922533
168
+ 13,11,0.2946906271576391,0.29945007545996805,0.302239423585306,0.29945007545996805,0.3106998028778017,0.38324680439526776,0.29468995860092484,0.29945007545996805
169
+ 13,-1,0.2946906271576391,0.29945007545996805,0.302239423585306,0.29945007545996805,0.3106998028778017,0.38324680439526776,0.29468995860092484,0.29945007545996805
170
+ 14,1,0.28997831906117344,0.2834301125752784,0.29892501100364366,0.2834301125752784,0.30837709084188114,0.3031469899718195,0.2899785232815844,0.2834301125752784
171
+ 14,2,0.2769846376493739,0.285894722249846,0.2876875730911927,0.285894722249846,0.29829487590775017,0.27850089322614313,0.27698437602929976,0.285894722249846
172
+ 14,3,0.2638386396405887,0.31177312383280625,0.2763305445295704,0.31177312383280625,0.2876879167121095,0.2957531609481166,0.263839229454118,0.31177312383280625
173
+ 14,4,0.2598088271326054,0.25139018680589914,0.272944760781639,0.25139018680589914,0.2841131560015727,0.25139018680589914,0.25980879983190597,0.25139018680589914
174
+ 14,5,0.25897743047787825,0.27850089322614313,0.27289043470934593,0.27850089322614313,0.28388144157234735,0.2834301125752784,0.25897710191968526,0.27850089322614313
175
+ 14,6,0.2538246204140274,0.2661778448533049,0.26858783466146463,0.2661778448533049,0.27923049862622495,0.2760362835515755,0.2538245790514121,0.2661778448533049
176
+ 14,7,0.23397283230104188,0.27850089322614313,0.24945647613612443,0.27850089322614313,0.2601048852921486,0.26124862550416966,0.2339724638545987,0.27850089322614313
177
+ 14,8,0.20838359433320197,0.23537022392120946,0.22536344277006914,0.23537022392120946,0.2350870109037186,0.27973319806342695,0.20838359641846393,0.23537022392120946
178
+ 14,9,0.19871764517056423,0.28096550290071076,0.2171620567845268,0.28096550290071076,0.22624034766696563,0.2760362835515755,0.19871781653776763,0.28096550290071076
179
+ 14,10,0.19838900723976444,0.2637132351787373,0.21785087922339205,0.2637132351787373,0.22524644687889922,0.285894722249846,0.19838923776011905,0.2637132351787373
180
+ 14,11,0.19672871204204365,0.23660252875849325,0.217143601432742,0.23660252875849325,0.22318950414545896,0.26864245452787255,0.19672820772809663,0.23660252875849325
181
+ 14,-1,0.19672871204204365,0.23660252875849325,0.217143601432742,0.23660252875849325,0.22318950414545896,0.26864245452787255,0.19672820772809663,0.23660252875849325
182
+ 15,1,0.19503257463756532,0.23660252875849325,0.21548027686736262,0.23660252875849325,0.2206777391060753,0.2563194061550344,0.19503340556019388,0.23660252875849325
183
+ 15,2,0.18988697082729133,0.2242794803856551,0.21035178389952178,0.2242794803856551,0.21476020249162053,0.23413791908392564,0.1898873515369506,0.2242794803856551
184
+ 15,3,0.19741021641410966,0.21565334652466836,0.21701316629206413,0.21565334652466836,0.22080340850174474,0.24646096745676385,0.19741018425582973,0.21565334652466836
185
+ 15,4,0.1953580112942006,0.21565334652466836,0.21384620675812227,0.21565334652466836,0.21707556877653672,0.24276405294491238,0.19535830888361316,0.21565334652466836
186
+ 15,5,0.1877325539454797,0.23783483359577712,0.20592383870066416,0.23783483359577712,0.20868343677291756,0.24276405294491238,0.18773158542902008,0.23783483359577712
187
+ 15,6,0.17756617572810574,0.2242794803856551,0.19592203122135046,0.2242794803856551,0.19795067871718697,0.25139018680589914,0.1775654801860428,0.2242794803856551
188
+ 15,7,0.1699315699857664,0.25139018680589914,0.18845620017366085,0.25139018680589914,0.1902692362567422,0.2908239415989813,0.16993153815570905,0.25139018680589914
189
+ 15,8,0.17157198245996325,0.29821777062268423,0.18950402583012455,0.29821777062268423,0.19119917981152523,0.2908239415989813,0.17157178367151898,0.29821777062268423
190
+ 15,9,0.1705367041433481,0.29821777062268423,0.1880821359315564,0.29821777062268423,0.1891420473643982,0.2908239415989813,0.17053719989279525,0.29821777062268423
191
+ 15,10,0.17187946032874635,0.2821978077379946,0.18908697875655167,0.2821978077379946,0.18951488616362966,0.3031469899718195,0.17187867020251252,0.2821978077379946
192
+ 15,11,0.17066503465620295,0.2821978077379946,0.1876710328788344,0.2821978077379946,0.18718130151530843,0.3031469899718195,0.17066489697873294,0.2821978077379946
193
+ 15,-1,0.17066503465620295,0.2821978077379946,0.1876710328788344,0.2821978077379946,0.18718130151530843,0.3031469899718195,0.17066489697873294,0.2821978077379946
194
+ 16,1,0.17338767189530443,0.24522866261948,0.18902409858003566,0.24522866261948,0.1878010470009884,0.29821777062268423,0.1733875774171117,0.24522866261948
195
+ 16,2,0.17268379673898512,0.2575517109923182,0.18745450771102484,0.2575517109923182,0.18556862169917554,0.29821777062268423,0.17268393567492094,0.2575517109923182
196
+ 16,3,0.17600620789071925,0.24892557713133145,0.18914768171109342,0.24892557713133145,0.18693829008720342,0.2735716738770078,0.17600580434875657,0.24892557713133145
197
+ 16,4,0.17976355573419417,0.2649455400160211,0.19196362337363815,0.2649455400160211,0.18968833544489155,0.2649455400160211,0.17976291513927878,0.2649455400160211
198
+ 16,5,0.18057125954218883,0.2649455400160211,0.1918776017661153,0.2649455400160211,0.1897211373191512,0.2600163206668858,0.18057203505297392,0.2649455400160211
199
+ 16,6,0.17650205576843508,0.2649455400160211,0.1870583604592812,0.2649455400160211,0.18490030662021745,0.2600163206668858,0.17650222321092598,0.2649455400160211
200
+ 16,7,0.16863928992922705,0.24892557713133145,0.17849101135829767,0.24892557713133145,0.17643774634539577,0.2439963577821962,0.16863876134113884,0.24892557713133145
201
+ 16,8,0.15937079658761877,0.24892557713133145,0.16873127591586043,0.24892557713133145,0.16675036196068388,0.2439963577821962,0.1593709216932636,0.24892557713133145
202
+ 16,9,0.1501914865867161,0.2575517109923182,0.15999001874200128,0.2575517109923182,0.15774778199787887,0.2526224916431829,0.1501915640242195,0.2575517109923182
203
+ 16,10,0.14529891119705526,0.2575517109923182,0.15566328059574827,0.2575517109923182,0.15276687495367886,0.2526224916431829,0.14529914626346238,0.2575517109923182
204
+ 16,11,0.14895495246001422,0.23537022392120946,0.16002710673848786,0.23537022392120946,0.1569167539225269,0.29945007545996805,0.1489550423122351,0.23537022392120946
205
+ 16,-1,0.14895495246001422,0.23537022392120946,0.16002710673848786,0.23537022392120946,0.1569167539225269,0.29945007545996805,0.1489550423122351,0.23537022392120946
206
+ 17,1,0.15064217278505948,0.20949182233824928,0.16245939806841442,0.20949182233824928,0.1593052723848119,0.2846624174125622,0.15064162278683177,0.20949182233824928
207
+ 17,2,0.16221823549137818,0.25508710131775053,0.17481418394033071,0.25508710131775053,0.17160463510532764,0.2600163206668858,0.16221729084806905,0.25508710131775053
208
+ 17,3,0.17148644380314465,0.23783483359577712,0.18462725019095133,0.23783483359577712,0.18101153023581507,0.21565334652466836,0.171486480079936,0.23783483359577712
209
+ 17,4,0.17570093311427817,0.2267440900602227,0.18976948493476253,0.2267440900602227,0.18592753081874108,0.2255117852229389,0.1757005137806591,0.2267440900602227
210
+ 17,5,0.17027567494803714,0.2538547964804667,0.18588215033804467,0.2538547964804667,0.18184707326351873,0.2255117852229389,0.17027527940569642,0.2538547964804667
211
+ 17,6,0.16167565654373323,0.2698747593651564,0.17860460903718817,0.2698747593651564,0.17505625298643945,0.2575517109923182,0.1616753126879184,0.2698747593651564
212
+ 17,7,0.15664057142378163,0.2698747593651564,0.17473164842186045,0.2698747593651564,0.17169154544259577,0.2575517109923182,0.15664043139996314,0.2698747593651564
213
+ 17,8,0.15128033327870438,0.2538547964804667,0.16989909329083874,0.2538547964804667,0.167391781278147,0.2119564320128169,0.15128041680930576,0.2538547964804667
214
+ 17,9,0.14336524501296147,0.2267440900602227,0.1624946069534659,0.2267440900602227,0.15993418210723984,0.20702721266368163,0.14336619818759674,0.2267440900602227
215
+ 17,10,0.135259177218501,0.1873103352671405,0.154737222809034,0.1873103352671405,0.15218493583701637,0.2242794803856551,0.13525974823202358,0.1873103352671405
216
+ 17,11,0.13924582254935303,0.20702721266368163,0.1588515313695164,0.20702721266368163,0.15655967368940926,0.2242794803856551,0.13924651689490974,0.20702721266368163
217
+ 17,-1,0.13924582254935303,0.20702721266368163,0.1588515313695164,0.20702721266368163,0.15655967368940926,0.2242794803856551,0.13924651689490974,0.20702721266368163
218
+ 18,1,0.15110342932993154,0.23290561424664183,0.16994597756851143,0.23290561424664183,0.16848457870003597,0.23290561424664183,0.1511037630316804,0.23290561424664183
219
+ 18,2,0.15757616908249386,0.21688565136195218,0.17626582674912197,0.21688565136195218,0.17537290143522202,0.21688565136195218,0.15757708007532634,0.21688565136195218
220
+ 18,3,0.16067912986114902,0.21688565136195218,0.17898960751437834,0.21688565136195218,0.17864178490717775,0.23783483359577712,0.160678629920663,0.21688565136195218
221
+ 18,4,0.1607633763090634,0.21688565136195218,0.17917435451845926,0.21688565136195218,0.1787859967866911,0.23783483359577712,0.16076263584930806,0.21688565136195218
222
+ 18,5,0.15941706970314357,0.21688565136195218,0.17758773334789624,0.21688565136195218,0.17698661718793862,0.23783483359577712,0.15941719742178911,0.21688565136195218
223
+ 18,6,0.1620615096442195,0.23413791908392564,0.18011324842534798,0.23413791908392564,0.17950018187577355,0.23413791908392564,0.16206113026310023,0.23413791908392564
224
+ 18,7,0.16341788195566573,0.25015788196861527,0.18133540219389374,0.25015788196861527,0.1809313379368166,0.28712702708712984,0.16341789771841655,0.25015788196861527
225
+ 18,8,0.169493438531871,0.2661778448533049,0.18680974791924337,0.2661778448533049,0.18678999433856394,0.2661778448533049,0.16949358266575004,0.2661778448533049
226
+ 18,9,0.17633308567704503,0.27850089322614313,0.19324374433721908,0.27850089322614313,0.19388076395993986,0.2846624174125622,0.17633301674990937,0.27850089322614313
227
+ 18,10,0.19087072484603676,0.27850089322614313,0.20684713796652016,0.27850089322614313,0.2080792309529989,0.2895916367616975,0.19087065134785836,0.27850089322614313
228
+ 18,11,0.19847189769190748,0.27850089322614313,0.21358797307939892,0.27850089322614313,0.21495282956391104,0.26741014969058874,0.19847155779476103,0.27850089322614313
229
+ 18,-1,0.19847189769190748,0.27850089322614313,0.21358797307939892,0.27850089322614313,0.21495282956391104,0.26741014969058874,0.19847155779476103,0.27850089322614313
230
+ 19,1,0.20244936317906648,0.27850089322614313,0.2169766868226567,0.27850089322614313,0.21833672480258043,0.26741014969058874,0.2024497304814934,0.27850089322614313
231
+ 19,2,0.20717821745179166,0.2624809303414535,0.22150790629448788,0.2624809303414535,0.22329130977022538,0.2895916367616975,0.20717825220868558,0.2624809303414535
232
+ 19,3,0.21166556150494323,0.24646096745676385,0.22579073761692903,0.24646096745676385,0.22762844349431688,0.2895916367616975,0.21166505130951552,0.24646096745676385
233
+ 19,4,0.2094032488069487,0.2957531609481166,0.22425675291207414,0.2957531609481166,0.22630522066498562,0.30068238029725186,0.20940348641131362,0.2957531609481166
234
+ 19,5,0.21580767568263376,0.30068238029725186,0.23076083181569595,0.30068238029725186,0.23342322160836892,0.30068238029725186,0.21580823076741448,0.30068238029725186
235
+ 19,6,0.21912483105765843,0.30068238029725186,0.23384615668243788,0.30068238029725186,0.2368192791076227,0.2895916367616975,0.21912436161550827,0.30068238029725186
236
+ 19,7,0.22090117985454755,0.2735716738770078,0.2352271294400819,0.2735716738770078,0.23841673738439748,0.2895916367616975,0.22090103277392714,0.2735716738770078
237
+ 19,8,0.2221542839392439,0.24646096745676385,0.23615373073196888,0.24646096745676385,0.23942767824250474,0.2735716738770078,0.2221541260354549,0.24646096745676385
238
+ 19,9,0.2213649278332078,0.24646096745676385,0.23500896648610176,0.24646096745676385,0.238281154803522,0.25139018680589914,0.22136517049075843,0.24646096745676385
239
+ 19,10,0.216893886276899,0.24153174810762854,0.23078852388663615,0.24153174810762854,0.23396423061748353,0.27850089322614313,0.21689334353050999,0.24153174810762854
240
+ 19,11,0.21644813890713882,0.2624809303414535,0.2308051268472922,0.2624809303414535,0.2336066539927412,0.2563194061550344,0.2164468838735985,0.2624809303414535
241
+ 19,-1,0.21644813890713882,0.2624809303414535,0.2308051268472922,0.2624809303414535,0.2336066539927412,0.2563194061550344,0.2164468838735985,0.2624809303414535
242
+ 20,1,0.21790285927274675,0.19716877396541108,0.23244938679824148,0.19716877396541108,0.23552089174602533,0.26124862550416966,0.21790310521255213,0.19716877396541108
243
+ 20,2,0.21661690352405896,0.17375498205701848,0.23150890844812766,0.17375498205701848,0.2352335237487625,0.2439963577821962,0.21661651850214148,0.17375498205701848
244
+ 20,3,0.2118239581397238,0.17991650624343758,0.22691869940977544,0.17991650624343758,0.23114758622038623,0.22797639489750657,0.21182467009057637,0.17991650624343758
245
+ 20,4,0.2058445273661786,0.2119564320128169,0.2211983916854636,0.2119564320128169,0.22603069887625216,0.18484572559257287,0.20584479820106372,0.2119564320128169
246
+ 20,5,0.20371418165212593,0.2119564320128169,0.2192829693997825,0.2119564320128169,0.22480494923874358,0.18484572559257287,0.2037155415399145,0.2119564320128169
247
+ 20,6,0.20467918265643606,0.2119564320128169,0.22062978060544908,0.2119564320128169,0.22673653353455409,0.20702721266368163,0.20467895666610034,0.2119564320128169
248
+ 20,7,0.20955665156204117,0.20456260298911397,0.2261021608754628,0.20456260298911397,0.23268459552475385,0.19470416429084345,0.20955757945327202,0.20456260298911397
249
+ 20,8,0.2125986852105663,0.20456260298911397,0.22955492069573025,0.20456260298911397,0.23656818572875907,0.19470416429084345,0.2125991877480225,0.20456260298911397
250
+ 20,9,0.21722226300209463,0.20456260298911397,0.23429320373932414,0.20456260298911397,0.24174042427377895,0.21565334652466836,0.21722187355869355,0.20456260298911397
251
+ 20,10,0.2237215245946915,0.20456260298911397,0.24100924272174207,0.20456260298911397,0.24912427872968312,0.22058256587380362,0.22372116350796867,0.20456260298911397
252
+ 20,11,0.2302263295968998,0.2255117852229389,0.24764749575111294,0.2255117852229389,0.2562414836920526,0.22058256587380362,0.23022669168962942,0.2255117852229389
253
+ 20,-1,0.2302263295968998,0.2255117852229389,0.24764749575111294,0.2255117852229389,0.2562414836920526,0.22058256587380362,0.23022669168962942,0.2255117852229389
254
+ 21,1,0.23763378144470457,0.2255117852229389,0.25499119653275837,0.2255117852229389,0.26374283859903125,0.22058256587380362,0.23763228964454897,0.2255117852229389
255
+ 21,2,0.24407832967629886,0.27973319806342695,0.2608249094883375,0.27973319806342695,0.26940575591226695,0.24769327229404767,0.2440784737553442,0.27973319806342695
256
+ 21,3,0.2420136401539721,0.306843904483671,0.2588365988706655,0.306843904483671,0.26699915456607504,0.24276405294491238,0.24201348129304034,0.306843904483671
257
+ 21,4,0.23723464413735587,0.28096550290071076,0.2541196147334817,0.28096550290071076,0.2615500319995164,0.2698747593651564,0.23723535938969398,0.28096550290071076
258
+ 21,5,0.23325120450541373,0.25878401582960203,0.2499295153112741,0.25878401582960203,0.2568259361714696,0.24769327229404767,0.23325100592943734,0.25878401582960203
259
+ 21,6,0.23111425353289133,0.2267440900602227,0.2474834890079771,0.2267440900602227,0.2539826794497123,0.2267440900602227,0.2311149993831389,0.2267440900602227
260
+ 21,7,0.22806507986798166,0.2439963577821962,0.24398820687645317,0.2439963577821962,0.2503697991142557,0.2390671384330609,0.2280643630706427,0.2439963577821962
261
+ 21,8,0.2228951048536255,0.2563194061550344,0.23819576733295053,0.2563194061550344,0.24454684490545658,0.2563194061550344,0.222894539932585,0.2563194061550344
262
+ 21,9,0.21849201286116784,0.25139018680589914,0.2332861727174405,0.25139018680589914,0.23951937590657216,0.2563194061550344,0.21849178670211572,0.25139018680589914
263
+ 21,10,0.214650723320612,0.25139018680589914,0.2291538831604797,0.25139018680589914,0.23492473308451406,0.2563194061550344,0.21465118866972177,0.25139018680589914
264
+ 21,11,0.2103444221273414,0.22920869973479036,0.22465915667387681,0.22920869973479036,0.23013727524168287,0.2390671384330609,0.21034418672481342,0.22920869973479036
265
+ 21,-1,0.2103444221273414,0.22920869973479036,0.22465915667387681,0.22920869973479036,0.23013727524168287,0.2390671384330609,0.21034418672481342,0.22920869973479036
266
+ 22,1,0.2104207218044331,0.21688565136195218,0.2248428395256356,0.21688565136195218,0.22992702646791924,0.2390671384330609,0.2104205301218043,0.21688565136195218
267
+ 22,2,0.20940669790741406,0.21688565136195218,0.2239864660763574,0.21688565136195218,0.22861968657423037,0.21688565136195218,0.20940599342972044,0.21688565136195218
268
+ 22,3,0.20942311429198787,0.18484572559257287,0.22420504527247376,0.18484572559257287,0.2282879311304352,0.23783483359577712,0.2094226017045157,0.18484572559257287
269
+ 22,4,0.21108458008856587,0.2057949078263978,0.22585550149903597,0.2057949078263978,0.22949999735210552,0.17375498205701848,0.21108502792356798,0.2057949078263978
270
+ 22,5,0.21751511703563597,0.17375498205701848,0.2318080192553722,0.17375498205701848,0.2350445574395592,0.15650271433504503,0.21751524361776645,0.17375498205701848
271
+ 22,6,0.22271676470759016,0.19840107880269486,0.23668969573407972,0.19840107880269486,0.23901450365001858,0.18114881108072142,0.22271640960212477,0.19840107880269486
272
+ 22,7,0.2298469611026396,0.16512884819603177,0.24330051771056962,0.16512884819603177,0.24494323469331097,0.16512884819603177,0.2298468215596437,0.16512884819603177
273
+ 22,8,0.2388270374481029,0.16512884819603177,0.25157652792705293,0.16512884819603177,0.25304323782433785,0.16512884819603177,0.23882714961394447,0.16512884819603177
274
+ 22,9,0.24995951845894052,0.18114881108072142,0.2617972715107902,0.18114881108072142,0.2631939023187334,0.19716877396541108,0.24995961394241392,0.18114881108072142
275
+ 22,10,0.2591824053795885,0.18114881108072142,0.2698246159039843,0.18114881108072142,0.27172620694827243,0.19716877396541108,0.25918177015407406,0.18114881108072142
276
+ 22,11,0.2691208475875445,0.19840107880269486,0.2784871273056604,0.19840107880269486,0.28094492719372377,0.19716877396541108,0.26912101411266404,0.19840107880269486
277
+ 22,-1,0.2691208475875445,0.19840107880269486,0.2784871273056604,0.19840107880269486,0.28094492719372377,0.19716877396541108,0.26912101411266404,0.19840107880269486
278
+ 23,1,0.27951696348716126,0.19840107880269486,0.2874724802773863,0.19840107880269486,0.2904700022119344,0.2131887368501007,0.27951828053209465,0.19840107880269486
279
+ 23,2,0.2901036702959942,0.24646096745676385,0.296815014878719,0.24646096745676385,0.3002665299817445,0.22920869973479036,0.29010506494720767,0.24646096745676385
280
+ 23,3,0.30348242986763196,0.24646096745676385,0.308412280512067,0.24646096745676385,0.31200020191287237,0.24646096745676385,0.3034819893828702,0.24646096745676385
281
+ 23,4,0.31523151432522956,0.23044100457207417,0.31859682448953863,0.23044100457207417,0.32271194201457276,0.24646096745676385,0.31523136253114287,0.23044100457207417
282
+ 23,5,0.32367584956137735,0.2624809303414535,0.32604240253877814,0.2624809303414535,0.33026200128166505,0.24646096745676385,0.323676017962642,0.2624809303414535
283
+ 23,6,0.33096988807106115,0.333954610903915,0.33252191572854595,0.333954610903915,0.3367184389145448,0.28096550290071076,0.33097107896000444,0.333954610903915
284
+ 23,7,0.3359157797545079,0.35120687862588845,0.3370854069441667,0.35120687862588845,0.34130647415966897,0.3672268415105781,0.3359159881548365,0.35120687862588845
285
+ 23,8,0.3401329754115403,0.35120687862588845,0.34070781583425225,0.35120687862588845,0.3449818113225649,0.3672268415105781,0.3401326385488521,0.35120687862588845
286
+ 23,9,0.34261784713191235,0.3388838302530503,0.3432250342183023,0.3388838302530503,0.34743211008113495,0.3672268415105781,0.34261780689957455,0.3388838302530503
287
+ 23,10,0.3424996149027535,0.3549037931377399,0.34345873460325316,0.3549037931377399,0.3478865745438525,0.3672268415105781,0.3424990193485822,0.3549037931377399
288
+ 23,11,0.343308324101755,0.37215606085971337,0.3440402517881779,0.37215606085971337,0.3498407762580593,0.37215606085971337,0.34330842675387746,0.37215606085971337
289
+ 23,-1,0.343308324101755,0.37215606085971337,0.3440402517881779,0.37215606085971337,0.3498407762580593,0.37215606085971337,0.34330842675387746,0.37215606085971337
290
+ 24,1,0.3439649634759213,0.3425807447649017,0.3442398114110046,0.3425807447649017,0.35138592583420686,0.37215606085971337,0.34396440526331007,0.3425807447649017
291
+ 24,2,0.34181682828936977,0.320399257693793,0.34158418640436083,0.320399257693793,0.35012750558891625,0.34997457378860464,0.3418170991644909,0.320399257693793
292
+ 24,3,0.33635724741261974,0.3142377335073739,0.3359892065761548,0.3142377335073739,0.34550393033325955,0.37215606085971337,0.33635765178446364,0.3142377335073739
293
+ 24,4,0.3309025778436569,0.3142377335073739,0.3308472732481368,0.3142377335073739,0.34080286621383715,0.3043792948091033,0.3309025610973583,0.3142377335073739
294
+ 24,5,0.3242081662428456,0.28712702708712984,0.32486211348671556,0.28712702708712984,0.3348636561205609,0.3043792948091033,0.3242080168754452,0.28712702708712984
295
+ 24,6,0.31814655846488615,0.28712702708712984,0.3196827779110608,0.28712702708712984,0.32990990242934415,0.3043792948091033,0.31814682395307475,0.28712702708712984
296
+ 24,7,0.31277102552033986,0.2821978077379946,0.31526278321598666,0.2821978077379946,0.3257423693601644,0.3043792948091033,0.3127715356205655,0.2821978077379946
297
+ 24,8,0.3060855375287062,0.2895916367616975,0.30992021779657863,0.2895916367616975,0.320369138954156,0.3043792948091033,0.3060861499384906,0.2895916367616975
298
+ 24,9,0.30019598932483915,0.2895916367616975,0.3053479317439901,0.2895916367616975,0.31558671512940195,0.272339369039724,0.3001954211286589,0.2895916367616975
299
+ 24,10,0.29652350246108544,0.31916695285650915,0.3027286159912135,0.31916695285650915,0.3125708151098439,0.26741014969058874,0.29652320096405627,0.31916695285650915
300
+ 24,11,0.29263871345754283,0.31916695285650915,0.3000254906361302,0.31916695285650915,0.3094614538671614,0.26741014969058874,0.29263794139622734,0.31916695285650915
301
+ 24,-1,0.29263871345754283,0.31916695285650915,0.3000254906361302,0.31916695285650915,0.3094614538671614,0.26741014969058874,0.29263794139622734,0.31916695285650915
302
+ 25,1,0.2903802129383176,0.30191468513453573,0.29863582416943973,0.30191468513453573,0.30753566626357587,0.24276405294491238,0.29038077628489917,0.30191468513453573
303
+ 25,2,0.28856986068156704,0.2895916367616975,0.2972386929398535,0.2895916367616975,0.305842982141285,0.24276405294491238,0.28857006071955693,0.2895916367616975
304
+ 25,3,0.2875341702305584,0.2895916367616975,0.2962747166399503,0.2895916367616975,0.3047978105992282,0.24044682104762805,0.28753415456942805,0.2895916367616975
305
+ 25,4,0.28755803779305145,0.2895916367616975,0.29577717806419035,0.2895916367616975,0.304539575161265,0.26741014969058874,0.2875577572158785,0.2895916367616975
306
+ 25,5,0.2868138450598453,0.2649455400160211,0.2946411436381652,0.2649455400160211,0.3036705785093783,0.29945007545996805,0.2868138028885059,0.2649455400160211
307
+ 25,6,0.28539600004216265,0.24522866261948,0.2930728043443759,0.24522866261948,0.3020287199412134,0.2945208561108328,0.2853951870623113,0.24522866261948
308
+ 25,7,0.28362151001036096,0.26741014969058874,0.29142279296381574,0.26741014969058874,0.29994433189704506,0.2945208561108328,0.2836204216577183,0.26741014969058874
309
+ 25,8,0.281418418049612,0.26741014969058874,0.28928451664759164,0.26741014969058874,0.29758516734689217,0.2945208561108328,0.2814192556189725,0.26741014969058874
310
+ 25,9,0.2789362295907404,0.25139018680589914,0.28686763684168837,0.25139018680589914,0.29501630862351386,0.2957531609481166,0.2789364219751162,0.25139018680589914
311
+ 25,10,0.2769259961029197,0.24646096745676385,0.28499079737230043,0.24646096745676385,0.29278248492570247,0.31177312383280625,0.27692574137214,0.24646096745676385
312
+ 25,11,0.276112180243415,0.26864245452787255,0.2846372276673429,0.26864245452787255,0.29191118330757293,0.2895916367616975,0.27611218678912763,0.26864245452787255
313
+ 25,-1,0.276112180243415,0.26864245452787255,0.2846372276673429,0.26864245452787255,0.29191118330757293,0.2895916367616975,0.27611218678912763,0.26864245452787255
314
+ 26,1,0.2745833417125594,0.26864245452787255,0.28385880794066465,0.26864245452787255,0.2904940848311658,0.26864245452787255,0.2745824549120313,0.26864245452787255
315
+ 26,2,0.27366914481120336,0.26864245452787255,0.2840331217926604,0.26864245452787255,0.29021726209225823,0.2735716738770078,0.2736691792868328,0.26864245452787255
316
+ 26,3,0.27059406053637064,0.2735716738770078,0.28216777692375755,0.2735716738770078,0.2878649434905406,0.2821978077379946,0.2705945321952107,0.2735716738770078
317
+ 26,4,0.2685898112561523,0.2821978077379946,0.2814363327543946,0.2821978077379946,0.2866888817489898,0.2821978077379946,0.2685898197021195,0.2821978077379946
318
+ 26,5,0.2659737983583446,0.3043792948091033,0.2797577737611009,0.3043792948091033,0.28460697369939053,0.2772685883888593,0.26597425150706505,0.3043792948091033
319
+ 26,6,0.2646113772564346,0.3364192205784826,0.27901067904672805,0.3364192205784826,0.28360719392348843,0.2575517109923182,0.2646110809820969,0.3364192205784826
320
+ 26,7,0.2632157887080327,0.3364192205784826,0.2780791030792208,0.3364192205784826,0.28255584029947955,0.2575517109923182,0.2632161794672556,0.3364192205784826
321
+ 26,8,0.2625864234935468,0.3142377335073739,0.2777892665806698,0.3142377335073739,0.28228413577834455,0.2624809303414535,0.2625860020163817,0.3142377335073739
322
+ 26,9,0.2633616626695314,0.3056115996463871,0.278783011489028,0.3056115996463871,0.28343431480089676,0.2624809303414535,0.26336241316375475,0.3056115996463871
323
+ 26,10,0.2648801227109782,0.29328855127354897,0.280507396765416,0.29328855127354897,0.2851343600883496,0.2575517109923182,0.26488003459015247,0.29328855127354897
324
+ 26,11,0.2628477920375865,0.2661778448533049,0.27919545790367234,0.2661778448533049,0.28387005607841975,0.27850089322614313,0.26284669829525276,0.2661778448533049
325
+ 26,-1,0.2628477920375865,0.2661778448533049,0.27919545790367234,0.2661778448533049,0.28387005607841975,0.27850089322614313,0.26284669829525276,0.2661778448533049
326
+ 27,1,0.26037732141491055,0.2661778448533049,0.27755300026619056,0.2661778448533049,0.2820474634554743,0.28712702708712984,0.260377562585797,0.2661778448533049
327
+ 27,2,0.25740733339210453,0.2698747593651564,0.2754057496006107,0.2698747593651564,0.27970746517934075,0.2698747593651564,0.25740772780865934,0.2698747593651564
328
+ 27,3,0.25430995839624326,0.2698747593651564,0.2731466469528294,0.2698747593651564,0.2772391927913399,0.2748039787142917,0.25431098334195956,0.2698747593651564
329
+ 27,4,0.2515557130791495,0.23167330940935799,0.2711435249191556,0.23167330940935799,0.275271157964333,0.25878401582960203,0.2515566084403732,0.23167330940935799
330
+ 27,5,0.24795630972765814,0.2526224916431829,0.26834474290942323,0.2526224916431829,0.27257585213271196,0.2748039787142917,0.24795653180095878,0.2526224916431829
331
+ 27,6,0.24431727862011657,0.2526224916431829,0.2652862304085236,0.2526224916431829,0.26961005958747364,0.2526224916431829,0.24431682065619834,0.2526224916431829
332
+ 27,7,0.24037666872998814,0.2526224916431829,0.2617620047894571,0.2526224916431829,0.26617167185192026,0.22058256587380362,0.24037718410557993,0.2526224916431829
333
+ 27,8,0.23675734229975923,0.2526224916431829,0.25850642398277146,0.2526224916431829,0.26292822950968353,0.22058256587380362,0.23675785254753007,0.2526224916431829
334
+ 27,9,0.2337052776007546,0.2526224916431829,0.25575007496968294,0.2526224916431829,0.26016519577529723,0.22058256587380362,0.23370487801579343,0.2526224916431829
335
+ 27,10,0.23094784311478866,0.2526224916431829,0.2532334950441617,0.2526224916431829,0.2576561772522699,0.22058256587380362,0.230947720638935,0.2526224916431829
336
+ 27,11,0.2285371396304352,0.2526224916431829,0.2510306046724148,0.2526224916431829,0.25546669642948494,0.22058256587380362,0.22853732676883884,0.2526224916431829
337
+ 27,-1,0.2285371396304352,0.2526224916431829,0.2510306046724148,0.2526224916431829,0.25546669642948494,0.22058256587380362,0.22853732676883884,0.2526224916431829
338
+ 28,1,0.22671416220017773,0.22058256587380362,0.2493890178805433,0.22058256587380362,0.25379685491873727,0.22058256587380362,0.2267139855732639,0.22058256587380362
339
+ 28,2,0.22441405588644991,0.22058256587380362,0.2472967855652912,0.22058256587380362,0.2516524478173412,0.22058256587380362,0.22441474387038493,0.22058256587380362
340
+ 28,3,0.22217052707157978,0.22058256587380362,0.24517857057074707,0.22058256587380362,0.2495164850432034,0.22058256587380362,0.2221705247858025,0.22058256587380362
341
+ 28,4,0.22028538695341332,0.22058256587380362,0.24325738255358736,0.22058256587380362,0.2475506742419486,0.22058256587380362,0.22028558568649578,0.22058256587380362
342
+ 28,5,0.218794071401049,0.24769327229404767,0.2415901744908704,0.24769327229404767,0.24587313810033767,0.22058256587380362,0.21879416077174835,0.24769327229404767
343
+ 28,6,0.21847669317492457,0.24769327229404767,0.24104103893406909,0.24769327229404767,0.24530630202103365,0.22058256587380362,0.21847680138730136,0.24769327229404767
344
+ 28,7,0.21876572427331004,0.24769327229404767,0.2410995358213498,0.24769327229404767,0.24532895391628842,0.22058256587380362,0.2187657193041901,0.24769327229404767
345
+ 28,8,0.21911666155492168,0.24769327229404767,0.24120236525248048,0.24769327229404767,0.24537016535573672,0.22058256587380362,0.2191166966553844,0.24769327229404767
346
+ 28,9,0.21962050421320783,0.24769327229404767,0.24153591848912082,0.24769327229404767,0.24563302044939683,0.19963338363997868,0.2196205168204713,0.24769327229404767
347
+ 28,10,0.22020786244408674,0.23167330940935799,0.24197380508288843,0.23167330940935799,0.24602256988355514,0.23167330940935799,0.22020854015503744,0.23167330940935799
348
+ 28,11,0.2194965683425255,0.24769327229404767,0.24119036455487552,0.24769327229404767,0.2451258465619992,0.2661778448533049,0.21949676579223765,0.24769327229404767
349
+ 28,-1,0.2194965683425255,0.24769327229404767,0.24119036455487552,0.24769327229404767,0.2451258465619992,0.2661778448533049,0.21949676579223765,0.24769327229404767
350
+ 29,1,0.2189417482187867,0.24769327229404767,0.24051494272512763,0.24769327229404767,0.24436732866609856,0.2661778448533049,0.21894168067996223,0.24769327229404767
351
+ 29,2,0.218534451413761,0.2267440900602227,0.2400271860639394,0.2267440900602227,0.24379931337072724,0.25015788196861527,0.2185342193800181,0.2267440900602227
352
+ 29,3,0.2182327383280021,0.2267440900602227,0.23961726660392196,0.2267440900602227,0.2433484693677768,0.25015788196861527,0.21823279345168153,0.2267440900602227
353
+ 29,4,0.21798074395425995,0.2267440900602227,0.23923333729821863,0.2267440900602227,0.24293855314676263,0.26741014969058874,0.21797974187603208,0.2267440900602227
354
+ 29,5,0.2178053263843235,0.2267440900602227,0.2389560534607959,0.2267440900602227,0.24264407529729784,0.26741014969058874,0.21780642424115354,0.2267440900602227
355
+ 29,6,0.21784143776488402,0.2267440900602227,0.23891253117268318,0.2267440900602227,0.24258756313928223,0.26741014969058874,0.2178412306884152,0.2267440900602227
356
+ 29,7,0.21802371347873684,0.2267440900602227,0.23902279173751895,0.2267440900602227,0.24269698186208666,0.26741014969058874,0.21802309810872045,0.2267440900602227
357
+ 29,8,0.2181027943948082,0.2267440900602227,0.23906475968158902,0.2267440900602227,0.2427297142378175,0.26741014969058874,0.21810302405206672,0.2267440900602227
358
+ 29,9,0.21817804982188027,0.2267440900602227,0.23911245162686506,0.2267440900602227,0.24276357179725064,0.26741014969058874,0.21817798353377005,0.2267440900602227
359
+ 29,10,0.218313225715346,0.2267440900602227,0.23921646200970836,0.2267440900602227,0.2428640200208659,0.26741014969058874,0.218312391063022,0.2267440900602227
360
+ 29,11,0.21837737313410713,0.2267440900602227,0.23926720400008036,0.2267440900602227,0.24291448739968716,0.26741014969058874,0.2183777195957137,0.2267440900602227
361
+ 29,-1,0.21837737313410713,0.2267440900602227,0.23926720400008036,0.2267440900602227,0.24291448739968716,0.26741014969058874,0.2183777195957137,0.2267440900602227
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/30_Epochs/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/30_Epochs/modules.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ },
14
+ {
15
+ "idx": 2,
16
+ "name": "2",
17
+ "path": "2_Normalize",
18
+ "type": "sentence_transformers.models.Normalize"
19
+ }
20
+ ]
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/30_Epochs/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:295f5328dc1fafb55fb2b0ee7d507e3ae71152af3345b301dc0f177c10f99cd2
3
+ size 328509745
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/30_Epochs/sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 512,
3
+ "do_lower_case": false
4
+ }
codeScripts/Dependencies/BERT-models/Prueba3/Prueba_anterior/Model_all-distilroberta-v1/30_Epochs/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "sep_token": "</s>", "pad_token": "<pad>", "cls_token": "<s>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": false}}