Mayhem50 commited on
Commit
e02dfe9
1 Parent(s): 91a3b7c

Upload 13 files

Browse files
1_Pooling/config.json CHANGED
@@ -1,9 +1,5 @@
1
  {
2
  "word_embedding_dimension": 1024,
3
- "pooling_mode_cls_token": false,
4
- "pooling_mode_mean_tokens": false,
5
- "pooling_mode_max_tokens": false,
6
- "pooling_mode_mean_sqrt_len_tokens": false,
7
- "pooling_mode_weightedmean_tokens": true,
8
- "pooling_mode_lasttoken": false
9
  }
 
1
  {
2
  "word_embedding_dimension": 1024,
3
+ "position_start": 0,
4
+ "num_positions": 512
 
 
 
 
5
  }
1_Pooling/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17b5b722c730d9037743f6cbc1222e6c304f9659ae02911530578368a513e65c
3
+ size 2877
README.md CHANGED
@@ -8,7 +8,7 @@ tags:
8
 
9
  # {MODEL_NAME}
10
 
11
- This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 1024 dimensional dense vector space and can be used for tasks like clustering or semantic search.
12
 
13
  <!--- Describe your model here -->
14
 
@@ -45,29 +45,31 @@ The model was trained with the parameters:
45
 
46
  **DataLoader**:
47
 
48
- `sentence_transformers.datasets.NoDuplicatesDataLoader.NoDuplicatesDataLoader` of length 1100 with parameters:
49
  ```
50
- {'batch_size': 512}
51
  ```
52
 
53
  **Loss**:
54
 
55
- `sentence_transformers.losses.MultipleNegativesRankingLoss.MNRLGradCache`
56
 
57
  Parameters of the fit()-Method:
58
  ```
59
  {
60
- "epochs": 1,
61
- "evaluation_steps": 110,
62
- "evaluator": "sentence_transformers.evaluation.EmbeddingSimilarityEvaluator.EmbeddingSimilarityEvaluator",
63
  "max_grad_norm": 1,
64
  "optimizer_class": "<class 'transformers.optimization.AdamW'>",
65
  "optimizer_params": {
66
- "lr": 0.00032
 
 
67
  },
68
  "scheduler": "WarmupLinear",
69
  "steps_per_epoch": null,
70
- "warmup_steps": 110,
71
  "weight_decay": 0.01
72
  }
73
  ```
@@ -76,8 +78,8 @@ Parameters of the fit()-Method:
76
  ## Full Model Architecture
77
  ```
78
  SentenceTransformer(
79
- (0): Transformer({'max_seq_length': 75, 'do_lower_case': False}) with Transformer model: BloomModel
80
- (1): Pooling({'word_embedding_dimension': 1024, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': True, 'pooling_mode_lasttoken': False})
81
  )
82
  ```
83
 
 
8
 
9
  # {MODEL_NAME}
10
 
11
+ This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a None dimensional dense vector space and can be used for tasks like clustering or semantic search.
12
 
13
  <!--- Describe your model here -->
14
 
 
45
 
46
  **DataLoader**:
47
 
48
+ `torch.utils.data.dataloader.DataLoader` of length 30617 with parameters:
49
  ```
50
+ {'batch_size': 16, 'sampler': 'torch.utils.data.sampler.RandomSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'}
51
  ```
52
 
53
  **Loss**:
54
 
55
+ `sentence_transformers.losses.MSELoss.MSELoss`
56
 
57
  Parameters of the fit()-Method:
58
  ```
59
  {
60
+ "epochs": 3,
61
+ "evaluation_steps": 1000,
62
+ "evaluator": "sentence_transformers.evaluation.SequentialEvaluator.SequentialEvaluator",
63
  "max_grad_norm": 1,
64
  "optimizer_class": "<class 'transformers.optimization.AdamW'>",
65
  "optimizer_params": {
66
+ "correct_bias": false,
67
+ "eps": 1e-06,
68
+ "lr": 2e-05
69
  },
70
  "scheduler": "WarmupLinear",
71
  "steps_per_epoch": null,
72
+ "warmup_steps": 10000,
73
  "weight_decay": 0.01
74
  }
75
  ```
 
78
  ## Full Model Architecture
79
  ```
80
  SentenceTransformer(
81
+ (0): Transformer({'max_seq_length': 128, 'do_lower_case': False}) with Transformer model: BloomModel
82
+ (1): WeightedMeanPooling()
83
  )
84
  ```
85
 
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "bigscience/bloom-560m",
3
  "apply_residual_connection_post_layernorm": false,
4
  "architectures": [
5
  "BloomModel"
 
1
  {
2
+ "_name_or_path": "Mayhem50/sgpt-bloom-560M-nli",
3
  "apply_residual_connection_post_layernorm": false,
4
  "architectures": [
5
  "BloomModel"
eval/mse_evaluation_TED2020-en-fr-dev.tsv.gz_results.csv ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch,steps,MSE
2
+ 0,1000,1272.7741241455078
3
+ 0,2000,1239.7001266479492
4
+ 0,3000,1244.563388824463
5
+ 0,4000,1255.099868774414
6
+ 0,5000,1284.8063468933105
7
+ 0,6000,1265.7715797424316
8
+ 0,7000,1308.1769943237305
9
+ 0,8000,1277.3436546325684
10
+ 0,9000,1263.1890296936035
11
+ 0,10000,1308.1955909729004
12
+ 0,11000,1283.3735466003418
13
+ 0,12000,1262.7530097961426
14
+ 0,13000,1260.5589866638184
15
+ 0,14000,1253.4588813781738
16
+ 0,15000,1239.389419555664
17
+ 0,16000,1284.862232208252
18
+ 0,17000,1215.8602714538574
19
+ 0,18000,1208.6514472961426
20
+ 0,19000,1207.7702522277832
21
+ 0,20000,1203.3350944519043
22
+ 0,21000,1202.7355194091797
23
+ 0,22000,1193.0546760559082
24
+ 0,23000,1168.9787864685059
25
+ 0,24000,1169.53125
26
+ 0,25000,1167.289161682129
27
+ 0,26000,1183.1158638000488
28
+ 0,27000,1167.3190116882324
29
+ 0,28000,1135.9935760498047
30
+ 0,29000,1140.6807899475098
31
+ 0,30000,1136.8371963500977
32
+ 0,-1,1127.4292945861816
33
+ 1,31000,1103.3071517944336
34
+ 1,32000,1101.201057434082
35
+ 1,33000,1093.484878540039
36
+ 1,34000,1085.3593826293945
37
+ 1,35000,1087.46919631958
38
+ 1,36000,1088.6981964111328
39
+ 1,37000,1085.8448028564453
40
+ 1,38000,1079.0802955627441
41
+ 1,39000,1078.3958435058594
42
+ 1,40000,1067.711067199707
43
+ 1,41000,1070.4686164855957
44
+ 1,42000,1062.2697830200195
45
+ 1,43000,1062.1541023254395
46
+ 1,44000,1061.3683700561523
47
+ 1,45000,1068.0513381958008
48
+ 1,46000,1060.914134979248
49
+ 1,47000,1047.8565216064453
50
+ 1,48000,1049.0880012512207
51
+ 1,49000,1041.7594909667969
52
+ 1,50000,1047.9501724243164
53
+ 1,51000,1041.0179138183594
54
+ 1,52000,1047.9998588562012
55
+ 1,53000,1035.096263885498
eval/translation_evaluation_TED2020-en-fr-dev.tsv.gz_results.csv ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch,steps,src2trg,trg2src
2
+ 0,1000,0.963,0.96
3
+ 0,2000,0.965,0.964
4
+ 0,3000,0.963,0.962
5
+ 0,4000,0.967,0.963
6
+ 0,5000,0.966,0.959
7
+ 0,6000,0.97,0.963
8
+ 0,7000,0.966,0.962
9
+ 0,8000,0.968,0.966
10
+ 0,9000,0.974,0.964
11
+ 0,10000,0.969,0.96
12
+ 0,11000,0.972,0.965
13
+ 0,12000,0.969,0.956
14
+ 0,13000,0.962,0.957
15
+ 0,14000,0.97,0.963
16
+ 0,15000,0.969,0.967
17
+ 0,16000,0.967,0.959
18
+ 0,17000,0.971,0.962
19
+ 0,18000,0.97,0.964
20
+ 0,19000,0.965,0.962
21
+ 0,20000,0.968,0.965
22
+ 0,21000,0.972,0.967
23
+ 0,22000,0.968,0.965
24
+ 0,23000,0.972,0.966
25
+ 0,24000,0.967,0.957
26
+ 0,25000,0.972,0.961
27
+ 0,26000,0.97,0.959
28
+ 0,27000,0.969,0.962
29
+ 0,28000,0.967,0.968
30
+ 0,29000,0.968,0.963
31
+ 0,30000,0.969,0.965
32
+ 0,-1,0.968,0.961
33
+ 1,31000,0.968,0.964
34
+ 1,32000,0.969,0.967
35
+ 1,33000,0.971,0.966
36
+ 1,34000,0.972,0.965
37
+ 1,35000,0.973,0.968
38
+ 1,36000,0.973,0.97
39
+ 1,37000,0.973,0.966
40
+ 1,38000,0.971,0.967
41
+ 1,39000,0.972,0.969
42
+ 1,40000,0.973,0.967
43
+ 1,41000,0.972,0.967
44
+ 1,42000,0.97,0.967
45
+ 1,43000,0.968,0.967
46
+ 1,44000,0.97,0.96
47
+ 1,45000,0.969,0.965
48
+ 1,46000,0.972,0.964
49
+ 1,47000,0.97,0.967
50
+ 1,48000,0.972,0.965
51
+ 1,49000,0.971,0.969
52
+ 1,50000,0.973,0.969
53
+ 1,51000,0.971,0.969
54
+ 1,52000,0.973,0.967
55
+ 1,53000,0.972,0.967
modules.json CHANGED
@@ -8,7 +8,7 @@
8
  {
9
  "idx": 1,
10
  "name": "1",
11
- "path": "1_Pooling",
12
- "type": "sentence_transformers.models.Pooling"
13
  }
14
  ]
 
8
  {
9
  "idx": 1,
10
  "name": "1",
11
+ "path": "1_WeightedMeanPooling",
12
+ "type": "sentence_transformers.models.WeightedMeanPooling"
13
  }
14
  ]
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:315daacf1871ff28875d33e477925ed63831ef9a6db45fd01d44ae3feed61e33
3
  size 2236953889
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1085ad4f1f628a9ff63ec9b12a5a0b6258f11c982ebbbcc6fe16f5ea7c49c4ed
3
  size 2236953889
sentence_bert_config.json CHANGED
@@ -1,4 +1,4 @@
1
  {
2
- "max_seq_length": 75,
3
  "do_lower_case": false
4
  }
 
1
  {
2
+ "max_seq_length": 128,
3
  "do_lower_case": false
4
  }
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:796518d81847a3370a556a1c4bc940286069aec6dbbc8cb61deb5f696dac82f7
3
- size 14500694
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:872aa95025502051626b31ba2ddf6709b05318c68e3cfd501dd0ee688f166724
3
+ size 14500695
tokenizer_config.json CHANGED
@@ -3,7 +3,7 @@
3
  "bos_token": "<s>",
4
  "eos_token": "</s>",
5
  "model_max_length": 1000000000000000019884624838656,
6
- "name_or_path": "bigscience/bloom-560m",
7
  "pad_token": "<pad>",
8
  "padding_side": "left",
9
  "special_tokens_map_file": null,
 
3
  "bos_token": "<s>",
4
  "eos_token": "</s>",
5
  "model_max_length": 1000000000000000019884624838656,
6
+ "name_or_path": "Mayhem50/sgpt-bloom-560M-nli",
7
  "pad_token": "<pad>",
8
  "padding_side": "left",
9
  "special_tokens_map_file": null,