imvladikon commited on
Commit
1691eda
1 Parent(s): 8912500

Upload 11 files

Browse files
Files changed (5) hide show
  1. README.md +5 -5
  2. config.json +1 -1
  3. pytorch_model.bin +1 -1
  4. sentence_bert_config.json +1 -1
  5. tokenizer.json +1 -1
README.md CHANGED
@@ -85,9 +85,9 @@ The model was trained with the parameters:
85
 
86
  **DataLoader**:
87
 
88
- `torch.utils.data.dataloader.DataLoader` of length 49088 with parameters:
89
  ```
90
- {'batch_size': 8, 'sampler': 'torch.utils.data.sampler.RandomSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'}
91
  ```
92
 
93
  **Loss**:
@@ -101,7 +101,7 @@ Parameters of the fit()-Method:
101
  ```
102
  {
103
  "epochs": 1,
104
- "evaluation_steps": 4908,
105
  "evaluator": "sentence_transformers.evaluation.EmbeddingSimilarityEvaluator.EmbeddingSimilarityEvaluator",
106
  "max_grad_norm": 1,
107
  "optimizer_class": "<class 'torch.optim.adamw.AdamW'>",
@@ -110,7 +110,7 @@ Parameters of the fit()-Method:
110
  },
111
  "scheduler": "WarmupLinear",
112
  "steps_per_epoch": null,
113
- "warmup_steps": 4908,
114
  "weight_decay": 0.01
115
  }
116
  ```
@@ -119,7 +119,7 @@ Parameters of the fit()-Method:
119
  ## Full Model Architecture
120
  ```
121
  SentenceTransformer(
122
- (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: BertModel
123
  (1): Pooling({'word_embedding_dimension': 512, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False})
124
  )
125
  ```
 
85
 
86
  **DataLoader**:
87
 
88
+ `__main__.MultiDatasetDataLoader` of length 10819 with parameters:
89
  ```
90
+ {'batch_size': 'unknown'}
91
  ```
92
 
93
  **Loss**:
 
101
  ```
102
  {
103
  "epochs": 1,
104
+ "evaluation_steps": 500,
105
  "evaluator": "sentence_transformers.evaluation.EmbeddingSimilarityEvaluator.EmbeddingSimilarityEvaluator",
106
  "max_grad_norm": 1,
107
  "optimizer_class": "<class 'torch.optim.adamw.AdamW'>",
 
110
  },
111
  "scheduler": "WarmupLinear",
112
  "steps_per_epoch": null,
113
+ "warmup_steps": 500,
114
  "weight_decay": 0.01
115
  }
116
  ```
 
119
  ## Full Model Architecture
120
  ```
121
  SentenceTransformer(
122
+ (0): Transformer({'max_seq_length': 128, 'do_lower_case': False}) with Transformer model: BertModel
123
  (1): Pooling({'word_embedding_dimension': 512, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False})
124
  )
125
  ```
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "imvladikon/sentence_transformers_alephbertgimmel_small",
3
  "architectures": [
4
  "BertModel"
5
  ],
 
1
  {
2
+ "_name_or_path": "/media/robert/BC7CA8E37CA899A2/dev/sentence_transformers_hebrew/sentence_transformers_hebrew/output/training_paraphrases_imvladikon-sentence_transformers_alephbertgimmel_small-2023-11-26_22-50-39",
3
  "architectures": [
4
  "BertModel"
5
  ],
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ddd2bf9323017d3d800a345155763df407446069445830014ff29ef899522e41
3
  size 314717513
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aa79393cdeb74b6af5bd7ec7ce5d036fb3f7d87c894958d3503da21c8affc689
3
  size 314717513
sentence_bert_config.json CHANGED
@@ -1,4 +1,4 @@
1
  {
2
- "max_seq_length": 512,
3
  "do_lower_case": false
4
  }
 
1
  {
2
+ "max_seq_length": 128,
3
  "do_lower_case": false
4
  }
tokenizer.json CHANGED
@@ -2,7 +2,7 @@
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
- "max_length": 512,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
 
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
+ "max_length": 128,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },