Fine-tuned on QA dataset
Browse files- README.md +3 -3
- model.safetensors +1 -1
README.md
CHANGED
@@ -83,7 +83,7 @@ The model was trained with the parameters:
|
|
83 |
|
84 |
**DataLoader**:
|
85 |
|
86 |
-
`torch.utils.data.dataloader.DataLoader` of length
|
87 |
```
|
88 |
{'batch_size': 8, 'sampler': 'torch.utils.data.sampler.RandomSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'}
|
89 |
```
|
@@ -99,7 +99,7 @@ Parameters of the fit()-Method:
|
|
99 |
```
|
100 |
{
|
101 |
"epochs": 5,
|
102 |
-
"evaluation_steps":
|
103 |
"evaluator": "sentence_transformers.evaluation.EmbeddingSimilarityEvaluator.EmbeddingSimilarityEvaluator",
|
104 |
"max_grad_norm": 1,
|
105 |
"optimizer_class": "<class 'torch.optim.adamw.AdamW'>",
|
@@ -108,7 +108,7 @@ Parameters of the fit()-Method:
|
|
108 |
},
|
109 |
"scheduler": "WarmupLinear",
|
110 |
"steps_per_epoch": null,
|
111 |
-
"warmup_steps":
|
112 |
"weight_decay": 0.01
|
113 |
}
|
114 |
```
|
|
|
83 |
|
84 |
**DataLoader**:
|
85 |
|
86 |
+
`torch.utils.data.dataloader.DataLoader` of length 1106 with parameters:
|
87 |
```
|
88 |
{'batch_size': 8, 'sampler': 'torch.utils.data.sampler.RandomSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'}
|
89 |
```
|
|
|
99 |
```
|
100 |
{
|
101 |
"epochs": 5,
|
102 |
+
"evaluation_steps": 10000,
|
103 |
"evaluator": "sentence_transformers.evaluation.EmbeddingSimilarityEvaluator.EmbeddingSimilarityEvaluator",
|
104 |
"max_grad_norm": 1,
|
105 |
"optimizer_class": "<class 'torch.optim.adamw.AdamW'>",
|
|
|
108 |
},
|
109 |
"scheduler": "WarmupLinear",
|
110 |
"steps_per_epoch": null,
|
111 |
+
"warmup_steps": 552,
|
112 |
"weight_decay": 0.01
|
113 |
}
|
114 |
```
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1340612432
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a89bea4c45e35de93083ce2515bec94634cc66f8caeb6dcce5e91ac3b27878f4
|
3 |
size 1340612432
|