96abhishekarora
commited on
Commit
•
e9a3f59
1
Parent(s):
fa3c456
Modified validation and training for linktransformer model
Browse files- .gitattributes +1 -0
- LT_training_config.json +2 -2
- README.md +5 -5
- pytorch_model.bin +1 -1
.gitattributes
CHANGED
@@ -34,3 +34,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
pytorch_model.bin filter=lfs diff=lfs merge=lfs -text
|
|
|
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
pytorch_model.bin filter=lfs diff=lfs merge=lfs -text
|
37 |
+
.git/lfs/objects/ac/e3/ace3c71b97a96f01b2feceec24958b690700ea37ad188af305b32ef472101546 filter=lfs diff=lfs merge=lfs -text
|
LT_training_config.json
CHANGED
@@ -1,10 +1,10 @@
|
|
1 |
{
|
2 |
"model_save_dir": "models",
|
3 |
"model_save_name": "linkage_un_data_en_fine_fine",
|
4 |
-
"opt_model_description": "This model was trained on a dataset prepared by linking product classifications from [UN stats](https://unstats.un.org/unsd/classifications/Econ). \n This model is designed to link different products together - trained on variation brought on by product level correspondance. It was trained for
|
5 |
"opt_model_lang": "en",
|
6 |
"train_batch_size": 64,
|
7 |
-
"num_epochs":
|
8 |
"warm_up_perc": 1,
|
9 |
"learning_rate": 2e-06,
|
10 |
"val_perc": 0.2,
|
|
|
1 |
{
|
2 |
"model_save_dir": "models",
|
3 |
"model_save_name": "linkage_un_data_en_fine_fine",
|
4 |
+
"opt_model_description": "This model was trained on a dataset prepared by linking product classifications from [UN stats](https://unstats.un.org/unsd/classifications/Econ). \n This model is designed to link different products together - trained on variation brought on by product level correspondance. It was trained for 100 epochs using other defaults that can be found in the repo's LinkTransformer config file - LT_training_config.json \n ",
|
5 |
"opt_model_lang": "en",
|
6 |
"train_batch_size": 64,
|
7 |
+
"num_epochs": 100,
|
8 |
"warm_up_perc": 1,
|
9 |
"learning_rate": 2e-06,
|
10 |
"val_perc": 0.2,
|
README.md
CHANGED
@@ -23,7 +23,7 @@ This model has been fine-tuned on the model : multi-qa-mpnet-base-dot-v1. It is
|
|
23 |
|
24 |
|
25 |
This model was trained on a dataset prepared by linking product classifications from [UN stats](https://unstats.un.org/unsd/classifications/Econ).
|
26 |
-
This model is designed to link different products together - trained on variation brought on by product level correspondance. It was trained for
|
27 |
|
28 |
|
29 |
## Usage (LinkTransformer)
|
@@ -97,7 +97,7 @@ The model was trained with the parameters:
|
|
97 |
|
98 |
**DataLoader**:
|
99 |
|
100 |
-
`torch.utils.data.dataloader.DataLoader` of length
|
101 |
```
|
102 |
{'batch_size': 64, 'sampler': 'torch.utils.data.dataloader._InfiniteConstantSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'}
|
103 |
```
|
@@ -109,8 +109,8 @@ The model was trained with the parameters:
|
|
109 |
Parameters of the fit()-Method:
|
110 |
```
|
111 |
{
|
112 |
-
"epochs":
|
113 |
-
"evaluation_steps":
|
114 |
"evaluator": "sentence_transformers.evaluation.SequentialEvaluator.SequentialEvaluator",
|
115 |
"max_grad_norm": 1,
|
116 |
"optimizer_class": "<class 'torch.optim.adamw.AdamW'>",
|
@@ -119,7 +119,7 @@ Parameters of the fit()-Method:
|
|
119 |
},
|
120 |
"scheduler": "WarmupLinear",
|
121 |
"steps_per_epoch": null,
|
122 |
-
"warmup_steps":
|
123 |
"weight_decay": 0.01
|
124 |
}
|
125 |
```
|
|
|
23 |
|
24 |
|
25 |
This model was trained on a dataset prepared by linking product classifications from [UN stats](https://unstats.un.org/unsd/classifications/Econ).
|
26 |
+
This model is designed to link different products together - trained on variation brought on by product level correspondance. It was trained for 100 epochs using other defaults that can be found in the repo's LinkTransformer config file - LT_training_config.json
|
27 |
|
28 |
|
29 |
## Usage (LinkTransformer)
|
|
|
97 |
|
98 |
**DataLoader**:
|
99 |
|
100 |
+
`torch.utils.data.dataloader.DataLoader` of length 150 with parameters:
|
101 |
```
|
102 |
{'batch_size': 64, 'sampler': 'torch.utils.data.dataloader._InfiniteConstantSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'}
|
103 |
```
|
|
|
109 |
Parameters of the fit()-Method:
|
110 |
```
|
111 |
{
|
112 |
+
"epochs": 100,
|
113 |
+
"evaluation_steps": 1500,
|
114 |
"evaluator": "sentence_transformers.evaluation.SequentialEvaluator.SequentialEvaluator",
|
115 |
"max_grad_norm": 1,
|
116 |
"optimizer_class": "<class 'torch.optim.adamw.AdamW'>",
|
|
|
119 |
},
|
120 |
"scheduler": "WarmupLinear",
|
121 |
"steps_per_epoch": null,
|
122 |
+
"warmup_steps": 15000,
|
123 |
"weight_decay": 0.01
|
124 |
}
|
125 |
```
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 438009257
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a1caf77aeeb33f22a280e7dd98da99969ecbc0b510530aacd240d85af85b27f5
|
3 |
size 438009257
|