96abhishekarora commited on
Commit
e9a3f59
1 Parent(s): fa3c456

Modified validation and training for linktransformer model

Browse files
Files changed (4) hide show
  1. .gitattributes +1 -0
  2. LT_training_config.json +2 -2
  3. README.md +5 -5
  4. pytorch_model.bin +1 -1
.gitattributes CHANGED
@@ -34,3 +34,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  pytorch_model.bin filter=lfs diff=lfs merge=lfs -text
 
 
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  pytorch_model.bin filter=lfs diff=lfs merge=lfs -text
37
+ .git/lfs/objects/ac/e3/ace3c71b97a96f01b2feceec24958b690700ea37ad188af305b32ef472101546 filter=lfs diff=lfs merge=lfs -text
LT_training_config.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
  "model_save_dir": "models",
3
  "model_save_name": "linkage_un_data_en_fine_fine",
4
- "opt_model_description": "This model was trained on a dataset prepared by linking product classifications from [UN stats](https://unstats.un.org/unsd/classifications/Econ). \n This model is designed to link different products together - trained on variation brought on by product level correspondance. It was trained for 30 epochs using other defaults that can be found in the repo's LinkTransformer config file - LT_training_config.json \n ",
5
  "opt_model_lang": "en",
6
  "train_batch_size": 64,
7
- "num_epochs": 30,
8
  "warm_up_perc": 1,
9
  "learning_rate": 2e-06,
10
  "val_perc": 0.2,
 
1
  {
2
  "model_save_dir": "models",
3
  "model_save_name": "linkage_un_data_en_fine_fine",
4
+ "opt_model_description": "This model was trained on a dataset prepared by linking product classifications from [UN stats](https://unstats.un.org/unsd/classifications/Econ). \n This model is designed to link different products together - trained on variation brought on by product level correspondance. It was trained for 100 epochs using other defaults that can be found in the repo's LinkTransformer config file - LT_training_config.json \n ",
5
  "opt_model_lang": "en",
6
  "train_batch_size": 64,
7
+ "num_epochs": 100,
8
  "warm_up_perc": 1,
9
  "learning_rate": 2e-06,
10
  "val_perc": 0.2,
README.md CHANGED
@@ -23,7 +23,7 @@ This model has been fine-tuned on the model : multi-qa-mpnet-base-dot-v1. It is
23
 
24
 
25
  This model was trained on a dataset prepared by linking product classifications from [UN stats](https://unstats.un.org/unsd/classifications/Econ).
26
- This model is designed to link different products together - trained on variation brought on by product level correspondance. It was trained for 30 epochs using other defaults that can be found in the repo's LinkTransformer config file - LT_training_config.json
27
 
28
 
29
  ## Usage (LinkTransformer)
@@ -97,7 +97,7 @@ The model was trained with the parameters:
97
 
98
  **DataLoader**:
99
 
100
- `torch.utils.data.dataloader.DataLoader` of length 147 with parameters:
101
  ```
102
  {'batch_size': 64, 'sampler': 'torch.utils.data.dataloader._InfiniteConstantSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'}
103
  ```
@@ -109,8 +109,8 @@ The model was trained with the parameters:
109
  Parameters of the fit()-Method:
110
  ```
111
  {
112
- "epochs": 30,
113
- "evaluation_steps": 1470,
114
  "evaluator": "sentence_transformers.evaluation.SequentialEvaluator.SequentialEvaluator",
115
  "max_grad_norm": 1,
116
  "optimizer_class": "<class 'torch.optim.adamw.AdamW'>",
@@ -119,7 +119,7 @@ Parameters of the fit()-Method:
119
  },
120
  "scheduler": "WarmupLinear",
121
  "steps_per_epoch": null,
122
- "warmup_steps": 4410,
123
  "weight_decay": 0.01
124
  }
125
  ```
 
23
 
24
 
25
  This model was trained on a dataset prepared by linking product classifications from [UN stats](https://unstats.un.org/unsd/classifications/Econ).
26
+ This model is designed to link different products together - trained on variation brought on by product level correspondance. It was trained for 100 epochs using other defaults that can be found in the repo's LinkTransformer config file - LT_training_config.json
27
 
28
 
29
  ## Usage (LinkTransformer)
 
97
 
98
  **DataLoader**:
99
 
100
+ `torch.utils.data.dataloader.DataLoader` of length 150 with parameters:
101
  ```
102
  {'batch_size': 64, 'sampler': 'torch.utils.data.dataloader._InfiniteConstantSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'}
103
  ```
 
109
  Parameters of the fit()-Method:
110
  ```
111
  {
112
+ "epochs": 100,
113
+ "evaluation_steps": 1500,
114
  "evaluator": "sentence_transformers.evaluation.SequentialEvaluator.SequentialEvaluator",
115
  "max_grad_norm": 1,
116
  "optimizer_class": "<class 'torch.optim.adamw.AdamW'>",
 
119
  },
120
  "scheduler": "WarmupLinear",
121
  "steps_per_epoch": null,
122
+ "warmup_steps": 15000,
123
  "weight_decay": 0.01
124
  }
125
  ```
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ace3c71b97a96f01b2feceec24958b690700ea37ad188af305b32ef472101546
3
  size 438009257
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1caf77aeeb33f22a280e7dd98da99969ecbc0b510530aacd240d85af85b27f5
3
  size 438009257