marquesafonso commited on
Commit
44f29a3
1 Parent(s): 2739c56

Add new SentenceTransformer model.

Browse files
Files changed (4) hide show
  1. README.md +4 -3
  2. config.json +1 -1
  3. model.safetensors +1 -1
  4. tokenizer_config.json +7 -0
README.md CHANGED
@@ -6,7 +6,8 @@ tags:
6
  - feature-extraction
7
  - sentence-similarity
8
  - transformers
9
-
 
10
  ---
11
 
12
  # marquesafonso/albertina-sts
@@ -101,7 +102,7 @@ The model was trained with the parameters:
101
  Parameters of the fit()-Method:
102
  ```
103
  {
104
- "epochs": 10,
105
  "evaluation_steps": 800,
106
  "evaluator": "sentence_transformers.evaluation.EmbeddingSimilarityEvaluator.EmbeddingSimilarityEvaluator",
107
  "max_grad_norm": 1,
@@ -111,7 +112,7 @@ Parameters of the fit()-Method:
111
  },
112
  "scheduler": "WarmupLinear",
113
  "steps_per_epoch": null,
114
- "warmup_steps": 51,
115
  "weight_decay": 0.01
116
  }
117
  ```
 
6
  - feature-extraction
7
  - sentence-similarity
8
  - transformers
9
+ datasets:
10
+ - assin2
11
  ---
12
 
13
  # marquesafonso/albertina-sts
 
102
  Parameters of the fit()-Method:
103
  ```
104
  {
105
+ "epochs": 5,
106
  "evaluation_steps": 800,
107
  "evaluator": "sentence_transformers.evaluation.EmbeddingSimilarityEvaluator.EmbeddingSimilarityEvaluator",
108
  "max_grad_norm": 1,
 
112
  },
113
  "scheduler": "WarmupLinear",
114
  "steps_per_epoch": null,
115
+ "warmup_steps": 25,
116
  "weight_decay": 0.01
117
  }
118
  ```
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "PORTULAN/albertina-100m-portuguese-ptpt-encoder",
3
  "architectures": [
4
  "DebertaModel"
5
  ],
 
1
  {
2
+ "_name_or_path": "marquesafonso/albertina-sts",
3
  "architectures": [
4
  "DebertaModel"
5
  ],
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:621a687aa7a9909a9d7ec65d2f69592c8dd91427aae35773c7b11382c5adf76a
3
  size 554429144
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f680335c2520653500d13083ae637696ad679d9174cf8730c4c7c6b94797f9a
3
  size 554429144
tokenizer_config.json CHANGED
@@ -50,10 +50,17 @@
50
  "eos_token": "[SEP]",
51
  "errors": "replace",
52
  "mask_token": "[MASK]",
 
53
  "model_max_length": 512,
 
54
  "pad_token": "[PAD]",
 
 
55
  "sep_token": "[SEP]",
 
56
  "tokenizer_class": "DebertaTokenizer",
 
 
57
  "unk_token": "[UNK]",
58
  "vocab_type": "gpt2"
59
  }
 
50
  "eos_token": "[SEP]",
51
  "errors": "replace",
52
  "mask_token": "[MASK]",
53
+ "max_length": 512,
54
  "model_max_length": 512,
55
+ "pad_to_multiple_of": null,
56
  "pad_token": "[PAD]",
57
+ "pad_token_type_id": 0,
58
+ "padding_side": "right",
59
  "sep_token": "[SEP]",
60
+ "stride": 0,
61
  "tokenizer_class": "DebertaTokenizer",
62
+ "truncation_side": "right",
63
+ "truncation_strategy": "longest_first",
64
  "unk_token": "[UNK]",
65
  "vocab_type": "gpt2"
66
  }