hieule commited on
Commit
ee9b6bd
1 Parent(s): a47edda

Push model

Browse files
1_Pooling/config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 768,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": true,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false
7
+ }
README.md CHANGED
Binary files a/README.md and b/README.md differ
 
added_tokens.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"<mask>": 64000}
bpe.codes ADDED
The diff for this file is too large to render. See raw diff
 
config.json CHANGED
@@ -1,24 +1,29 @@
1
  {
 
2
  "architectures": [
3
- "BertModel"
4
  ],
5
  "attention_probs_dropout_prob": 0.1,
 
6
  "classifier_dropout": null,
 
 
7
  "hidden_act": "gelu",
8
  "hidden_dropout_prob": 0.1,
9
  "hidden_size": 768,
10
  "initializer_range": 0.02,
11
  "intermediate_size": 3072,
12
- "layer_norm_eps": 1e-12,
13
- "max_position_embeddings": 512,
14
- "model_type": "bert",
15
  "num_attention_heads": 12,
16
  "num_hidden_layers": 12,
17
- "pad_token_id": 0,
18
  "position_embedding_type": "absolute",
 
19
  "torch_dtype": "float32",
20
- "transformers_version": "4.12.5",
21
- "type_vocab_size": 2,
22
  "use_cache": true,
23
- "vocab_size": 30522
24
  }
 
1
  {
2
+ "_name_or_path": "sentence_phobert_nli\\",
3
  "architectures": [
4
+ "RobertaModel"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
  "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
+ "gradient_checkpointing": false,
11
  "hidden_act": "gelu",
12
  "hidden_dropout_prob": 0.1,
13
  "hidden_size": 768,
14
  "initializer_range": 0.02,
15
  "intermediate_size": 3072,
16
+ "layer_norm_eps": 1e-05,
17
+ "max_position_embeddings": 258,
18
+ "model_type": "roberta",
19
  "num_attention_heads": 12,
20
  "num_hidden_layers": 12,
21
+ "pad_token_id": 1,
22
  "position_embedding_type": "absolute",
23
+ "tokenizer_class": "PhobertTokenizer",
24
  "torch_dtype": "float32",
25
+ "transformers_version": "4.11.3",
26
+ "type_vocab_size": 1,
27
  "use_cache": true,
28
+ "vocab_size": 64001
29
  }
config_sentence_transformers.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "2.2.0",
4
+ "transformers": "4.11.3",
5
+ "pytorch": "1.10.2+cpu"
6
+ }
7
+ }
eval/similarity_evaluation_results.csv ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ epoch,steps,cosine_pearson,cosine_spearman,euclidean_pearson,euclidean_spearman,manhattan_pearson,manhattan_spearman,dot_pearson,dot_spearman
2
+ 0,-1,0.8078286933987654,0.8068675782136558,0.7826592675151728,0.7868028226819643,0.7807693868686497,0.7844094389688113,0.750560538312266,0.7567512236263219
3
+ 1,-1,0.8141910092889064,0.8132247090159028,0.7927358135349715,0.7967180091229377,0.7912690373690494,0.7950951810589162,0.7629149612065824,0.7688506534784273
4
+ 2,-1,0.8213888261354507,0.8200802249399388,0.8014694226634118,0.8063164716361716,0.8002064148053362,0.8046039741807846,0.7731418158158698,0.7791783534950835
5
+ 3,-1,0.8230233654608273,0.821893616196692,0.8035552172896905,0.8082517954348161,0.8024512539715857,0.806933904809289,0.7800862039068843,0.7844405198389021
modules.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ }
14
+ ]
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0b8ae6acf1e3eef45dab826964740341eb3484bd3d09845b3bca2caa0dacbe09
3
- size 438007537
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f442320a1ae3cd68748245e47ca504703bb182f661af0e9acfd3e438aafeb104
3
+ size 540069681
sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 256,
3
+ "do_lower_case": false
4
+ }
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "sep_token": "</s>", "pad_token": "<pad>", "cls_token": "<s>", "mask_token": "<mask>"}
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "sep_token": "</s>", "cls_token": "<s>", "pad_token": "<pad>", "mask_token": "<mask>", "model_max_length": 256, "special_tokens_map_file": null, "tokenizer_file": null, "name_or_path": "sentence_phobert_nli\\", "tokenizer_class": "PhobertTokenizer"}
vocab.txt ADDED
The diff for this file is too large to render. See raw diff