asahi417 commited on
Commit
70289de
1 Parent(s): 5bb1f0f

model update

Browse files
analogy.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"distance_function": "cosine_similarity", "sat/test": 0.5675675675675675, "sat/valid": 0.6172106824925816, "u2/test": 0.5416666666666666, "u2/valid": 0.6096491228070176, "u4/test": 0.625, "u4/valid": 0.6203703703703703, "google/test": 0.98, "google/valid": 0.924, "bats/test": 0.8241206030150754, "bats/valid": 0.7787659811006115, "sat_full": 0.6122994652406417}
config.json CHANGED
@@ -1,11 +1,10 @@
1
  {
2
- "_name_or_path": "relbert_output/models/e.nce_logout.mask.roberta-large.0.000005.8.0.05.640/best_model",
3
  "architectures": [
4
  "RobertaModel"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
  "bos_token_id": 0,
8
- "classifier_dropout": null,
9
  "eos_token_id": 2,
10
  "gradient_checkpointing": false,
11
  "hidden_act": "gelu",
@@ -25,8 +24,7 @@
25
  "template": "I wasn\u2019t aware of this relationship, but I just read in the encyclopedia that <obj> is <subj>\u2019s <mask>",
26
  "template_mode": "manual"
27
  },
28
- "torch_dtype": "float32",
29
- "transformers_version": "4.20.1",
30
  "type_vocab_size": 1,
31
  "use_cache": true,
32
  "vocab_size": 50265
 
1
  {
2
+ "_name_or_path": "roberta-large",
3
  "architectures": [
4
  "RobertaModel"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
  "bos_token_id": 0,
 
8
  "eos_token_id": 2,
9
  "gradient_checkpointing": false,
10
  "hidden_act": "gelu",
 
24
  "template": "I wasn\u2019t aware of this relationship, but I just read in the encyclopedia that <obj> is <subj>\u2019s <mask>",
25
  "template_mode": "manual"
26
  },
27
+ "transformers_version": "4.6.1",
 
28
  "type_vocab_size": 1,
29
  "use_cache": true,
30
  "vocab_size": 50265
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:29859f297e143db63c248b437a38d26a5f01d66d1ed81e033bb42f5ad2420f2f
3
- size 1421566897
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0873367c71e0b57e0d3d69def9e2dd35173c4d7d819001004e83679b4ffc81b8
3
+ size 1421595889
special_tokens_map.json CHANGED
@@ -1,15 +1 @@
1
- {
2
- "bos_token": "<s>",
3
- "cls_token": "<s>",
4
- "eos_token": "</s>",
5
- "mask_token": {
6
- "content": "<mask>",
7
- "lstrip": true,
8
- "normalized": false,
9
- "rstrip": false,
10
- "single_word": false
11
- },
12
- "pad_token": "<pad>",
13
- "sep_token": "</s>",
14
- "unk_token": "<unk>"
15
- }
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "sep_token": "</s>", "pad_token": "<pad>", "cls_token": "<s>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": false}}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1,16 +1 @@
1
- {
2
- "add_prefix_space": false,
3
- "bos_token": "<s>",
4
- "cls_token": "<s>",
5
- "eos_token": "</s>",
6
- "errors": "replace",
7
- "mask_token": "<mask>",
8
- "model_max_length": 512,
9
- "name_or_path": "relbert_output/models/e.nce_logout.mask.roberta-large.0.000005.8.0.05.640/best_model",
10
- "pad_token": "<pad>",
11
- "sep_token": "</s>",
12
- "special_tokens_map_file": null,
13
- "tokenizer_class": "RobertaTokenizer",
14
- "trim_offsets": true,
15
- "unk_token": "<unk>"
16
- }
 
1
+ {"unk_token": "<unk>", "bos_token": "<s>", "eos_token": "</s>", "add_prefix_space": false, "errors": "replace", "sep_token": "</s>", "cls_token": "<s>", "pad_token": "<pad>", "mask_token": "<mask>", "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "roberta-large"}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
trainer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"model": "roberta-large", "max_length": 64, "mode": "mask", "data": "relbert/semeval2012_relational_similarity", "template_mode": "manual", "template": "I wasn\u2019t aware of this relationship, but I just read in the encyclopedia that <obj> is <subj>\u2019s <mask>", "loss_function": "nce_logout", "temperature_nce_constant": 0.05, "temperature_nce_rank": {"min": 0.01, "max": 0.05, "type": "linear"}, "epoch": 23, "batch": 128, "lr": 5e-06, "lr_decay": false, "lr_warmup": 1, "weight_decay": 0, "random_seed": 0, "exclude_relation": null, "n_sample": 640, "gradient_accumulation": 8}
validation_loss.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"validation_loss": 5.019965439700009, "validation_data": "relbert/semeval2012_relational_similarity", "validation_data/exclude_relation": null}