asahi417 commited on
Commit
c45b7e7
1 Parent(s): 6f78525

model update

Browse files
added_tokens.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"<hl>": 50265}
config.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "lmqg_output/optimized_no_answer/bart_large_squad/model_blrpij/epoch_2",
3
+ "activation_dropout": 0.1,
4
+ "activation_function": "gelu",
5
+ "add_bias_logits": false,
6
+ "add_final_layer_norm": false,
7
+ "architectures": [
8
+ "BartForConditionalGeneration"
9
+ ],
10
+ "attention_dropout": 0.1,
11
+ "bos_token_id": 0,
12
+ "classif_dropout": 0.1,
13
+ "classifier_dropout": 0.0,
14
+ "d_model": 1024,
15
+ "decoder_attention_heads": 16,
16
+ "decoder_ffn_dim": 4096,
17
+ "decoder_layerdrop": 0.0,
18
+ "decoder_layers": 12,
19
+ "decoder_start_token_id": 2,
20
+ "dropout": 0.1,
21
+ "early_stopping": true,
22
+ "encoder_attention_heads": 16,
23
+ "encoder_ffn_dim": 4096,
24
+ "encoder_layerdrop": 0.0,
25
+ "encoder_layers": 12,
26
+ "eos_token_id": 2,
27
+ "forced_eos_token_id": 2,
28
+ "gradient_checkpointing": false,
29
+ "id2label": {
30
+ "0": "LABEL_0",
31
+ "1": "LABEL_1",
32
+ "2": "LABEL_2"
33
+ },
34
+ "init_std": 0.02,
35
+ "is_encoder_decoder": true,
36
+ "label2id": {
37
+ "LABEL_0": 0,
38
+ "LABEL_1": 1,
39
+ "LABEL_2": 2
40
+ },
41
+ "max_position_embeddings": 1024,
42
+ "model_type": "bart",
43
+ "no_repeat_ngram_size": 3,
44
+ "normalize_before": false,
45
+ "num_beams": 4,
46
+ "num_hidden_layers": 12,
47
+ "pad_token_id": 1,
48
+ "scale_embedding": false,
49
+ "task_specific_params": {
50
+ "summarization": {
51
+ "length_penalty": 1.0,
52
+ "max_length": 128,
53
+ "min_length": 12,
54
+ "num_beams": 4
55
+ },
56
+ "summarization_cnn": {
57
+ "length_penalty": 2.0,
58
+ "max_length": 142,
59
+ "min_length": 56,
60
+ "num_beams": 4
61
+ },
62
+ "summarization_xsum": {
63
+ "length_penalty": 1.0,
64
+ "max_length": 62,
65
+ "min_length": 11,
66
+ "num_beams": 6
67
+ }
68
+ },
69
+ "torch_dtype": "float32",
70
+ "transformers_version": "4.12.5",
71
+ "use_cache": true,
72
+ "vocab_size": 50266
73
+ }
eval/metric.first.answer.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"dev": {"Bleu_1": 0.48406879120947427, "Bleu_2": 0.3327119786166825, "Bleu_3": 0.2511075798531371, "Bleu_4": 0.19671634242245814, "METEOR": 0.2266473382703044, "ROUGE_L": 0.46579175726471117}, "test": {"Bleu_1": 0.4622713223035384, "Bleu_2": 0.3107179177959037, "Bleu_3": 0.2304108919392574, "Bleu_4": 0.17667100940990094, "METEOR": 0.2188383319130062, "ROUGE_L": 0.44993420214100505}}
eval/metric.first.sentence.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"dev": {"Bleu_1": 0.5480164167410585, "Bleu_2": 0.3898850280789222, "Bleu_3": 0.3005491914855951, "Bleu_4": 0.2390563667332963, "METEOR": 0.2509887691266911, "ROUGE_L": 0.5063550621156588}, "test": {"Bleu_1": 0.5513066870053248, "Bleu_2": 0.3900073838681523, "Bleu_3": 0.29727850050142923, "Bleu_4": 0.23214955671664797, "METEOR": 0.24936501492606217, "ROUGE_L": 0.500863380109846}}
eval/samples.dev.hyp.txt ADDED
The diff for this file is too large to render. See raw diff
 
eval/samples.test.hyp.txt ADDED
The diff for this file is too large to render. See raw diff
 
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b620eec44bdefffa3e578d9db68ed9bbdcd9803b5e73defd73b38b61a5f5953
3
+ size 1625561409
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "sep_token": "</s>", "pad_token": "<pad>", "cls_token": "<s>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": false}, "additional_special_tokens": ["<hl>"]}
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "<unk>", "bos_token": "<s>", "eos_token": "</s>", "add_prefix_space": false, "errors": "replace", "sep_token": "</s>", "cls_token": "<s>", "pad_token": "<pad>", "mask_token": "<mask>", "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "lmqg_output/optimized_no_answer/bart_large_squad/model_blrpij/epoch_2", "tokenizer_class": "BartTokenizer"}
vocab.json ADDED
The diff for this file is too large to render. See raw diff