mengzhouxia commited on
Commit
afb7f55
1 Parent(s): 8a488df

first commit

Browse files
Files changed (4) hide show
  1. config.json +131 -0
  2. pytorch_model.bin +3 -0
  3. tokenizer_config.json +1 -0
  4. vocab.txt +0 -0
config.json ADDED
@@ -0,0 +1,131 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "NewBertForSequenceClassification"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "decompose_qk": false,
7
+ "decompose_vo": false,
8
+ "do_distill": true,
9
+ "do_emb_distill": false,
10
+ "do_layer_distill": true,
11
+ "do_mha_distill": false,
12
+ "do_mha_layer_distill": false,
13
+ "finetuning_task": "mnli",
14
+ "gradient_checkpointing": false,
15
+ "hidden_act": "gelu",
16
+ "hidden_dropout_prob": 0.1,
17
+ "hidden_size": 768,
18
+ "id2label": {
19
+ "0": "LABEL_0",
20
+ "1": "LABEL_1",
21
+ "2": "LABEL_2"
22
+ },
23
+ "initializer_range": 0.02,
24
+ "intermediate_size": 3072,
25
+ "label2id": {
26
+ "LABEL_0": 0,
27
+ "LABEL_1": 1,
28
+ "LABEL_2": 2
29
+ },
30
+ "layer_norm_eps": 1e-12,
31
+ "max_position_embeddings": 512,
32
+ "model_type": "bert",
33
+ "num_attention_heads": 12,
34
+ "num_hidden_layers": 12,
35
+ "output_attentions": true,
36
+ "output_hidden_states": true,
37
+ "pad_token_id": 0,
38
+ "pruned_heads": {
39
+ "0": [
40
+ 1,
41
+ 2,
42
+ 4,
43
+ 5,
44
+ 7,
45
+ 11
46
+ ],
47
+ "1": [
48
+ 0,
49
+ 2,
50
+ 3,
51
+ 5,
52
+ 6,
53
+ 7
54
+ ],
55
+ "2": [
56
+ 3,
57
+ 4,
58
+ 7,
59
+ 8,
60
+ 11
61
+ ],
62
+ "3": [
63
+ 2,
64
+ 4,
65
+ 6,
66
+ 7
67
+ ],
68
+ "4": [
69
+ 8,
70
+ 1
71
+ ],
72
+ "5": [
73
+ 1,
74
+ 2,
75
+ 11
76
+ ],
77
+ "6": [
78
+ 3,
79
+ 7,
80
+ 9,
81
+ 10,
82
+ 11
83
+ ],
84
+ "7": [
85
+ 2,
86
+ 4,
87
+ 6,
88
+ 7,
89
+ 11
90
+ ],
91
+ "8": [
92
+ 0,
93
+ 2,
94
+ 3,
95
+ 7,
96
+ 8,
97
+ 10
98
+ ],
99
+ "9": [
100
+ 1,
101
+ 3,
102
+ 4,
103
+ 5,
104
+ 7,
105
+ 9
106
+ ],
107
+ "10": [
108
+ 0,
109
+ 1,
110
+ 2,
111
+ 5,
112
+ 6,
113
+ 9
114
+ ],
115
+ "11": [
116
+ 0,
117
+ 3,
118
+ 5,
119
+ 6,
120
+ 7,
121
+ 8,
122
+ 10,
123
+ 11
124
+ ]
125
+ },
126
+ "qk_denominator": "ori",
127
+ "sephidden_pruned": false,
128
+ "transform_embedding": false,
129
+ "type_vocab_size": 2,
130
+ "vocab_size": 30522
131
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3914bd84cf7eebf7d24a63f91003a2b552409aa0a633ed3e4824af0a9e81fee9
3
+ size 233981203
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"do_lower_case": true, "model_max_length": 512}
vocab.txt ADDED
The diff for this file is too large to render. See raw diff