avichr commited on
Commit
7ba5ff3
1 Parent(s): e90c7a6
Files changed (5) hide show
  1. config.json +41 -0
  2. log_history.json +173 -0
  3. pytorch_model.bin +3 -0
  4. training_args.bin +3 -0
  5. vocab.txt +0 -0
config.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "BertForTokenClassification"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "gradient_checkpointing": false,
7
+ "hidden_act": "gelu",
8
+ "hidden_dropout_prob": 0.1,
9
+ "hidden_size": 768,
10
+ "id2label": {
11
+ "0": "B_TIME",
12
+ "1": "B_PERS",
13
+ "2": "B_MONEY",
14
+ "3": "B_ORG",
15
+ "4": "B_PERCENT",
16
+ "5": "O",
17
+ "6": "B_LOC",
18
+ "7": "B_DATE"
19
+ },
20
+ "initializer_range": 0.02,
21
+ "intermediate_size": 3072,
22
+ "label2id": {
23
+ "B_DATE": 7,
24
+ "B_LOC": 6,
25
+ "B_MONEY": 2,
26
+ "B_ORG": 3,
27
+ "B_PERCENT": 4,
28
+ "B_PERS": 1,
29
+ "B_TIME": 0,
30
+ "O": 5
31
+ },
32
+ "layer_norm_eps": 1e-12,
33
+ "max_position_embeddings": 512,
34
+ "model_type": "bert",
35
+ "num_attention_heads": 12,
36
+ "num_hidden_layers": 12,
37
+ "pad_token_id": 0,
38
+ "total_flos": 6997313242916978688,
39
+ "type_vocab_size": 2,
40
+ "vocab_size": 30522
41
+ }
log_history.json ADDED
@@ -0,0 +1,173 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "loss": 2.1632102966308593,
4
+ "learning_rate": 1.0000000000000002e-06,
5
+ "epoch": 0.2127659574468085,
6
+ "total_flos": 4181675520000,
7
+ "step": 10
8
+ },
9
+ {
10
+ "loss": 1.9698543548583984,
11
+ "learning_rate": 2.0000000000000003e-06,
12
+ "epoch": 0.425531914893617,
13
+ "total_flos": 8363351040000,
14
+ "step": 20
15
+ },
16
+ {
17
+ "loss": 1.5943157196044921,
18
+ "learning_rate": 3e-06,
19
+ "epoch": 0.6382978723404256,
20
+ "total_flos": 12545026560000,
21
+ "step": 30
22
+ },
23
+ {
24
+ "loss": 1.1324241638183594,
25
+ "learning_rate": 4.000000000000001e-06,
26
+ "epoch": 0.851063829787234,
27
+ "total_flos": 16726702080000,
28
+ "step": 40
29
+ },
30
+ {
31
+ "loss": 0.7799140930175781,
32
+ "learning_rate": 5e-06,
33
+ "epoch": 1.0638297872340425,
34
+ "total_flos": 20620887408000,
35
+ "step": 50
36
+ },
37
+ {
38
+ "loss": 0.5984169006347656,
39
+ "learning_rate": 6e-06,
40
+ "epoch": 1.2765957446808511,
41
+ "total_flos": 24802562928000,
42
+ "step": 60
43
+ },
44
+ {
45
+ "loss": 0.4769012451171875,
46
+ "learning_rate": 7.000000000000001e-06,
47
+ "epoch": 1.4893617021276595,
48
+ "total_flos": 28984238448000,
49
+ "step": 70
50
+ },
51
+ {
52
+ "loss": 0.3644599914550781,
53
+ "learning_rate": 8.000000000000001e-06,
54
+ "epoch": 1.702127659574468,
55
+ "total_flos": 33165913968000,
56
+ "step": 80
57
+ },
58
+ {
59
+ "loss": 0.29639739990234376,
60
+ "learning_rate": 9e-06,
61
+ "epoch": 1.9148936170212765,
62
+ "total_flos": 37347589488000,
63
+ "step": 90
64
+ },
65
+ {
66
+ "loss": 0.26825637817382814,
67
+ "learning_rate": 1e-05,
68
+ "epoch": 2.127659574468085,
69
+ "total_flos": 41241774816000,
70
+ "step": 100
71
+ },
72
+ {
73
+ "loss": 0.19923171997070313,
74
+ "learning_rate": 1.1000000000000001e-05,
75
+ "epoch": 2.3404255319148937,
76
+ "total_flos": 45423450336000,
77
+ "step": 110
78
+ },
79
+ {
80
+ "loss": 0.19310073852539061,
81
+ "learning_rate": 1.2e-05,
82
+ "epoch": 2.5531914893617023,
83
+ "total_flos": 49605125856000,
84
+ "step": 120
85
+ },
86
+ {
87
+ "loss": 0.15804977416992189,
88
+ "learning_rate": 1.3000000000000001e-05,
89
+ "epoch": 2.7659574468085104,
90
+ "total_flos": 53786801376000,
91
+ "step": 130
92
+ },
93
+ {
94
+ "loss": 0.17403030395507812,
95
+ "learning_rate": 1.4000000000000001e-05,
96
+ "epoch": 2.978723404255319,
97
+ "total_flos": 57968476896000,
98
+ "step": 140
99
+ },
100
+ {
101
+ "loss": 0.11231231689453125,
102
+ "learning_rate": 1.5e-05,
103
+ "epoch": 3.1914893617021276,
104
+ "total_flos": 61862662224000,
105
+ "step": 150
106
+ },
107
+ {
108
+ "loss": 0.0972991943359375,
109
+ "learning_rate": 1.6000000000000003e-05,
110
+ "epoch": 3.404255319148936,
111
+ "total_flos": 66044337744000,
112
+ "step": 160
113
+ },
114
+ {
115
+ "loss": 0.09285507202148438,
116
+ "learning_rate": 1.7000000000000003e-05,
117
+ "epoch": 3.617021276595745,
118
+ "total_flos": 70226013264000,
119
+ "step": 170
120
+ },
121
+ {
122
+ "loss": 0.10490646362304687,
123
+ "learning_rate": 1.8e-05,
124
+ "epoch": 3.829787234042553,
125
+ "total_flos": 74407688784000,
126
+ "step": 180
127
+ },
128
+ {
129
+ "loss": 0.08623809814453125,
130
+ "learning_rate": 1.9e-05,
131
+ "epoch": 4.042553191489362,
132
+ "total_flos": 78301874112000,
133
+ "step": 190
134
+ },
135
+ {
136
+ "loss": 0.05904541015625,
137
+ "learning_rate": 2e-05,
138
+ "epoch": 4.25531914893617,
139
+ "total_flos": 82483549632000,
140
+ "step": 200
141
+ },
142
+ {
143
+ "loss": 0.052008056640625,
144
+ "learning_rate": 2.1e-05,
145
+ "epoch": 4.468085106382979,
146
+ "total_flos": 86665225152000,
147
+ "step": 210
148
+ },
149
+ {
150
+ "loss": 0.06222076416015625,
151
+ "learning_rate": 2.2000000000000003e-05,
152
+ "epoch": 4.680851063829787,
153
+ "total_flos": 90846900672000,
154
+ "step": 220
155
+ },
156
+ {
157
+ "loss": 0.05680923461914063,
158
+ "learning_rate": 2.3000000000000003e-05,
159
+ "epoch": 4.8936170212765955,
160
+ "total_flos": 95028576192000,
161
+ "step": 230
162
+ },
163
+ {
164
+ "eval_loss": 0.0899211324525602,
165
+ "eval_accuracy_score": 0.974121282348397,
166
+ "eval_precision": 0.9776031434184675,
167
+ "eval_recall": 0.9749216300940439,
168
+ "eval_f1": 0.9762605454188739,
169
+ "epoch": 5.0,
170
+ "total_flos": 96831923760000,
171
+ "step": 235
172
+ }
173
+ ]
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d385d4e8a6fd0c04dff5ab0ef9b091d66ad239a1d17ef4c6fee85d860bb6efee
3
+ size 435679343
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d683997fff536a8cb258bffa83dc61ab5f5b2167c2bb0fbd8408703c1e902c4
3
+ size 1775
vocab.txt ADDED
The diff for this file is too large to render. See raw diff