asuvarna31 commited on
Commit
125a643
1 Parent(s): 9deb17d

added model

Browse files
config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "roberta-large",
3
+ "architectures": [
4
+ "RobertaForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
+ "finetuning_task": "comp",
11
+ "hidden_act": "gelu",
12
+ "hidden_dropout_prob": 0.1,
13
+ "hidden_size": 1024,
14
+ "id2label": {
15
+ "0": "LABEL_0",
16
+ "1": "LABEL_1",
17
+ "2": "LABEL_2"
18
+ },
19
+ "initializer_range": 0.02,
20
+ "intermediate_size": 4096,
21
+ "label2id": {
22
+ "LABEL_0": 0,
23
+ "LABEL_1": 1,
24
+ "LABEL_2": 2
25
+ },
26
+ "layer_norm_eps": 1e-05,
27
+ "max_position_embeddings": 514,
28
+ "model_type": "roberta",
29
+ "num_attention_heads": 16,
30
+ "num_hidden_layers": 24,
31
+ "pad_token_id": 1,
32
+ "position_embedding_type": "absolute",
33
+ "problem_type": "single_label_classification",
34
+ "torch_dtype": "float32",
35
+ "transformers_version": "4.34.1",
36
+ "type_vocab_size": 1,
37
+ "use_cache": true,
38
+ "vocab_size": 50265
39
+ }
eval_results ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ acc_best = 0.745398773006135
2
+ num_best = 326
3
+ correct_best = 243
4
+ prec_0_best = 0.8007662835249042
5
+ rec_0_best = 0.9372197309417041
6
+ f1_0_best = 0.8636363636363638
7
+ prec_1_best = 0.34782608695652173
8
+ rec_1_best = 0.26666666666666666
9
+ f1_1_best = 0.30188679245283023
10
+ prec_2_best = 0.6190476190476191
11
+ rec_2_best = 0.3561643835616438
12
+ f1_2_best = 0.45217391304347826
13
+ macro_f1_best = 0.5392323563775574
14
+ acc_MaxLen256 = 0.7147239263803681
15
+ num_MaxLen256 = 326
16
+ correct_MaxLen256 = 233
17
+ prec_0_MaxLen256 = 0.7672727272727272
18
+ rec_0_MaxLen256 = 0.9461883408071748
19
+ f1_0_MaxLen256 = 0.8473895582329316
20
+ prec_1_MaxLen256 = 0.20833333333333334
21
+ rec_1_MaxLen256 = 0.16666666666666666
22
+ f1_1_MaxLen256 = 0.1851851851851852
23
+ prec_2_MaxLen256 = 0.6296296296296297
24
+ rec_2_MaxLen256 = 0.2328767123287671
25
+ f1_2_MaxLen256 = 0.34
26
+ macro_f1_MaxLen256 = 0.4575249144727056
27
+ Best checkpoint is ./outputs/oversample/comp/roberta-large-LR1e-5-epoch10-MaxLen256/checkpoint-best, best accuracy is 0.745398773006135
eval_test_results ADDED
@@ -0,0 +1,187 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ======= Predict using the model from checkpoint-20:
3
+
4
+ ======= Predict using the model from checkpoint-20:
5
+ {"acc": 0.6698292220113852, "num": 527, "correct": 353, "prec_0": 0.6698292220113852, "rec_0": 1.0, "f1_0": 0.8022727272727272, "prec_1": 0.0, "rec_1": 0.0, "f1_1": 0, "prec_2": 0.0, "rec_2": 0.0, "f1_2": 0, "macro_f1": 0.2674242424242424}
6
+ ======= Predict using the model from checkpoint-40:
7
+ {"acc": 0.6698292220113852, "num": 527, "correct": 353, "prec_0": 0.6698292220113852, "rec_0": 1.0, "f1_0": 0.8022727272727272, "prec_1": 0.0, "rec_1": 0.0, "f1_1": 0, "prec_2": 0.0, "rec_2": 0.0, "f1_2": 0, "macro_f1": 0.2674242424242424}
8
+ ======= Predict using the model from checkpoint-60:
9
+ {"acc": 0.6698292220113852, "num": 527, "correct": 353, "prec_0": 0.6698292220113852, "rec_0": 1.0, "f1_0": 0.8022727272727272, "prec_1": 0.0, "rec_1": 0.0, "f1_1": 0, "prec_2": 0.0, "rec_2": 0.0, "f1_2": 0, "macro_f1": 0.2674242424242424}
10
+ ======= Predict using the model from checkpoint-80:
11
+ {"acc": 0.6660341555977229, "num": 527, "correct": 351, "prec_0": 0.6737044145873321, "rec_0": 0.9943342776203966, "f1_0": 0.8032036613272312, "prec_1": 0.0, "rec_1": 0.0, "f1_1": 0, "prec_2": 0.0, "rec_2": 0.0, "f1_2": 0, "macro_f1": 0.26773455377574373}
12
+ ======= Predict using the model from checkpoint-100:
13
+ {"acc": 0.6736242884250474, "num": 527, "correct": 355, "prec_0": 0.7063655030800822, "rec_0": 0.9745042492917847, "f1_0": 0.819047619047619, "prec_1": 0.0967741935483871, "rec_1": 0.08108108108108109, "f1_1": 0.08823529411764705, "prec_2": 0.8888888888888888, "rec_2": 0.058394160583941604, "f1_2": 0.1095890410958904, "macro_f1": 0.33895731808705215}
14
+ ======= Predict using the model from checkpoint-120:
15
+ {"acc": 0.7229601518026565, "num": 527, "correct": 381, "prec_0": 0.7797029702970297, "rec_0": 0.8923512747875354, "f1_0": 0.8322324966974901, "prec_1": 0.16666666666666666, "rec_1": 0.08108108108108109, "f1_1": 0.1090909090909091, "prec_2": 0.6, "rec_2": 0.45985401459854014, "f1_2": 0.5206611570247933, "macro_f1": 0.48732818760439756}
16
+ ======= Predict using the model from checkpoint-140:
17
+ {"acc": 0.7115749525616698, "num": 527, "correct": 375, "prec_0": 0.7686746987951807, "rec_0": 0.9036827195467422, "f1_0": 0.8307291666666667, "prec_1": 0.17142857142857143, "rec_1": 0.16216216216216217, "f1_1": 0.16666666666666669, "prec_2": 0.6493506493506493, "rec_2": 0.36496350364963503, "f1_2": 0.4672897196261682, "macro_f1": 0.48822851765316727}
18
+ ======= Predict using the model from checkpoint-160:
19
+ {"acc": 0.6793168880455408, "num": 527, "correct": 358, "prec_0": 0.8378378378378378, "rec_0": 0.7903682719546742, "f1_0": 0.8134110787172012, "prec_1": 0.18333333333333332, "rec_1": 0.2972972972972973, "f1_1": 0.2268041237113402, "prec_2": 0.5074626865671642, "rec_2": 0.49635036496350365, "f1_2": 0.5018450184501846, "macro_f1": 0.514020073626242}
20
+ ======= Predict using the model from checkpoint-180:
21
+ {"acc": 0.7020872865275142, "num": 527, "correct": 370, "prec_0": 0.7181628392484343, "rec_0": 0.9745042492917847, "f1_0": 0.826923076923077, "prec_1": 0.125, "rec_1": 0.05405405405405406, "f1_1": 0.07547169811320754, "prec_2": 0.75, "rec_2": 0.17518248175182483, "f1_2": 0.28402366863905326, "macro_f1": 0.3954728145584459}
22
+ ======= Predict using the model from checkpoint-200:
23
+ {"acc": 0.7115749525616698, "num": 527, "correct": 375, "prec_0": 0.7783251231527094, "rec_0": 0.8951841359773371, "f1_0": 0.8326745718050066, "prec_1": 0.19230769230769232, "rec_1": 0.13513513513513514, "f1_1": 0.15873015873015872, "prec_2": 0.5684210526315789, "rec_2": 0.39416058394160586, "f1_2": 0.4655172413793104, "macro_f1": 0.48564065730482525}
24
+ ======= Predict using the model from checkpoint-220:
25
+ {"acc": 0.713472485768501, "num": 527, "correct": 376, "prec_0": 0.7511520737327189, "rec_0": 0.9235127478753541, "f1_0": 0.8284625158831004, "prec_1": 0.1724137931034483, "rec_1": 0.13513513513513514, "f1_1": 0.15151515151515152, "prec_2": 0.703125, "rec_2": 0.3284671532846715, "f1_2": 0.44776119402985076, "macro_f1": 0.47591295380936754}
26
+ ======= Predict using the model from checkpoint-240:
27
+ {"acc": 0.6963946869070209, "num": 527, "correct": 367, "prec_0": 0.7663551401869159, "rec_0": 0.9291784702549575, "f1_0": 0.8399487836107554, "prec_1": 0.1346153846153846, "rec_1": 0.1891891891891892, "f1_1": 0.15730337078651685, "prec_2": 0.6808510638297872, "rec_2": 0.23357664233576642, "f1_2": 0.34782608695652173, "macro_f1": 0.44835941378459804}
28
+ ======= Predict using the model from checkpoint-260:
29
+ {"acc": 0.7058823529411765, "num": 527, "correct": 372, "prec_0": 0.748314606741573, "rec_0": 0.943342776203966, "f1_0": 0.8345864661654135, "prec_1": 0.16666666666666666, "rec_1": 0.16216216216216217, "f1_1": 0.16438356164383564, "prec_2": 0.717391304347826, "rec_2": 0.24087591240875914, "f1_2": 0.36065573770491804, "macro_f1": 0.4532085885047224}
30
+ ======= Predict using the model from checkpoint-280:
31
+ {"acc": 0.7172675521821632, "num": 527, "correct": 378, "prec_0": 0.7580645161290323, "rec_0": 0.9320113314447592, "f1_0": 0.8360864040660737, "prec_1": 0.20588235294117646, "rec_1": 0.1891891891891892, "f1_1": 0.19718309859154928, "prec_2": 0.711864406779661, "rec_2": 0.30656934306569344, "f1_2": 0.4285714285714286, "macro_f1": 0.4872803104096839}
32
+ ======= Predict using the model from checkpoint-300:
33
+ {"acc": 0.7305502846299811, "num": 527, "correct": 385, "prec_0": 0.74235807860262, "rec_0": 0.9631728045325779, "f1_0": 0.8384710234278668, "prec_1": 0.3333333333333333, "rec_1": 0.13513513513513514, "f1_1": 0.1923076923076923, "prec_2": 0.7407407407407407, "rec_2": 0.291970802919708, "f1_2": 0.418848167539267, "macro_f1": 0.48320896109160866}
34
+ ======= Predict using the model from checkpoint-320:
35
+ {"acc": 0.7267552182163188, "num": 527, "correct": 383, "prec_0": 0.7422907488986784, "rec_0": 0.9546742209631728, "f1_0": 0.8351920693928129, "prec_1": 0.3333333333333333, "rec_1": 0.10810810810810811, "f1_1": 0.163265306122449, "prec_2": 0.6885245901639344, "rec_2": 0.30656934306569344, "f1_2": 0.42424242424242425, "macro_f1": 0.47423326658589543}
36
+ ======= Predict using the model from checkpoint-340:
37
+ {"acc": 0.7248576850094877, "num": 527, "correct": 382, "prec_0": 0.7630331753554502, "rec_0": 0.9121813031161473, "f1_0": 0.8309677419354838, "prec_1": 0.35, "rec_1": 0.1891891891891892, "f1_1": 0.24561403508771934, "prec_2": 0.6235294117647059, "rec_2": 0.38686131386861317, "f1_2": 0.47747747747747743, "macro_f1": 0.5180197515002268}
38
+ ======= Predict using the model from checkpoint-360:
39
+ {"acc": 0.7210626185958254, "num": 527, "correct": 380, "prec_0": 0.76, "rec_0": 0.9150141643059491, "f1_0": 0.8303341902313625, "prec_1": 0.34782608695652173, "rec_1": 0.21621621621621623, "f1_1": 0.26666666666666666, "prec_2": 0.620253164556962, "rec_2": 0.35766423357664234, "f1_2": 0.45370370370370366, "macro_f1": 0.5169015202005777}
40
+ ======= Predict using the model from checkpoint-380:
41
+ {"acc": 0.7096774193548387, "num": 527, "correct": 374, "prec_0": 0.7454954954954955, "rec_0": 0.9376770538243626, "f1_0": 0.8306148055207028, "prec_1": 0.21212121212121213, "rec_1": 0.1891891891891892, "f1_1": 0.20000000000000004, "prec_2": 0.72, "rec_2": 0.26277372262773724, "f1_2": 0.3850267379679144, "macro_f1": 0.4718805144962057}
42
+ ======= Predict using the model from checkpoint-400:
43
+ {"acc": 0.7115749525616698, "num": 527, "correct": 375, "prec_0": 0.7327586206896551, "rec_0": 0.9631728045325779, "f1_0": 0.832313341493268, "prec_1": 0.24, "rec_1": 0.16216216216216217, "f1_1": 0.1935483870967742, "prec_2": 0.7631578947368421, "rec_2": 0.2116788321167883, "f1_2": 0.33142857142857146, "macro_f1": 0.45243010000620454}
44
+ ======= Predict using the model from checkpoint-420:
45
+ {"acc": 0.7191650853889943, "num": 527, "correct": 379, "prec_0": 0.7482993197278912, "rec_0": 0.9348441926345609, "f1_0": 0.8312342569269521, "prec_1": 0.2692307692307692, "rec_1": 0.1891891891891892, "f1_1": 0.22222222222222224, "prec_2": 0.7, "rec_2": 0.30656934306569344, "f1_2": 0.4263959390862944, "macro_f1": 0.49328413941182286}
46
+ ======= Predict using the model from checkpoint-440:
47
+ {"acc": 0.7172675521821632, "num": 527, "correct": 378, "prec_0": 0.7615740740740741, "rec_0": 0.9320113314447592, "f1_0": 0.8382165605095542, "prec_1": 0.23076923076923078, "rec_1": 0.24324324324324326, "f1_1": 0.23684210526315788, "prec_2": 0.7142857142857143, "rec_2": 0.291970802919708, "f1_2": 0.41450777202072536, "macro_f1": 0.4965221459311458}
48
+ ======= Predict using the model from checkpoint-460:
49
+ {"acc": 0.7210626185958254, "num": 527, "correct": 380, "prec_0": 0.738562091503268, "rec_0": 0.9603399433427762, "f1_0": 0.834975369458128, "prec_1": 0.2608695652173913, "rec_1": 0.16216216216216217, "f1_1": 0.2, "prec_2": 0.7777777777777778, "rec_2": 0.25547445255474455, "f1_2": 0.38461538461538464, "macro_f1": 0.4731969180245042}
50
+ ======= Predict using the model from checkpoint-480:
51
+ {"acc": 0.7286527514231499, "num": 527, "correct": 384, "prec_0": 0.7641509433962265, "rec_0": 0.9178470254957507, "f1_0": 0.8339768339768339, "prec_1": 0.3, "rec_1": 0.24324324324324326, "f1_1": 0.26865671641791045, "prec_2": 0.6986301369863014, "rec_2": 0.3722627737226277, "f1_2": 0.4857142857142857, "macro_f1": 0.5294492787030101}
52
+ ======= Predict using the model from checkpoint-500:
53
+ {"acc": 0.7191650853889943, "num": 527, "correct": 379, "prec_0": 0.7444933920704846, "rec_0": 0.9575070821529745, "f1_0": 0.8376703841387856, "prec_1": 0.2413793103448276, "rec_1": 0.1891891891891892, "f1_1": 0.21212121212121213, "prec_2": 0.7727272727272727, "rec_2": 0.24817518248175183, "f1_2": 0.37569060773480667, "macro_f1": 0.47516073466493475}
54
+ ======= Predict using the model from checkpoint-520:
55
+ {"acc": 0.7248576850094877, "num": 527, "correct": 382, "prec_0": 0.7466666666666667, "rec_0": 0.9518413597733711, "f1_0": 0.8368617683686178, "prec_1": 0.2692307692307692, "rec_1": 0.1891891891891892, "f1_1": 0.22222222222222224, "prec_2": 0.7647058823529411, "rec_2": 0.2846715328467153, "f1_2": 0.4148936170212766, "macro_f1": 0.4913258692040389}
56
+ ======= Predict using the model from checkpoint-20:
57
+ {"acc": 0.6679316888045541, "num": 527, "correct": 352, "prec_0": 0.6717557251908397, "rec_0": 0.9971671388101983, "f1_0": 0.8027366020524516, "prec_1": 0.0, "rec_1": 0.0, "f1_1": 0, "prec_2": 0.0, "rec_2": 0.0, "f1_2": 0, "macro_f1": 0.2675788673508172}
58
+ ======= Predict using the model from checkpoint-40:
59
+ {"acc": 0.6698292220113852, "num": 527, "correct": 353, "prec_0": 0.6698292220113852, "rec_0": 1.0, "f1_0": 0.8022727272727272, "prec_1": 0.0, "rec_1": 0.0, "f1_1": 0, "prec_2": 0.0, "rec_2": 0.0, "f1_2": 0, "macro_f1": 0.2674242424242424}
60
+ ======= Predict using the model from checkpoint-60:
61
+ {"acc": 0.6698292220113852, "num": 527, "correct": 353, "prec_0": 0.6698292220113852, "rec_0": 1.0, "f1_0": 0.8022727272727272, "prec_1": 0.0, "rec_1": 0.0, "f1_1": 0, "prec_2": 0.0, "rec_2": 0.0, "f1_2": 0, "macro_f1": 0.2674242424242424}
62
+ ======= Predict using the model from checkpoint-80:
63
+ {"acc": 0.6698292220113852, "num": 527, "correct": 353, "prec_0": 0.6698292220113852, "rec_0": 1.0, "f1_0": 0.8022727272727272, "prec_1": 0.0, "rec_1": 0.0, "f1_1": 0, "prec_2": 0.0, "rec_2": 0.0, "f1_2": 0, "macro_f1": 0.2674242424242424}
64
+ ======= Predict using the model from checkpoint-100:
65
+ {"acc": 0.6698292220113852, "num": 527, "correct": 353, "prec_0": 0.6698292220113852, "rec_0": 1.0, "f1_0": 0.8022727272727272, "prec_1": 0.0, "rec_1": 0.0, "f1_1": 0, "prec_2": 0.0, "rec_2": 0.0, "f1_2": 0, "macro_f1": 0.2674242424242424}
66
+ ======= Predict using the model from checkpoint-120:
67
+ {"acc": 0.4364326375711575, "num": 527, "correct": 230, "prec_0": 0.7239057239057239, "rec_0": 0.6090651558073654, "f1_0": 0.6615384615384615, "prec_1": 0.06521739130434782, "rec_1": 0.40540540540540543, "f1_1": 0.11235955056179775, "prec_2": 0.0, "rec_2": 0.0, "f1_2": 0, "macro_f1": 0.25796600403341974}
68
+ ======= Predict using the model from checkpoint-140:
69
+ {"acc": 0.6527514231499051, "num": 527, "correct": 344, "prec_0": 0.6854838709677419, "rec_0": 0.9631728045325779, "f1_0": 0.800942285041225, "prec_1": 0.0, "rec_1": 0.0, "f1_1": 0, "prec_2": 0.6666666666666666, "rec_2": 0.029197080291970802, "f1_2": 0.055944055944055944, "macro_f1": 0.285628780328427}
70
+ ======= Predict using the model from checkpoint-160:
71
+ {"acc": 0.6850094876660342, "num": 527, "correct": 361, "prec_0": 0.6862745098039216, "rec_0": 0.9915014164305949, "f1_0": 0.8111239860950175, "prec_1": 0.0, "rec_1": 0.0, "f1_1": 0, "prec_2": 0.8461538461538461, "rec_2": 0.08029197080291971, "f1_2": 0.14666666666666667, "macro_f1": 0.3192635509205614}
72
+ ======= Predict using the model from checkpoint-180:
73
+ {"acc": 0.6679316888045541, "num": 527, "correct": 352, "prec_0": 0.7777777777777778, "rec_0": 0.7932011331444759, "f1_0": 0.7854137447405329, "prec_1": 0.125, "rec_1": 0.08108108108108109, "f1_1": 0.09836065573770492, "prec_2": 0.4825174825174825, "rec_2": 0.5036496350364964, "f1_2": 0.4928571428571429, "macro_f1": 0.4588771811117936}
74
+ ======= Predict using the model from checkpoint-200:
75
+ {"acc": 0.6584440227703985, "num": 527, "correct": 347, "prec_0": 0.7264770240700219, "rec_0": 0.9405099150141643, "f1_0": 0.8197530864197531, "prec_1": 0.11864406779661017, "rec_1": 0.1891891891891892, "f1_1": 0.14583333333333334, "prec_2": 0.7272727272727273, "rec_2": 0.058394160583941604, "f1_2": 0.1081081081081081, "macro_f1": 0.35789817595373147}
76
+ ======= Predict using the model from checkpoint-220:
77
+ {"acc": 0.6204933586337761, "num": 527, "correct": 327, "prec_0": 0.7487437185929648, "rec_0": 0.8441926345609065, "f1_0": 0.7936085219707057, "prec_1": 0.07526881720430108, "rec_1": 0.1891891891891892, "f1_1": 0.1076923076923077, "prec_2": 0.6111111111111112, "rec_2": 0.16058394160583941, "f1_2": 0.2543352601156069, "macro_f1": 0.3852120299262068}
78
+ ======= Predict using the model from checkpoint-240:
79
+ {"acc": 0.7020872865275142, "num": 527, "correct": 370, "prec_0": 0.7580246913580246, "rec_0": 0.8696883852691218, "f1_0": 0.8100263852242745, "prec_1": 0.3333333333333333, "rec_1": 0.10810810810810811, "f1_1": 0.163265306122449, "prec_2": 0.5363636363636364, "rec_2": 0.4306569343065693, "f1_2": 0.47773279352226716, "macro_f1": 0.4836748282896635}
80
+ ======= Predict using the model from checkpoint-260:
81
+ {"acc": 0.681214421252372, "num": 527, "correct": 359, "prec_0": 0.75, "rec_0": 0.9263456090651558, "f1_0": 0.8288973384030418, "prec_1": 0.09803921568627451, "rec_1": 0.13513513513513514, "f1_1": 0.11363636363636365, "prec_2": 0.675, "rec_2": 0.19708029197080293, "f1_2": 0.3050847457627119, "macro_f1": 0.4158728159340391}
82
+ ======= Predict using the model from checkpoint-280:
83
+ {"acc": 0.6679316888045541, "num": 527, "correct": 352, "prec_0": 0.7028688524590164, "rec_0": 0.9716713881019831, "f1_0": 0.8156956004756243, "prec_1": 0.034482758620689655, "rec_1": 0.02702702702702703, "f1_1": 0.030303030303030304, "prec_2": 0.8, "rec_2": 0.058394160583941604, "f1_2": 0.10884353741496598, "macro_f1": 0.31828072273120683}
84
+ ======= Predict using the model from checkpoint-300:
85
+ {"acc": 0.6679316888045541, "num": 527, "correct": 352, "prec_0": 0.7557603686635944, "rec_0": 0.9291784702549575, "f1_0": 0.8335451080050825, "prec_1": 0.08955223880597014, "rec_1": 0.16216216216216217, "f1_1": 0.11538461538461539, "prec_2": 0.6923076923076923, "rec_2": 0.13138686131386862, "f1_2": 0.22085889570552145, "macro_f1": 0.38992953969840644}
86
+ ======= Predict using the model from checkpoint-320:
87
+ {"acc": 0.7001897533206831, "num": 527, "correct": 369, "prec_0": 0.7160751565762005, "rec_0": 0.9716713881019831, "f1_0": 0.8245192307692307, "prec_1": 0.1875, "rec_1": 0.08108108108108109, "f1_1": 0.11320754716981132, "prec_2": 0.71875, "rec_2": 0.1678832116788321, "f1_2": 0.272189349112426, "macro_f1": 0.4033053756838227}
88
+ ======= Predict using the model from checkpoint-340:
89
+ {"acc": 0.6888045540796964, "num": 527, "correct": 363, "prec_0": 0.7482678983833718, "rec_0": 0.9178470254957507, "f1_0": 0.8244274809160306, "prec_1": 0.1388888888888889, "rec_1": 0.13513513513513514, "f1_1": 0.136986301369863, "prec_2": 0.5862068965517241, "rec_2": 0.24817518248175183, "f1_2": 0.3487179487179487, "macro_f1": 0.43671057700128074}
90
+ ======= Predict using the model from checkpoint-360:
91
+ {"acc": 0.6869070208728653, "num": 527, "correct": 362, "prec_0": 0.7136842105263158, "rec_0": 0.9603399433427762, "f1_0": 0.818840579710145, "prec_1": 0.125, "rec_1": 0.08108108108108109, "f1_1": 0.09836065573770492, "prec_2": 0.7142857142857143, "rec_2": 0.145985401459854, "f1_2": 0.24242424242424243, "macro_f1": 0.3865418259573641}
92
+ ======= Predict using the model from checkpoint-380:
93
+ {"acc": 0.6888045540796964, "num": 527, "correct": 363, "prec_0": 0.7063655030800822, "rec_0": 0.9745042492917847, "f1_0": 0.819047619047619, "prec_1": 0.058823529411764705, "rec_1": 0.02702702702702703, "f1_1": 0.03703703703703704, "prec_2": 0.782608695652174, "rec_2": 0.13138686131386862, "f1_2": 0.225, "macro_f1": 0.3603615520282187}
94
+ ======= Predict using the model from checkpoint-400:
95
+ {"acc": 0.6850094876660342, "num": 527, "correct": 361, "prec_0": 0.6983805668016194, "rec_0": 0.9773371104815864, "f1_0": 0.8146399055489965, "prec_1": 0.07692307692307693, "rec_1": 0.02702702702702703, "f1_1": 0.04, "prec_2": 0.75, "rec_2": 0.10948905109489052, "f1_2": 0.19108280254777074, "macro_f1": 0.34857423603225574}
96
+ ======= Predict using the model from checkpoint-420:
97
+ {"acc": 0.6850094876660342, "num": 527, "correct": 361, "prec_0": 0.7161016949152542, "rec_0": 0.9575070821529745, "f1_0": 0.8193939393939395, "prec_1": 0.07142857142857142, "rec_1": 0.05405405405405406, "f1_1": 0.06153846153846154, "prec_2": 0.7777777777777778, "rec_2": 0.15328467153284672, "f1_2": 0.25609756097560976, "macro_f1": 0.3790099873026702}
98
+ ======= Predict using the model from checkpoint-440:
99
+ {"acc": 0.6944971537001897, "num": 527, "correct": 366, "prec_0": 0.7089397089397089, "rec_0": 0.9660056657223796, "f1_0": 0.8177458033573142, "prec_1": 0.14285714285714285, "rec_1": 0.05405405405405406, "f1_1": 0.0784313725490196, "prec_2": 0.71875, "rec_2": 0.1678832116788321, "f1_2": 0.272189349112426, "macro_f1": 0.3894555083395866}
100
+ ======= Predict using the model from checkpoint-460:
101
+ {"acc": 0.683111954459203, "num": 527, "correct": 360, "prec_0": 0.7304347826086957, "rec_0": 0.9518413597733711, "f1_0": 0.8265682656826568, "prec_1": 0.1111111111111111, "rec_1": 0.10810810810810811, "f1_1": 0.1095890410958904, "prec_2": 0.6451612903225806, "rec_2": 0.145985401459854, "f1_2": 0.23809523809523808, "macro_f1": 0.3914175149579284}
102
+ ======= Predict using the model from checkpoint-480:
103
+ {"acc": 0.6774193548387096, "num": 527, "correct": 357, "prec_0": 0.7127882599580713, "rec_0": 0.9631728045325779, "f1_0": 0.8192771084337349, "prec_1": 0.0967741935483871, "rec_1": 0.08108108108108109, "f1_1": 0.08823529411764705, "prec_2": 0.7368421052631579, "rec_2": 0.10218978102189781, "f1_2": 0.1794871794871795, "macro_f1": 0.3623331940128538}
104
+ ======= Predict using the model from checkpoint-500:
105
+ {"acc": 0.6963946869070209, "num": 527, "correct": 367, "prec_0": 0.7296703296703296, "rec_0": 0.9405099150141643, "f1_0": 0.8217821782178218, "prec_1": 0.125, "rec_1": 0.08108108108108109, "f1_1": 0.09836065573770492, "prec_2": 0.6666666666666666, "rec_2": 0.23357664233576642, "f1_2": 0.3459459459459459, "macro_f1": 0.4220295933004909}
106
+ ======= Predict using the model from checkpoint-520:
107
+ {"acc": 0.6888045540796964, "num": 527, "correct": 363, "prec_0": 0.7268817204301076, "rec_0": 0.9575070821529745, "f1_0": 0.8264058679706602, "prec_1": 0.0967741935483871, "rec_1": 0.08108108108108109, "f1_1": 0.08823529411764705, "prec_2": 0.7096774193548387, "rec_2": 0.16058394160583941, "f1_2": 0.2619047619047619, "macro_f1": 0.3921819746643564}
108
+ ======= Predict using the model from checkpoint-540:
109
+ {"acc": 0.6793168880455408, "num": 527, "correct": 358, "prec_0": 0.7432432432432432, "rec_0": 0.9348441926345609, "f1_0": 0.8281053952321203, "prec_1": 0.10204081632653061, "rec_1": 0.13513513513513514, "f1_1": 0.11627906976744186, "prec_2": 0.6764705882352942, "rec_2": 0.1678832116788321, "f1_2": 0.26900584795321636, "macro_f1": 0.4044634376509262}
110
+ ======= Predict using the model from checkpoint-560:
111
+ {"acc": 0.6963946869070209, "num": 527, "correct": 367, "prec_0": 0.7362637362637363, "rec_0": 0.9490084985835694, "f1_0": 0.8292079207920793, "prec_1": 0.12903225806451613, "rec_1": 0.10810810810810811, "f1_1": 0.11764705882352941, "prec_2": 0.6829268292682927, "rec_2": 0.20437956204379562, "f1_2": 0.31460674157303375, "macro_f1": 0.42048724039621416}
112
+ ======= Predict using the model from checkpoint-580:
113
+ {"acc": 0.7020872865275142, "num": 527, "correct": 370, "prec_0": 0.7545871559633027, "rec_0": 0.9320113314447592, "f1_0": 0.8339670468948036, "prec_1": 0.17142857142857143, "rec_1": 0.16216216216216217, "f1_1": 0.16666666666666669, "prec_2": 0.625, "rec_2": 0.25547445255474455, "f1_2": 0.3626943005181347, "macro_f1": 0.4544426713598683}
114
+ ======= Predict using the model from checkpoint-600:
115
+ {"acc": 0.6641366223908919, "num": 527, "correct": 350, "prec_0": 0.7565632458233891, "rec_0": 0.8980169971671388, "f1_0": 0.8212435233160622, "prec_1": 0.11764705882352941, "rec_1": 0.21621621621621623, "f1_1": 0.15238095238095237, "prec_2": 0.625, "rec_2": 0.18248175182481752, "f1_2": 0.2824858757062147, "macro_f1": 0.4187034504677431}
116
+ ======= Predict using the model from checkpoint-620:
117
+ {"acc": 0.6907020872865275, "num": 527, "correct": 364, "prec_0": 0.7416481069042317, "rec_0": 0.943342776203966, "f1_0": 0.830423940149626, "prec_1": 0.13953488372093023, "rec_1": 0.16216216216216217, "f1_1": 0.15, "prec_2": 0.7142857142857143, "rec_2": 0.18248175182481752, "f1_2": 0.29069767441860467, "macro_f1": 0.4237072048560769}
118
+ ======= Predict using the model from checkpoint-640:
119
+ {"acc": 0.7001897533206831, "num": 527, "correct": 369, "prec_0": 0.7257383966244726, "rec_0": 0.9745042492917847, "f1_0": 0.8319226118500604, "prec_1": 0.19230769230769232, "rec_1": 0.13513513513513514, "f1_1": 0.15873015873015872, "prec_2": 0.7407407407407407, "rec_2": 0.145985401459854, "f1_2": 0.2439024390243902, "macro_f1": 0.41151840320153643}
120
+ ======= Predict using the model from checkpoint-660:
121
+ {"acc": 0.6944971537001897, "num": 527, "correct": 366, "prec_0": 0.7259100642398287, "rec_0": 0.9603399433427762, "f1_0": 0.8268292682926829, "prec_1": 0.16129032258064516, "rec_1": 0.13513513513513514, "f1_1": 0.14705882352941174, "prec_2": 0.7586206896551724, "rec_2": 0.16058394160583941, "f1_2": 0.2650602409638554, "macro_f1": 0.4129827775953167}
122
+ ======= Predict using the model from checkpoint-680:
123
+ {"acc": 0.6774193548387096, "num": 527, "correct": 357, "prec_0": 0.7308533916849015, "rec_0": 0.9461756373937678, "f1_0": 0.8246913580246913, "prec_1": 0.11627906976744186, "rec_1": 0.13513513513513514, "f1_1": 0.125, "prec_2": 0.6666666666666666, "rec_2": 0.13138686131386862, "f1_2": 0.21951219512195125, "macro_f1": 0.3897345177155475}
124
+ ======= Predict using the model from checkpoint-700:
125
+ {"acc": 0.6888045540796964, "num": 527, "correct": 363, "prec_0": 0.7304347826086957, "rec_0": 0.9518413597733711, "f1_0": 0.8265682656826568, "prec_1": 0.14705882352941177, "rec_1": 0.13513513513513514, "f1_1": 0.14084507042253522, "prec_2": 0.6666666666666666, "rec_2": 0.16058394160583941, "f1_2": 0.2588235294117647, "macro_f1": 0.4087456218389856}
126
+ ======= Predict using the model from checkpoint-720:
127
+ {"acc": 0.6907020872865275, "num": 527, "correct": 364, "prec_0": 0.721868365180467, "rec_0": 0.9631728045325779, "f1_0": 0.8252427184466019, "prec_1": 0.16129032258064516, "rec_1": 0.13513513513513514, "f1_1": 0.14705882352941174, "prec_2": 0.76, "rec_2": 0.1386861313868613, "f1_2": 0.23456790123456792, "macro_f1": 0.4022898144035272}
128
+ ======= Predict using the model from checkpoint-740:
129
+ {"acc": 0.6793168880455408, "num": 527, "correct": 358, "prec_0": 0.7403189066059226, "rec_0": 0.9206798866855525, "f1_0": 0.8207070707070707, "prec_1": 0.10638297872340426, "rec_1": 0.13513513513513514, "f1_1": 0.11904761904761907, "prec_2": 0.6829268292682927, "rec_2": 0.20437956204379562, "f1_2": 0.31460674157303375, "macro_f1": 0.4181204771092412}
130
+ ======= Predict using the model from checkpoint-760:
131
+ {"acc": 0.6944971537001897, "num": 527, "correct": 366, "prec_0": 0.7371937639198218, "rec_0": 0.9376770538243626, "f1_0": 0.8254364089775561, "prec_1": 0.14285714285714285, "rec_1": 0.13513513513513514, "f1_1": 0.1388888888888889, "prec_2": 0.6976744186046512, "rec_2": 0.21897810218978103, "f1_2": 0.33333333333333337, "macro_f1": 0.4325528770665928}
132
+ ======= Predict using the model from checkpoint-780:
133
+ {"acc": 0.6907020872865275, "num": 527, "correct": 364, "prec_0": 0.7262931034482759, "rec_0": 0.9546742209631728, "f1_0": 0.824969400244798, "prec_1": 0.14285714285714285, "rec_1": 0.13513513513513514, "f1_1": 0.1388888888888889, "prec_2": 0.7857142857142857, "rec_2": 0.16058394160583941, "f1_2": 0.2666666666666666, "macro_f1": 0.4101749852667845}
134
+ ======= Predict using the model from checkpoint-800:
135
+ {"acc": 0.6717267552182163, "num": 527, "correct": 354, "prec_0": 0.747072599531616, "rec_0": 0.9036827195467422, "f1_0": 0.8179487179487179, "prec_1": 0.09259259259259259, "rec_1": 0.13513513513513514, "f1_1": 0.1098901098901099, "prec_2": 0.6521739130434783, "rec_2": 0.21897810218978103, "f1_2": 0.3278688524590164, "macro_f1": 0.41856922676594815}
136
+ ======= Predict using the model from checkpoint-820:
137
+ {"acc": 0.6944971537001897, "num": 527, "correct": 366, "prec_0": 0.7298474945533769, "rec_0": 0.9490084985835694, "f1_0": 0.8251231527093597, "prec_1": 0.15151515151515152, "rec_1": 0.13513513513513514, "f1_1": 0.14285714285714285, "prec_2": 0.7428571428571429, "rec_2": 0.1897810218978102, "f1_2": 0.3023255813953488, "macro_f1": 0.42343529232061705}
138
+ ======= Predict using the model from checkpoint-840:
139
+ {"acc": 0.6888045540796964, "num": 527, "correct": 363, "prec_0": 0.7366071428571429, "rec_0": 0.9348441926345609, "f1_0": 0.8239700374531835, "prec_1": 0.12195121951219512, "rec_1": 0.13513513513513514, "f1_1": 0.1282051282051282, "prec_2": 0.7368421052631579, "rec_2": 0.20437956204379562, "f1_2": 0.32, "macro_f1": 0.4240583885527706}
140
+ ======= Predict using the model from checkpoint-860:
141
+ {"acc": 0.683111954459203, "num": 527, "correct": 360, "prec_0": 0.7546728971962616, "rec_0": 0.9150141643059491, "f1_0": 0.827144686299616, "prec_1": 0.10714285714285714, "rec_1": 0.16216216216216217, "f1_1": 0.12903225806451613, "prec_2": 0.7209302325581395, "rec_2": 0.22627737226277372, "f1_2": 0.34444444444444444, "macro_f1": 0.43354046293619214}
142
+ ======= Predict using the model from checkpoint-880:
143
+ {"acc": 0.7001897533206831, "num": 527, "correct": 369, "prec_0": 0.7336244541484717, "rec_0": 0.9518413597733711, "f1_0": 0.8286066584463626, "prec_1": 0.16129032258064516, "rec_1": 0.13513513513513514, "f1_1": 0.14705882352941174, "prec_2": 0.7368421052631579, "rec_2": 0.20437956204379562, "f1_2": 0.32, "macro_f1": 0.4318884939919248}
144
+ ======= Predict using the model from checkpoint-900:
145
+ {"acc": 0.6793168880455408, "num": 527, "correct": 358, "prec_0": 0.7658536585365854, "rec_0": 0.8895184135977338, "f1_0": 0.8230668414154653, "prec_1": 0.12903225806451613, "rec_1": 0.21621621621621623, "f1_1": 0.16161616161616163, "prec_2": 0.6545454545454545, "rec_2": 0.26277372262773724, "f1_2": 0.375, "macro_f1": 0.45322766767720896}
146
+ ======= Predict using the model from checkpoint-920:
147
+ {"acc": 0.6944971537001897, "num": 527, "correct": 366, "prec_0": 0.7389380530973452, "rec_0": 0.9461756373937678, "f1_0": 0.8298136645962733, "prec_1": 0.13157894736842105, "rec_1": 0.13513513513513514, "f1_1": 0.13333333333333333, "prec_2": 0.7297297297297297, "rec_2": 0.19708029197080293, "f1_2": 0.3103448275862069, "macro_f1": 0.4244972751719378}
148
+ ======= Predict using the model from checkpoint-940:
149
+ {"acc": 0.6907020872865275, "num": 527, "correct": 364, "prec_0": 0.7454545454545455, "rec_0": 0.9291784702549575, "f1_0": 0.8272383354350568, "prec_1": 0.11363636363636363, "rec_1": 0.13513513513513514, "f1_1": 0.12345679012345678, "prec_2": 0.7209302325581395, "rec_2": 0.22627737226277372, "f1_2": 0.34444444444444444, "macro_f1": 0.431713190000986}
150
+ ======= Predict using the model from checkpoint-960:
151
+ {"acc": 0.6907020872865275, "num": 527, "correct": 364, "prec_0": 0.7477272727272727, "rec_0": 0.9320113314447592, "f1_0": 0.8297604035308953, "prec_1": 0.10869565217391304, "rec_1": 0.13513513513513514, "f1_1": 0.12048192771084337, "prec_2": 0.7317073170731707, "rec_2": 0.21897810218978103, "f1_2": 0.33707865168539325, "macro_f1": 0.429106994309044}
152
+ ======= Predict using the model from checkpoint-980:
153
+ {"acc": 0.6925996204933587, "num": 527, "correct": 365, "prec_0": 0.746031746031746, "rec_0": 0.9320113314447592, "f1_0": 0.8287153652392947, "prec_1": 0.11627906976744186, "rec_1": 0.13513513513513514, "f1_1": 0.125, "prec_2": 0.7209302325581395, "rec_2": 0.22627737226277372, "f1_2": 0.34444444444444444, "macro_f1": 0.4327199365612464}
154
+ ======= Predict using the model from checkpoint-1000:
155
+ {"acc": 0.6907020872865275, "num": 527, "correct": 364, "prec_0": 0.738359201773836, "rec_0": 0.943342776203966, "f1_0": 0.8283582089552238, "prec_1": 0.125, "rec_1": 0.13513513513513514, "f1_1": 0.12987012987012989, "prec_2": 0.7222222222222222, "rec_2": 0.1897810218978102, "f1_2": 0.3005780346820809, "macro_f1": 0.4196021245024782}
156
+ ======= Predict using the model from checkpoint-1020:
157
+ {"acc": 0.6869070208728653, "num": 527, "correct": 362, "prec_0": 0.7404921700223713, "rec_0": 0.9376770538243626, "f1_0": 0.8274999999999999, "prec_1": 0.11363636363636363, "rec_1": 0.13513513513513514, "f1_1": 0.12345679012345678, "prec_2": 0.7222222222222222, "rec_2": 0.1897810218978102, "f1_2": 0.3005780346820809, "macro_f1": 0.41717827493517917}
158
+ ======= Predict using the model from checkpoint-1040:
159
+ {"acc": 0.6944971537001897, "num": 527, "correct": 366, "prec_0": 0.74, "rec_0": 0.943342776203966, "f1_0": 0.8293897882938979, "prec_1": 0.1282051282051282, "rec_1": 0.13513513513513514, "f1_1": 0.13157894736842105, "prec_2": 0.7368421052631579, "rec_2": 0.20437956204379562, "f1_2": 0.32, "macro_f1": 0.4269895785541063}
160
+ ======= Predict using the model from checkpoint-1060:
161
+ {"acc": 0.6944971537001897, "num": 527, "correct": 366, "prec_0": 0.74, "rec_0": 0.943342776203966, "f1_0": 0.8293897882938979, "prec_1": 0.1282051282051282, "rec_1": 0.13513513513513514, "f1_1": 0.13157894736842105, "prec_2": 0.7368421052631579, "rec_2": 0.20437956204379562, "f1_2": 0.32, "macro_f1": 0.4269895785541063}
162
+ ======= Predict using the model from checkpoint-20:
163
+ {"acc": 0.6698292220113852, "num": 527, "correct": 353, "prec_0": 0.6698292220113852, "rec_0": 1.0, "f1_0": 0.8022727272727272, "prec_1": 0.0, "rec_1": 0.0, "f1_1": 0, "prec_2": 0.0, "rec_2": 0.0, "f1_2": 0, "macro_f1": 0.2674242424242424}
164
+ ======= Predict using the model from checkpoint-40:
165
+ {"acc": 0.6698292220113852, "num": 527, "correct": 353, "prec_0": 0.6698292220113852, "rec_0": 1.0, "f1_0": 0.8022727272727272, "prec_1": 0.0, "rec_1": 0.0, "f1_1": 0, "prec_2": 0.0, "rec_2": 0.0, "f1_2": 0, "macro_f1": 0.2674242424242424}
166
+ ======= Predict using the model from checkpoint-60:
167
+ {"acc": 0.6717267552182163, "num": 527, "correct": 354, "prec_0": 0.6711026615969582, "rec_0": 1.0, "f1_0": 0.8031854379977247, "prec_1": 1.0, "rec_1": 0.02702702702702703, "f1_1": 0.052631578947368425, "prec_2": 0.0, "rec_2": 0.0, "f1_2": 0, "macro_f1": 0.28527233898169774}
168
+ ======= Predict using the model from checkpoint-80:
169
+ {"acc": 0.6774193548387096, "num": 527, "correct": 357, "prec_0": 0.6749521988527725, "rec_0": 1.0, "f1_0": 0.8059360730593608, "prec_1": 1.0, "rec_1": 0.02702702702702703, "f1_1": 0.052631578947368425, "prec_2": 1.0, "rec_2": 0.021897810218978103, "f1_2": 0.04285714285714286, "macro_f1": 0.3004749316212907}
170
+ ======= Predict using the model from checkpoint-100:
171
+ {"acc": 0.7001897533206831, "num": 527, "correct": 369, "prec_0": 0.7251585623678647, "rec_0": 0.9716713881019831, "f1_0": 0.8305084745762712, "prec_1": 0.14814814814814814, "rec_1": 0.10810810810810811, "f1_1": 0.125, "prec_2": 0.8148148148148148, "rec_2": 0.16058394160583941, "f1_2": 0.26829268292682923, "macro_f1": 0.40793371916770016}
172
+ ======= Predict using the model from checkpoint-120:
173
+ {"acc": 0.7286527514231499, "num": 527, "correct": 384, "prec_0": 0.7386609071274298, "rec_0": 0.9688385269121813, "f1_0": 0.838235294117647, "prec_1": 0.3, "rec_1": 0.16216216216216217, "f1_1": 0.2105263157894737, "prec_2": 0.8181818181818182, "rec_2": 0.26277372262773724, "f1_2": 0.39779005524861877, "macro_f1": 0.48218388838524645}
174
+ ======= Predict using the model from checkpoint-140:
175
+ {"acc": 0.7058823529411765, "num": 527, "correct": 372, "prec_0": 0.7540229885057471, "rec_0": 0.9291784702549575, "f1_0": 0.83248730964467, "prec_1": 0.23255813953488372, "rec_1": 0.2702702702702703, "f1_1": 0.25, "prec_2": 0.6938775510204082, "rec_2": 0.24817518248175183, "f1_2": 0.3655913978494624, "macro_f1": 0.48269290249804414}
176
+ ======= Predict using the model from checkpoint-160:
177
+ {"acc": 0.715370018975332, "num": 527, "correct": 377, "prec_0": 0.7321814254859611, "rec_0": 0.9603399433427762, "f1_0": 0.8308823529411764, "prec_1": 0.1875, "rec_1": 0.08108108108108109, "f1_1": 0.11320754716981132, "prec_2": 0.7291666666666666, "rec_2": 0.25547445255474455, "f1_2": 0.3783783783783784, "macro_f1": 0.44082275949645533}
178
+ ======= Predict using the model from checkpoint-180:
179
+ {"acc": 0.713472485768501, "num": 527, "correct": 376, "prec_0": 0.7307692307692307, "rec_0": 0.9688385269121813, "f1_0": 0.8331303288672349, "prec_1": 0.21739130434782608, "rec_1": 0.13513513513513514, "f1_1": 0.16666666666666666, "prec_2": 0.8055555555555556, "rec_2": 0.2116788321167883, "f1_2": 0.33526011560693636, "macro_f1": 0.44501903704694595}
180
+ ======= Predict using the model from checkpoint-200:
181
+ {"acc": 0.7115749525616698, "num": 527, "correct": 375, "prec_0": 0.7511210762331838, "rec_0": 0.9490084985835694, "f1_0": 0.8385481852315393, "prec_1": 0.24390243902439024, "rec_1": 0.2702702702702703, "f1_1": 0.2564102564102564, "prec_2": 0.75, "rec_2": 0.21897810218978103, "f1_2": 0.33898305084745767, "macro_f1": 0.47798049749641774}
182
+ ======= Predict using the model from checkpoint-220:
183
+ {"acc": 0.6963946869070209, "num": 527, "correct": 367, "prec_0": 0.7721822541966427, "rec_0": 0.9121813031161473, "f1_0": 0.8363636363636363, "prec_1": 0.21311475409836064, "rec_1": 0.35135135135135137, "f1_1": 0.2653061224489796, "prec_2": 0.6530612244897959, "rec_2": 0.23357664233576642, "f1_2": 0.3440860215053763, "macro_f1": 0.4819185934393307}
184
+ ======= Predict using the model from checkpoint-240:
185
+ {"acc": 0.715370018975332, "num": 527, "correct": 377, "prec_0": 0.7461368653421634, "rec_0": 0.9575070821529745, "f1_0": 0.8387096774193549, "prec_1": 0.28205128205128205, "rec_1": 0.2972972972972973, "f1_1": 0.2894736842105264, "prec_2": 0.8, "rec_2": 0.20437956204379562, "f1_2": 0.3255813953488372, "macro_f1": 0.4845882523262395}
186
+ ======= Predict using the model from checkpoint-260:
187
+ {"acc": 0.715370018975332, "num": 527, "correct": 377, "prec_0": 0.7483443708609272, "rec_0": 0.9603399433427762, "f1_0": 0.8411910669975187, "prec_1": 0.275, "rec_1": 0.2972972972972973, "f1_1": 0.28571428571428575, "prec_2": 0.7941176470588235, "rec_2": 0.19708029197080293, "f1_2": 0.31578947368421056, "macro_f1": 0.48089827546533837}
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5db803ee9652ac234917018ae4447f811f56649cc9401ab9b74b71fe5c5c7a63
3
+ size 1421583983
special_tokens_map.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "cls_token": {
10
+ "content": "<s>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "eos_token": {
17
+ "content": "</s>",
18
+ "lstrip": false,
19
+ "normalized": true,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "mask_token": {
24
+ "content": "<mask>",
25
+ "lstrip": true,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "pad_token": {
31
+ "content": "<pad>",
32
+ "lstrip": false,
33
+ "normalized": true,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ },
37
+ "sep_token": {
38
+ "content": "</s>",
39
+ "lstrip": false,
40
+ "normalized": true,
41
+ "rstrip": false,
42
+ "single_word": false
43
+ },
44
+ "unk_token": {
45
+ "content": "<unk>",
46
+ "lstrip": false,
47
+ "normalized": true,
48
+ "rstrip": false,
49
+ "single_word": false
50
+ }
51
+ }
test.tsv ADDED
@@ -0,0 +1,527 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2
2
+ 1
3
+ 1
4
+ 1
5
+ 1
6
+ 1
7
+ 3
8
+ 1
9
+ 1
10
+ 1
11
+ 1
12
+ 3
13
+ 1
14
+ 1
15
+ 1
16
+ 2
17
+ 1
18
+ 1
19
+ 1
20
+ 1
21
+ 1
22
+ 2
23
+ 2
24
+ 1
25
+ 1
26
+ 1
27
+ 1
28
+ 1
29
+ 1
30
+ 1
31
+ 1
32
+ 1
33
+ 1
34
+ 1
35
+ 2
36
+ 1
37
+ 1
38
+ 2
39
+ 1
40
+ 1
41
+ 1
42
+ 1
43
+ 1
44
+ 1
45
+ 1
46
+ 1
47
+ 1
48
+ 3
49
+ 1
50
+ 2
51
+ 1
52
+ 1
53
+ 1
54
+ 1
55
+ 1
56
+ 1
57
+ 1
58
+ 1
59
+ 1
60
+ 1
61
+ 1
62
+ 1
63
+ 1
64
+ 1
65
+ 1
66
+ 1
67
+ 1
68
+ 1
69
+ 1
70
+ 1
71
+ 1
72
+ 1
73
+ 1
74
+ 1
75
+ 1
76
+ 1
77
+ 1
78
+ 1
79
+ 1
80
+ 1
81
+ 1
82
+ 1
83
+ 1
84
+ 1
85
+ 1
86
+ 1
87
+ 1
88
+ 1
89
+ 1
90
+ 1
91
+ 1
92
+ 1
93
+ 1
94
+ 1
95
+ 1
96
+ 1
97
+ 1
98
+ 1
99
+ 1
100
+ 1
101
+ 1
102
+ 1
103
+ 1
104
+ 1
105
+ 1
106
+ 1
107
+ 1
108
+ 1
109
+ 1
110
+ 1
111
+ 1
112
+ 1
113
+ 1
114
+ 2
115
+ 1
116
+ 1
117
+ 1
118
+ 1
119
+ 1
120
+ 1
121
+ 1
122
+ 1
123
+ 1
124
+ 1
125
+ 1
126
+ 1
127
+ 1
128
+ 1
129
+ 1
130
+ 1
131
+ 1
132
+ 1
133
+ 3
134
+ 1
135
+ 1
136
+ 1
137
+ 1
138
+ 1
139
+ 1
140
+ 1
141
+ 1
142
+ 1
143
+ 1
144
+ 1
145
+ 1
146
+ 1
147
+ 1
148
+ 1
149
+ 1
150
+ 1
151
+ 1
152
+ 1
153
+ 1
154
+ 1
155
+ 1
156
+ 1
157
+ 1
158
+ 1
159
+ 1
160
+ 1
161
+ 1
162
+ 1
163
+ 1
164
+ 1
165
+ 1
166
+ 3
167
+ 1
168
+ 1
169
+ 1
170
+ 1
171
+ 3
172
+ 1
173
+ 1
174
+ 1
175
+ 1
176
+ 1
177
+ 1
178
+ 1
179
+ 1
180
+ 1
181
+ 1
182
+ 2
183
+ 1
184
+ 1
185
+ 1
186
+ 1
187
+ 1
188
+ 1
189
+ 1
190
+ 1
191
+ 1
192
+ 2
193
+ 1
194
+ 1
195
+ 1
196
+ 1
197
+ 2
198
+ 2
199
+ 1
200
+ 1
201
+ 1
202
+ 1
203
+ 1
204
+ 1
205
+ 1
206
+ 3
207
+ 1
208
+ 1
209
+ 1
210
+ 1
211
+ 1
212
+ 1
213
+ 1
214
+ 1
215
+ 1
216
+ 1
217
+ 1
218
+ 1
219
+ 1
220
+ 1
221
+ 1
222
+ 3
223
+ 3
224
+ 1
225
+ 1
226
+ 1
227
+ 1
228
+ 1
229
+ 1
230
+ 2
231
+ 3
232
+ 2
233
+ 1
234
+ 1
235
+ 1
236
+ 2
237
+ 1
238
+ 1
239
+ 1
240
+ 1
241
+ 1
242
+ 1
243
+ 1
244
+ 1
245
+ 1
246
+ 1
247
+ 1
248
+ 1
249
+ 1
250
+ 1
251
+ 1
252
+ 1
253
+ 1
254
+ 1
255
+ 1
256
+ 1
257
+ 2
258
+ 1
259
+ 1
260
+ 1
261
+ 1
262
+ 1
263
+ 1
264
+ 1
265
+ 3
266
+ 1
267
+ 1
268
+ 1
269
+ 1
270
+ 1
271
+ 1
272
+ 1
273
+ 1
274
+ 1
275
+ 1
276
+ 1
277
+ 1
278
+ 1
279
+ 1
280
+ 1
281
+ 1
282
+ 1
283
+ 1
284
+ 1
285
+ 2
286
+ 1
287
+ 1
288
+ 1
289
+ 1
290
+ 3
291
+ 1
292
+ 1
293
+ 1
294
+ 1
295
+ 2
296
+ 1
297
+ 1
298
+ 1
299
+ 1
300
+ 1
301
+ 1
302
+ 1
303
+ 1
304
+ 1
305
+ 1
306
+ 1
307
+ 1
308
+ 1
309
+ 1
310
+ 1
311
+ 1
312
+ 1
313
+ 1
314
+ 1
315
+ 1
316
+ 1
317
+ 1
318
+ 1
319
+ 1
320
+ 1
321
+ 1
322
+ 1
323
+ 1
324
+ 1
325
+ 1
326
+ 1
327
+ 1
328
+ 3
329
+ 1
330
+ 1
331
+ 1
332
+ 1
333
+ 1
334
+ 1
335
+ 1
336
+ 1
337
+ 1
338
+ 1
339
+ 1
340
+ 1
341
+ 1
342
+ 1
343
+ 1
344
+ 1
345
+ 1
346
+ 1
347
+ 1
348
+ 1
349
+ 1
350
+ 1
351
+ 1
352
+ 1
353
+ 1
354
+ 1
355
+ 3
356
+ 2
357
+ 2
358
+ 2
359
+ 1
360
+ 1
361
+ 1
362
+ 1
363
+ 1
364
+ 1
365
+ 1
366
+ 2
367
+ 1
368
+ 2
369
+ 1
370
+ 2
371
+ 2
372
+ 1
373
+ 2
374
+ 3
375
+ 3
376
+ 3
377
+ 1
378
+ 1
379
+ 1
380
+ 1
381
+ 2
382
+ 1
383
+ 2
384
+ 1
385
+ 1
386
+ 1
387
+ 1
388
+ 2
389
+ 2
390
+ 2
391
+ 1
392
+ 1
393
+ 3
394
+ 2
395
+ 1
396
+ 1
397
+ 3
398
+ 1
399
+ 2
400
+ 1
401
+ 1
402
+ 1
403
+ 1
404
+ 1
405
+ 3
406
+ 2
407
+ 3
408
+ 3
409
+ 1
410
+ 1
411
+ 1
412
+ 1
413
+ 1
414
+ 1
415
+ 1
416
+ 1
417
+ 1
418
+ 1
419
+ 3
420
+ 2
421
+ 1
422
+ 1
423
+ 1
424
+ 2
425
+ 2
426
+ 1
427
+ 1
428
+ 1
429
+ 1
430
+ 1
431
+ 1
432
+ 1
433
+ 2
434
+ 2
435
+ 1
436
+ 2
437
+ 1
438
+ 2
439
+ 1
440
+ 1
441
+ 1
442
+ 2
443
+ 1
444
+ 1
445
+ 3
446
+ 2
447
+ 2
448
+ 3
449
+ 1
450
+ 1
451
+ 1
452
+ 1
453
+ 1
454
+ 1
455
+ 1
456
+ 1
457
+ 1
458
+ 1
459
+ 3
460
+ 1
461
+ 1
462
+ 1
463
+ 2
464
+ 1
465
+ 1
466
+ 1
467
+ 1
468
+ 3
469
+ 3
470
+ 2
471
+ 3
472
+ 1
473
+ 2
474
+ 1
475
+ 3
476
+ 3
477
+ 1
478
+ 3
479
+ 1
480
+ 1
481
+ 1
482
+ 1
483
+ 2
484
+ 2
485
+ 2
486
+ 1
487
+ 3
488
+ 3
489
+ 3
490
+ 3
491
+ 2
492
+ 1
493
+ 2
494
+ 1
495
+ 3
496
+ 3
497
+ 2
498
+ 2
499
+ 2
500
+ 1
501
+ 1
502
+ 3
503
+ 2
504
+ 1
505
+ 3
506
+ 2
507
+ 3
508
+ 1
509
+ 3
510
+ 2
511
+ 1
512
+ 1
513
+ 3
514
+ 3
515
+ 3
516
+ 2
517
+ 3
518
+ 3
519
+ 2
520
+ 3
521
+ 3
522
+ 2
523
+ 1
524
+ 1
525
+ 1
526
+ 1
527
+ 1
test_results.txt ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ ======= Predict using the model from ./outputs/oversample/comp/roberta-large-LR1e-5-epoch10-MaxLen256/checkpoint-best for test:
2
+ {"acc": 0.7096774193548387, "num": 527, "correct": 374, "prec_0": 0.7454954954954955, "rec_0": 0.9376770538243626, "f1_0": 0.8306148055207028, "prec_1": 0.21212121212121213, "rec_1": 0.1891891891891892, "f1_1": 0.20000000000000004, "prec_2": 0.72, "rec_2": 0.26277372262773724, "f1_2": 0.3850267379679144, "macro_f1": 0.4718805144962057}======= Predict using the model from ./outputs/oversample/comp/roberta-large-LR1e-5-epoch10-MaxLen256/checkpoint-best for test:
3
+ {"acc": 0.6907020872865275, "num": 527, "correct": 364, "prec_0": 0.7454545454545455, "rec_0": 0.9291784702549575, "f1_0": 0.8272383354350568, "prec_1": 0.11363636363636363, "rec_1": 0.13513513513513514, "f1_1": 0.12345679012345678, "prec_2": 0.7209302325581395, "rec_2": 0.22627737226277372, "f1_2": 0.34444444444444444, "macro_f1": 0.431713190000986}======= Predict using the model from ./outputs/oversample/comp/roberta-large-LR1e-5-epoch10-MaxLen256/checkpoint-best for test:
4
+ {"acc": 0.6963946869070209, "num": 527, "correct": 367, "prec_0": 0.7721822541966427, "rec_0": 0.9121813031161473, "f1_0": 0.8363636363636363, "prec_1": 0.21311475409836064, "rec_1": 0.35135135135135137, "f1_1": 0.2653061224489796, "prec_2": 0.6530612244897959, "rec_2": 0.23357664233576642, "f1_2": 0.3440860215053763, "macro_f1": 0.4819185934393307}
tokenizer_config.json ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "<s>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "<pad>",
14
+ "lstrip": false,
15
+ "normalized": true,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "2": {
21
+ "content": "</s>",
22
+ "lstrip": false,
23
+ "normalized": true,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "3": {
29
+ "content": "<unk>",
30
+ "lstrip": false,
31
+ "normalized": true,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": true
35
+ },
36
+ "50264": {
37
+ "content": "<mask>",
38
+ "lstrip": true,
39
+ "normalized": false,
40
+ "rstrip": false,
41
+ "single_word": false,
42
+ "special": true
43
+ }
44
+ },
45
+ "bos_token": "<s>",
46
+ "clean_up_tokenization_spaces": true,
47
+ "cls_token": "<s>",
48
+ "do_lower_case": false,
49
+ "eos_token": "</s>",
50
+ "errors": "replace",
51
+ "mask_token": "<mask>",
52
+ "model_max_length": 512,
53
+ "pad_token": "<pad>",
54
+ "sep_token": "</s>",
55
+ "tokenizer_class": "RobertaTokenizer",
56
+ "unk_token": "<unk>"
57
+ }
train ADDED
The diff for this file is too large to render. See raw diff
 
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98d224803526afb65f31b0849d3c380c5b8b23fbbd5cdae8ced87d6ff5d88193
3
+ size 1711
vocab.json ADDED
The diff for this file is too large to render. See raw diff