anniew666 commited on
Commit
25c46ac
1 Parent(s): 8d5bf10

End of training

Browse files
Files changed (8) hide show
  1. README.md +28 -28
  2. all_results.json +40 -0
  3. cf.png +0 -0
  4. cf.txt +7 -7
  5. class_report.txt +8 -8
  6. eval_results.json +35 -0
  7. train_results.json +8 -0
  8. trainer_state.json +1287 -0
README.md CHANGED
@@ -19,34 +19,34 @@ should probably proofread and complete it, then remove this comment. -->
19
 
20
  This model is a fine-tuned version of [roberta-large](https://huggingface.co/roberta-large) on an unknown dataset.
21
  It achieves the following results on the evaluation set:
22
- - Loss: 0.7423
23
- - Accuracy: 0.7389
24
- - Prec: 0.7379
25
- - Recall: 0.7389
26
- - F1: 0.7381
27
- - B Acc: 0.6013
28
- - Micro F1: 0.7389
29
- - Prec Joy: 0.7415
30
- - Recall Joy: 0.7893
31
- - F1 Joy: 0.7646
32
- - Prec Anger: 0.6020
33
- - Recall Anger: 0.5728
34
- - F1 Anger: 0.5871
35
- - Prec Disgust: 0.4013
36
- - Recall Disgust: 0.3896
37
- - F1 Disgust: 0.3953
38
- - Prec Fear: 0.5341
39
- - Recall Fear: 0.5296
40
- - F1 Fear: 0.5318
41
- - Prec Neutral: 0.8416
42
- - Recall Neutral: 0.8279
43
- - F1 Neutral: 0.8347
44
- - Prec Sadness: 0.6410
45
- - Recall Sadness: 0.6338
46
- - F1 Sadness: 0.6374
47
- - Prec Surprise: 0.5093
48
- - Recall Surprise: 0.4663
49
- - F1 Surprise: 0.4869
50
 
51
  ## Model description
52
 
 
19
 
20
  This model is a fine-tuned version of [roberta-large](https://huggingface.co/roberta-large) on an unknown dataset.
21
  It achieves the following results on the evaluation set:
22
+ - Loss: 0.7106
23
+ - Accuracy: 0.7405
24
+ - Prec: 0.7387
25
+ - Recall: 0.7405
26
+ - F1: 0.7387
27
+ - B Acc: 0.5982
28
+ - Micro F1: 0.7405
29
+ - Prec Joy: 0.7558
30
+ - Recall Joy: 0.7617
31
+ - F1 Joy: 0.7587
32
+ - Prec Anger: 0.6294
33
+ - Recall Anger: 0.5631
34
+ - F1 Anger: 0.5944
35
+ - Prec Disgust: 0.4637
36
+ - Recall Disgust: 0.3854
37
+ - F1 Disgust: 0.4209
38
+ - Prec Fear: 0.4892
39
+ - Recall Fear: 0.5817
40
+ - F1 Fear: 0.5315
41
+ - Prec Neutral: 0.8292
42
+ - Recall Neutral: 0.8481
43
+ - F1 Neutral: 0.8385
44
+ - Prec Sadness: 0.6600
45
+ - Recall Sadness: 0.6140
46
+ - F1 Sadness: 0.6362
47
+ - Prec Surprise: 0.5320
48
+ - Recall Surprise: 0.4331
49
+ - F1 Surprise: 0.4775
50
 
51
  ## Model description
52
 
all_results.json ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 24.99,
3
+ "eval_accuracy": 0.7404651958269198,
4
+ "eval_b_acc": 0.598152273776674,
5
+ "eval_f1": 0.7387479859330638,
6
+ "eval_f1_anger": 0.594402837997635,
7
+ "eval_f1_disgust": 0.4209328782707622,
8
+ "eval_f1_fear": 0.5314520870076425,
9
+ "eval_f1_joy": 0.758746779689188,
10
+ "eval_f1_neutral": 0.8385366079274779,
11
+ "eval_f1_sadness": 0.6361726954492415,
12
+ "eval_f1_surprise": 0.4774640911342249,
13
+ "eval_loss": 0.7106355428695679,
14
+ "eval_micro_f1": 0.7404651958269197,
15
+ "eval_prec": 0.7386648012356265,
16
+ "eval_prec_anger": 0.6293823038397329,
17
+ "eval_prec_disgust": 0.46365914786967416,
18
+ "eval_prec_fear": 0.48917748917748916,
19
+ "eval_prec_joy": 0.7557947019867549,
20
+ "eval_prec_neutral": 0.8292169776802049,
21
+ "eval_prec_sadness": 0.660048426150121,
22
+ "eval_prec_surprise": 0.5320088300220751,
23
+ "eval_recall": 0.7404651958269198,
24
+ "eval_recall_anger": 0.5631067961165048,
25
+ "eval_recall_disgust": 0.3854166666666667,
26
+ "eval_recall_fear": 0.5817245817245817,
27
+ "eval_recall_joy": 0.7617220090105122,
28
+ "eval_recall_neutral": 0.8480681074001309,
29
+ "eval_recall_sadness": 0.6139639639639639,
30
+ "eval_recall_surprise": 0.4330637915543576,
31
+ "eval_runtime": 54.1882,
32
+ "eval_samples": 23388,
33
+ "eval_samples_per_second": 431.607,
34
+ "eval_steps_per_second": 13.49,
35
+ "train_loss": 0.6590935281284558,
36
+ "train_runtime": 28744.9325,
37
+ "train_samples": 209849,
38
+ "train_samples_per_second": 182.51,
39
+ "train_steps_per_second": 1.425
40
+ }
cf.png CHANGED
cf.txt CHANGED
@@ -1,7 +1,7 @@
1
- 5.728155339805824919e-01 7.766990291262135249e-02 2.165795369678864823e-02 8.439133681852128976e-02 1.202389843166542238e-01 7.692307692307692735e-02 4.630321135175503866e-02
2
- 1.729166666666666630e-01 3.895833333333333370e-01 1.270833333333333259e-01 6.875000000000000555e-02 6.666666666666666574e-02 1.583333333333333259e-01 1.666666666666666644e-02
3
- 2.509652509652509494e-02 2.445302445302445157e-02 5.296010296010296159e-01 1.126126126126126142e-01 1.370656370656370693e-01 1.351351351351351426e-01 3.603603603603603572e-02
4
- 1.001168029367595531e-02 4.505256132154179978e-03 2.052394460203570831e-02 7.892541298181211529e-01 1.331553479058902034e-01 2.469547805773402230e-02 1.785416319038878735e-02
5
- 1.871082421180653155e-02 4.397043689774534150e-03 1.861727009074749806e-02 9.027972682196651333e-02 8.278604172513799320e-01 1.777528300121620358e-02 2.235943493310880298e-02
6
- 3.738738738738738715e-02 2.477477477477477499e-02 1.031531531531531543e-01 7.297297297297297702e-02 1.153153153153153171e-01 6.337837837837837496e-01 1.261261261261261216e-02
7
- 3.773584905660377214e-02 7.187780772686433450e-03 6.918238993710691676e-02 1.805929919137466422e-01 1.841868823000898381e-01 5.480682839173404985e-02 4.663072776280323684e-01
 
1
+ 5.631067961165048263e-01 6.049290515309932448e-02 4.032860343539955428e-02 8.439133681852128976e-02 1.359223300970873738e-01 7.542942494398804931e-02 4.032860343539955428e-02
2
+ 1.520833333333333204e-01 3.854166666666666852e-01 1.708333333333333370e-01 5.416666666666666852e-02 6.666666666666666574e-02 1.562500000000000000e-01 1.458333333333333356e-02
3
+ 2.380952380952380820e-02 1.351351351351351426e-02 5.817245817245817463e-01 1.003861003861003798e-01 1.306306306306306286e-01 1.196911196911196879e-01 3.024453024453024538e-02
4
+ 9.344234940764225075e-03 4.171533455698314859e-03 3.237109961621892351e-02 7.617220090105122354e-01 1.493408977139996652e-01 2.252628066077089902e-02 2.052394460203570831e-02
5
+ 1.543642997474038714e-02 2.900177752830012182e-03 2.301431378052203047e-02 8.008232762653194614e-02 8.480681074001309439e-01 1.440733464309102742e-02 1.609130882215361463e-02
6
+ 3.558558558558558294e-02 2.207207207207207214e-02 1.247747747747747771e-01 6.531531531531531432e-02 1.288288288288288175e-01 6.139639639639639102e-01 9.459459459459459985e-03
7
+ 3.054806828391734216e-02 6.289308176100629269e-03 8.176100628930817182e-02 1.608265947888589398e-01 2.416891284815813057e-01 4.582210242587601151e-02 4.330637915543575911e-01
class_report.txt CHANGED
@@ -1,13 +1,13 @@
1
  precision recall f1-score support
2
 
3
- 0anger 0.60 0.57 0.59 1339
4
- 1disgust 0.40 0.39 0.40 480
5
- 2fear 0.53 0.53 0.53 1554
6
- 3joy 0.74 0.79 0.76 5993
7
- 4neutral 0.84 0.83 0.83 10689
8
- 5sadness 0.64 0.63 0.64 2220
9
- 6surprise 0.51 0.47 0.49 1113
10
 
11
  accuracy 0.74 23388
12
- macro avg 0.61 0.60 0.61 23388
13
  weighted avg 0.74 0.74 0.74 23388
 
1
  precision recall f1-score support
2
 
3
+ 0anger 0.63 0.56 0.59 1339
4
+ 1disgust 0.46 0.39 0.42 480
5
+ 2fear 0.49 0.58 0.53 1554
6
+ 3joy 0.76 0.76 0.76 5993
7
+ 4neutral 0.83 0.85 0.84 10689
8
+ 5sadness 0.66 0.61 0.64 2220
9
+ 6surprise 0.53 0.43 0.48 1113
10
 
11
  accuracy 0.74 23388
12
+ macro avg 0.62 0.60 0.61 23388
13
  weighted avg 0.74 0.74 0.74 23388
eval_results.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 24.99,
3
+ "eval_accuracy": 0.7404651958269198,
4
+ "eval_b_acc": 0.598152273776674,
5
+ "eval_f1": 0.7387479859330638,
6
+ "eval_f1_anger": 0.594402837997635,
7
+ "eval_f1_disgust": 0.4209328782707622,
8
+ "eval_f1_fear": 0.5314520870076425,
9
+ "eval_f1_joy": 0.758746779689188,
10
+ "eval_f1_neutral": 0.8385366079274779,
11
+ "eval_f1_sadness": 0.6361726954492415,
12
+ "eval_f1_surprise": 0.4774640911342249,
13
+ "eval_loss": 0.7106355428695679,
14
+ "eval_micro_f1": 0.7404651958269197,
15
+ "eval_prec": 0.7386648012356265,
16
+ "eval_prec_anger": 0.6293823038397329,
17
+ "eval_prec_disgust": 0.46365914786967416,
18
+ "eval_prec_fear": 0.48917748917748916,
19
+ "eval_prec_joy": 0.7557947019867549,
20
+ "eval_prec_neutral": 0.8292169776802049,
21
+ "eval_prec_sadness": 0.660048426150121,
22
+ "eval_prec_surprise": 0.5320088300220751,
23
+ "eval_recall": 0.7404651958269198,
24
+ "eval_recall_anger": 0.5631067961165048,
25
+ "eval_recall_disgust": 0.3854166666666667,
26
+ "eval_recall_fear": 0.5817245817245817,
27
+ "eval_recall_joy": 0.7617220090105122,
28
+ "eval_recall_neutral": 0.8480681074001309,
29
+ "eval_recall_sadness": 0.6139639639639639,
30
+ "eval_recall_surprise": 0.4330637915543576,
31
+ "eval_runtime": 54.1882,
32
+ "eval_samples": 23388,
33
+ "eval_samples_per_second": 431.607,
34
+ "eval_steps_per_second": 13.49
35
+ }
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 24.99,
3
+ "train_loss": 0.6590935281284558,
4
+ "train_runtime": 28744.9325,
5
+ "train_samples": 209849,
6
+ "train_samples_per_second": 182.51,
7
+ "train_steps_per_second": 1.425
8
+ }
trainer_state.json ADDED
@@ -0,0 +1,1287 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.7387479859330638,
3
+ "best_model_checkpoint": ".exp/lora/lora-roberta-large-no-anger-f4-0927/checkpoint-24588",
4
+ "epoch": 24.99237572430619,
5
+ "eval_steps": 2049,
6
+ "global_step": 40975,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.25,
13
+ "learning_rate": 0.00019863347974621767,
14
+ "loss": 1.157,
15
+ "step": 410
16
+ },
17
+ {
18
+ "epoch": 0.5,
19
+ "learning_rate": 0.0003977550024402147,
20
+ "loss": 0.8773,
21
+ "step": 820
22
+ },
23
+ {
24
+ "epoch": 0.75,
25
+ "learning_rate": 0.0005978526110297706,
26
+ "loss": 0.8365,
27
+ "step": 1230
28
+ },
29
+ {
30
+ "epoch": 1.0,
31
+ "learning_rate": 0.0007979502196193266,
32
+ "loss": 0.8167,
33
+ "step": 1640
34
+ },
35
+ {
36
+ "epoch": 1.25,
37
+ "eval_accuracy": 0.7129724645117154,
38
+ "eval_b_acc": 0.5271913381858306,
39
+ "eval_f1": 0.7029940172494629,
40
+ "eval_f1_anger": 0.4622467771639043,
41
+ "eval_f1_disgust": 0.3897550111358575,
42
+ "eval_f1_fear": 0.4693951466859834,
43
+ "eval_f1_joy": 0.7340311547020522,
44
+ "eval_f1_neutral": 0.8173319746577467,
45
+ "eval_f1_sadness": 0.5948575399583043,
46
+ "eval_f1_surprise": 0.4043659043659043,
47
+ "eval_loss": 0.7755911946296692,
48
+ "eval_micro_f1": 0.7129724645117155,
49
+ "eval_prec": 0.7003428461797957,
50
+ "eval_prec_anger": 0.602641056422569,
51
+ "eval_prec_disgust": 0.41866028708133973,
52
+ "eval_prec_fear": 0.5368682684341343,
53
+ "eval_prec_joy": 0.7252442996742671,
54
+ "eval_prec_neutral": 0.7763002861471133,
55
+ "eval_prec_sadness": 0.6123032904148784,
56
+ "eval_prec_surprise": 0.4796547472256473,
57
+ "eval_recall": 0.7129724645117154,
58
+ "eval_recall_anger": 0.37490664675130697,
59
+ "eval_recall_disgust": 0.3645833333333333,
60
+ "eval_recall_fear": 0.416988416988417,
61
+ "eval_recall_joy": 0.7430335391289838,
62
+ "eval_recall_neutral": 0.8629432126485171,
63
+ "eval_recall_sadness": 0.5783783783783784,
64
+ "eval_recall_surprise": 0.3495058400718778,
65
+ "eval_runtime": 52.691,
66
+ "eval_samples_per_second": 443.871,
67
+ "eval_steps_per_second": 13.873,
68
+ "step": 2049
69
+ },
70
+ {
71
+ "epoch": 1.25,
72
+ "learning_rate": 0.0009980478282088825,
73
+ "loss": 0.8037,
74
+ "step": 2050
75
+ },
76
+ {
77
+ "epoch": 1.5,
78
+ "learning_rate": 0.000989569953244618,
79
+ "loss": 0.8023,
80
+ "step": 2460
81
+ },
82
+ {
83
+ "epoch": 1.75,
84
+ "learning_rate": 0.0009790371474079022,
85
+ "loss": 0.7911,
86
+ "step": 2870
87
+ },
88
+ {
89
+ "epoch": 2.0,
90
+ "learning_rate": 0.0009685043415711864,
91
+ "loss": 0.79,
92
+ "step": 3280
93
+ },
94
+ {
95
+ "epoch": 2.25,
96
+ "learning_rate": 0.0009579972255048041,
97
+ "loss": 0.7639,
98
+ "step": 3690
99
+ },
100
+ {
101
+ "epoch": 2.5,
102
+ "eval_accuracy": 0.7293056268171711,
103
+ "eval_b_acc": 0.5661640466626477,
104
+ "eval_f1": 0.7224400663459573,
105
+ "eval_f1_anger": 0.564935064935065,
106
+ "eval_f1_disgust": 0.40478564307078757,
107
+ "eval_f1_fear": 0.49217002237136465,
108
+ "eval_f1_joy": 0.7486674866748666,
109
+ "eval_f1_neutral": 0.8254268594855161,
110
+ "eval_f1_sadness": 0.6280878316559926,
111
+ "eval_f1_surprise": 0.428341384863124,
112
+ "eval_loss": 0.7301676869392395,
113
+ "eval_micro_f1": 0.7293056268171713,
114
+ "eval_prec": 0.7205685165171505,
115
+ "eval_prec_anger": 0.6186666666666667,
116
+ "eval_prec_disgust": 0.3881453154875717,
117
+ "eval_prec_fear": 0.5851063829787234,
118
+ "eval_prec_joy": 0.7360528861657529,
119
+ "eval_prec_neutral": 0.796054918317692,
120
+ "eval_prec_sadness": 0.6380111524163569,
121
+ "eval_prec_surprise": 0.532,
122
+ "eval_recall": 0.7293056268171711,
123
+ "eval_recall_anger": 0.5197908887229276,
124
+ "eval_recall_disgust": 0.42291666666666666,
125
+ "eval_recall_fear": 0.4247104247104247,
126
+ "eval_recall_joy": 0.7617220090105122,
127
+ "eval_recall_neutral": 0.8570493030217982,
128
+ "eval_recall_sadness": 0.6184684684684685,
129
+ "eval_recall_surprise": 0.3584905660377358,
130
+ "eval_runtime": 52.3088,
131
+ "eval_samples_per_second": 447.114,
132
+ "eval_steps_per_second": 13.975,
133
+ "step": 4098
134
+ },
135
+ {
136
+ "epoch": 2.5,
137
+ "learning_rate": 0.0009474644196680881,
138
+ "loss": 0.7619,
139
+ "step": 4100
140
+ },
141
+ {
142
+ "epoch": 2.75,
143
+ "learning_rate": 0.0009369316138313723,
144
+ "loss": 0.7521,
145
+ "step": 4510
146
+ },
147
+ {
148
+ "epoch": 3.0,
149
+ "learning_rate": 0.0009263988079946566,
150
+ "loss": 0.7603,
151
+ "step": 4920
152
+ },
153
+ {
154
+ "epoch": 3.25,
155
+ "learning_rate": 0.0009158660021579407,
156
+ "loss": 0.7425,
157
+ "step": 5330
158
+ },
159
+ {
160
+ "epoch": 3.5,
161
+ "learning_rate": 0.0009053331963212249,
162
+ "loss": 0.7395,
163
+ "step": 5740
164
+ },
165
+ {
166
+ "epoch": 3.75,
167
+ "eval_accuracy": 0.7286642722763811,
168
+ "eval_b_acc": 0.5792854233134773,
169
+ "eval_f1": 0.7270638257323497,
170
+ "eval_f1_anger": 0.5326678765880217,
171
+ "eval_f1_disgust": 0.4021621621621622,
172
+ "eval_f1_fear": 0.5162393162393162,
173
+ "eval_f1_joy": 0.7518889745566691,
174
+ "eval_f1_neutral": 0.8303328398834822,
175
+ "eval_f1_sadness": 0.6280009507962918,
176
+ "eval_f1_surprise": 0.46755819712729074,
177
+ "eval_loss": 0.7348284721374512,
178
+ "eval_micro_f1": 0.7286642722763811,
179
+ "eval_prec": 0.7327576856576087,
180
+ "eval_prec_anger": 0.6786127167630058,
181
+ "eval_prec_disgust": 0.41797752808988764,
182
+ "eval_prec_fear": 0.46319018404907975,
183
+ "eval_prec_joy": 0.6988677081840332,
184
+ "eval_prec_neutral": 0.8480296527506828,
185
+ "eval_prec_sadness": 0.6648213387015601,
186
+ "eval_prec_surprise": 0.5209713024282561,
187
+ "eval_recall": 0.7286642722763811,
188
+ "eval_recall_anger": 0.438386855862584,
189
+ "eval_recall_disgust": 0.3875,
190
+ "eval_recall_fear": 0.583011583011583,
191
+ "eval_recall_joy": 0.8136158851993993,
192
+ "eval_recall_neutral": 0.8133595284872298,
193
+ "eval_recall_sadness": 0.5950450450450451,
194
+ "eval_recall_surprise": 0.42407906558849956,
195
+ "eval_runtime": 52.2656,
196
+ "eval_samples_per_second": 447.483,
197
+ "eval_steps_per_second": 13.986,
198
+ "step": 6147
199
+ },
200
+ {
201
+ "epoch": 3.75,
202
+ "learning_rate": 0.0008948003904845092,
203
+ "loss": 0.7441,
204
+ "step": 6150
205
+ },
206
+ {
207
+ "epoch": 4.0,
208
+ "learning_rate": 0.0008842675846477932,
209
+ "loss": 0.7408,
210
+ "step": 6560
211
+ },
212
+ {
213
+ "epoch": 4.25,
214
+ "learning_rate": 0.0008737347788110774,
215
+ "loss": 0.7383,
216
+ "step": 6970
217
+ },
218
+ {
219
+ "epoch": 4.5,
220
+ "learning_rate": 0.0008632019729743616,
221
+ "loss": 0.7479,
222
+ "step": 7380
223
+ },
224
+ {
225
+ "epoch": 4.75,
226
+ "learning_rate": 0.0008527205466783127,
227
+ "loss": 0.789,
228
+ "step": 7790
229
+ },
230
+ {
231
+ "epoch": 5.0,
232
+ "eval_accuracy": 0.7274670771335727,
233
+ "eval_b_acc": 0.55106785419059,
234
+ "eval_f1": 0.7180205205088754,
235
+ "eval_f1_anger": 0.556758924989972,
236
+ "eval_f1_disgust": 0.40292275574112735,
237
+ "eval_f1_fear": 0.4922737306843267,
238
+ "eval_f1_joy": 0.7450199203187251,
239
+ "eval_f1_neutral": 0.8237609262733973,
240
+ "eval_f1_sadness": 0.6308104161596496,
241
+ "eval_f1_surprise": 0.37617554858934166,
242
+ "eval_loss": 0.7419339418411255,
243
+ "eval_micro_f1": 0.7274670771335727,
244
+ "eval_prec": 0.7206445931565618,
245
+ "eval_prec_anger": 0.6013864818024264,
246
+ "eval_prec_disgust": 0.40376569037656906,
247
+ "eval_prec_fear": 0.5747422680412371,
248
+ "eval_prec_joy": 0.6887661141804788,
249
+ "eval_prec_neutral": 0.806307113420534,
250
+ "eval_prec_sadness": 0.6860772895712017,
251
+ "eval_prec_surprise": 0.6224066390041494,
252
+ "eval_recall": 0.7274670771335727,
253
+ "eval_recall_anger": 0.5182972367438387,
254
+ "eval_recall_disgust": 0.40208333333333335,
255
+ "eval_recall_fear": 0.4305019305019305,
256
+ "eval_recall_joy": 0.8112798264642083,
257
+ "eval_recall_neutral": 0.8419870895312939,
258
+ "eval_recall_sadness": 0.5837837837837838,
259
+ "eval_recall_surprise": 0.2695417789757412,
260
+ "eval_runtime": 52.2211,
261
+ "eval_samples_per_second": 447.865,
262
+ "eval_steps_per_second": 13.998,
263
+ "step": 8196
264
+ },
265
+ {
266
+ "epoch": 5.0,
267
+ "learning_rate": 0.0008421877408415969,
268
+ "loss": 0.7457,
269
+ "step": 8200
270
+ },
271
+ {
272
+ "epoch": 5.25,
273
+ "learning_rate": 0.000831706314545548,
274
+ "loss": 0.7311,
275
+ "step": 8610
276
+ },
277
+ {
278
+ "epoch": 5.5,
279
+ "learning_rate": 0.0008211735087088321,
280
+ "loss": 0.7408,
281
+ "step": 9020
282
+ },
283
+ {
284
+ "epoch": 5.75,
285
+ "learning_rate": 0.0008106407028721164,
286
+ "loss": 0.7309,
287
+ "step": 9430
288
+ },
289
+ {
290
+ "epoch": 6.0,
291
+ "learning_rate": 0.0008001078970354006,
292
+ "loss": 0.7439,
293
+ "step": 9840
294
+ },
295
+ {
296
+ "epoch": 6.25,
297
+ "eval_accuracy": 0.7207114759705832,
298
+ "eval_b_acc": 0.5857766516959649,
299
+ "eval_f1": 0.7223601133500003,
300
+ "eval_f1_anger": 0.5537190082644629,
301
+ "eval_f1_disgust": 0.3732162458836443,
302
+ "eval_f1_fear": 0.5119819565830279,
303
+ "eval_f1_joy": 0.7459492509935799,
304
+ "eval_f1_neutral": 0.8227773982629062,
305
+ "eval_f1_sadness": 0.6234614418487817,
306
+ "eval_f1_surprise": 0.47541684480547247,
307
+ "eval_loss": 0.7608067989349365,
308
+ "eval_micro_f1": 0.720711475970583,
309
+ "eval_prec": 0.7316542206741192,
310
+ "eval_prec_anger": 0.6197964847363552,
311
+ "eval_prec_disgust": 0.39443155452436196,
312
+ "eval_prec_fear": 0.45559458103361766,
313
+ "eval_prec_joy": 0.6881963051755746,
314
+ "eval_prec_neutral": 0.8598674145843958,
315
+ "eval_prec_sadness": 0.7047132311186826,
316
+ "eval_prec_surprise": 0.4535073409461664,
317
+ "eval_recall": 0.7207114759705832,
318
+ "eval_recall_anger": 0.5003734129947722,
319
+ "eval_recall_disgust": 0.3541666666666667,
320
+ "eval_recall_fear": 0.5842985842985843,
321
+ "eval_recall_joy": 0.8142833305523111,
322
+ "eval_recall_neutral": 0.7887547946487042,
323
+ "eval_recall_sadness": 0.559009009009009,
324
+ "eval_recall_surprise": 0.4995507637017071,
325
+ "eval_runtime": 52.2461,
326
+ "eval_samples_per_second": 447.651,
327
+ "eval_steps_per_second": 13.991,
328
+ "step": 10245
329
+ },
330
+ {
331
+ "epoch": 6.25,
332
+ "learning_rate": 0.0007895750911986846,
333
+ "loss": 0.7377,
334
+ "step": 10250
335
+ },
336
+ {
337
+ "epoch": 6.5,
338
+ "learning_rate": 0.0007790422853619688,
339
+ "loss": 0.7175,
340
+ "step": 10660
341
+ },
342
+ {
343
+ "epoch": 6.75,
344
+ "learning_rate": 0.0007685094795252531,
345
+ "loss": 0.7252,
346
+ "step": 11070
347
+ },
348
+ {
349
+ "epoch": 7.0,
350
+ "learning_rate": 0.0007579766736885373,
351
+ "loss": 0.7264,
352
+ "step": 11480
353
+ },
354
+ {
355
+ "epoch": 7.25,
356
+ "learning_rate": 0.0007474438678518214,
357
+ "loss": 0.712,
358
+ "step": 11890
359
+ },
360
+ {
361
+ "epoch": 7.5,
362
+ "eval_accuracy": 0.7298187104498033,
363
+ "eval_b_acc": 0.5809376473431288,
364
+ "eval_f1": 0.7262773858274297,
365
+ "eval_f1_anger": 0.5449915110356537,
366
+ "eval_f1_disgust": 0.4205128205128206,
367
+ "eval_f1_fear": 0.5043103448275862,
368
+ "eval_f1_joy": 0.751774155813772,
369
+ "eval_f1_neutral": 0.8270924591399369,
370
+ "eval_f1_sadness": 0.6288094716071038,
371
+ "eval_f1_surprise": 0.4750733137829912,
372
+ "eval_loss": 0.7239968180656433,
373
+ "eval_micro_f1": 0.7298187104498032,
374
+ "eval_prec": 0.7269696751543612,
375
+ "eval_prec_anger": 0.6312684365781711,
376
+ "eval_prec_disgust": 0.41414141414141414,
377
+ "eval_prec_fear": 0.5707317073170731,
378
+ "eval_prec_joy": 0.705710102489019,
379
+ "eval_prec_neutral": 0.8328590400303548,
380
+ "eval_prec_sadness": 0.6125587355830842,
381
+ "eval_prec_surprise": 0.5209003215434084,
382
+ "eval_recall": 0.7298187104498033,
383
+ "eval_recall_anger": 0.479462285287528,
384
+ "eval_recall_disgust": 0.4270833333333333,
385
+ "eval_recall_fear": 0.4517374517374517,
386
+ "eval_recall_joy": 0.8042716502586351,
387
+ "eval_recall_neutral": 0.8214051828983067,
388
+ "eval_recall_sadness": 0.6459459459459459,
389
+ "eval_recall_surprise": 0.4366576819407008,
390
+ "eval_runtime": 52.2383,
391
+ "eval_samples_per_second": 447.718,
392
+ "eval_steps_per_second": 13.994,
393
+ "step": 12294
394
+ },
395
+ {
396
+ "epoch": 7.5,
397
+ "learning_rate": 0.0007369110620151056,
398
+ "loss": 0.7081,
399
+ "step": 12300
400
+ },
401
+ {
402
+ "epoch": 7.75,
403
+ "learning_rate": 0.0007263782561783898,
404
+ "loss": 0.7139,
405
+ "step": 12710
406
+ },
407
+ {
408
+ "epoch": 8.0,
409
+ "learning_rate": 0.0007158454503416739,
410
+ "loss": 0.7194,
411
+ "step": 13120
412
+ },
413
+ {
414
+ "epoch": 8.25,
415
+ "learning_rate": 0.0007053126445049581,
416
+ "loss": 0.6946,
417
+ "step": 13530
418
+ },
419
+ {
420
+ "epoch": 8.5,
421
+ "learning_rate": 0.0006947798386682424,
422
+ "loss": 0.7032,
423
+ "step": 13940
424
+ },
425
+ {
426
+ "epoch": 8.75,
427
+ "eval_accuracy": 0.7343937061741064,
428
+ "eval_b_acc": 0.5832769765478361,
429
+ "eval_f1": 0.7316967540687647,
430
+ "eval_f1_anger": 0.5795918367346938,
431
+ "eval_f1_disgust": 0.3876146788990826,
432
+ "eval_f1_fear": 0.5310033821871477,
433
+ "eval_f1_joy": 0.7517611539751762,
434
+ "eval_f1_neutral": 0.8340448973970528,
435
+ "eval_f1_sadness": 0.6301433194637078,
436
+ "eval_f1_surprise": 0.45488521089161776,
437
+ "eval_loss": 0.7095019817352295,
438
+ "eval_micro_f1": 0.7343937061741064,
439
+ "eval_prec": 0.7327970959588296,
440
+ "eval_prec_anger": 0.639063906390639,
441
+ "eval_prec_disgust": 0.43112244897959184,
442
+ "eval_prec_fear": 0.4724172517552658,
443
+ "eval_prec_joy": 0.7556904400606981,
444
+ "eval_prec_neutral": 0.8188209843158464,
445
+ "eval_prec_sadness": 0.6471984805318138,
446
+ "eval_prec_surprise": 0.5605263157894737,
447
+ "eval_recall": 0.7343937061741064,
448
+ "eval_recall_anger": 0.5302464525765497,
449
+ "eval_recall_disgust": 0.35208333333333336,
450
+ "eval_recall_fear": 0.6061776061776062,
451
+ "eval_recall_joy": 0.7478725179375939,
452
+ "eval_recall_neutral": 0.8498456357002526,
453
+ "eval_recall_sadness": 0.6139639639639639,
454
+ "eval_recall_surprise": 0.38274932614555257,
455
+ "eval_runtime": 52.2132,
456
+ "eval_samples_per_second": 447.933,
457
+ "eval_steps_per_second": 14.0,
458
+ "step": 14343
459
+ },
460
+ {
461
+ "epoch": 8.75,
462
+ "learning_rate": 0.0006842470328315265,
463
+ "loss": 0.7012,
464
+ "step": 14350
465
+ },
466
+ {
467
+ "epoch": 9.0,
468
+ "learning_rate": 0.0006737142269948107,
469
+ "loss": 0.7012,
470
+ "step": 14760
471
+ },
472
+ {
473
+ "epoch": 9.25,
474
+ "learning_rate": 0.0006631814211580948,
475
+ "loss": 0.6766,
476
+ "step": 15170
477
+ },
478
+ {
479
+ "epoch": 9.5,
480
+ "learning_rate": 0.0006526486153213791,
481
+ "loss": 0.6987,
482
+ "step": 15580
483
+ },
484
+ {
485
+ "epoch": 9.75,
486
+ "learning_rate": 0.0006421158094846632,
487
+ "loss": 0.6972,
488
+ "step": 15990
489
+ },
490
+ {
491
+ "epoch": 10.0,
492
+ "eval_accuracy": 0.7343081922353344,
493
+ "eval_b_acc": 0.5923420613622242,
494
+ "eval_f1": 0.7316601301789013,
495
+ "eval_f1_anger": 0.5625479662317728,
496
+ "eval_f1_disgust": 0.40974729241877256,
497
+ "eval_f1_fear": 0.5255474452554744,
498
+ "eval_f1_joy": 0.7572113495244833,
499
+ "eval_f1_neutral": 0.8305052967094779,
500
+ "eval_f1_sadness": 0.6358187485015584,
501
+ "eval_f1_surprise": 0.46601941747572817,
502
+ "eval_loss": 0.7108047008514404,
503
+ "eval_micro_f1": 0.7343081922353344,
504
+ "eval_prec": 0.7324880260431258,
505
+ "eval_prec_anger": 0.5785319652722968,
506
+ "eval_prec_disgust": 0.3614649681528662,
507
+ "eval_prec_fear": 0.5714285714285714,
508
+ "eval_prec_joy": 0.7157503714710253,
509
+ "eval_prec_neutral": 0.8322217003287928,
510
+ "eval_prec_sadness": 0.6796514607893388,
511
+ "eval_prec_surprise": 0.5402843601895735,
512
+ "eval_recall": 0.7343081922353344,
513
+ "eval_recall_anger": 0.5474234503360716,
514
+ "eval_recall_disgust": 0.47291666666666665,
515
+ "eval_recall_fear": 0.4864864864864865,
516
+ "eval_recall_joy": 0.8037710662439512,
517
+ "eval_recall_neutral": 0.8287959584619703,
518
+ "eval_recall_sadness": 0.5972972972972973,
519
+ "eval_recall_surprise": 0.40970350404312667,
520
+ "eval_runtime": 52.2344,
521
+ "eval_samples_per_second": 447.751,
522
+ "eval_steps_per_second": 13.995,
523
+ "step": 16392
524
+ },
525
+ {
526
+ "epoch": 10.0,
527
+ "learning_rate": 0.0006315830036479474,
528
+ "loss": 0.689,
529
+ "step": 16400
530
+ },
531
+ {
532
+ "epoch": 10.25,
533
+ "learning_rate": 0.0006210501978112316,
534
+ "loss": 0.678,
535
+ "step": 16810
536
+ },
537
+ {
538
+ "epoch": 10.5,
539
+ "learning_rate": 0.0006105173919745158,
540
+ "loss": 0.6774,
541
+ "step": 17220
542
+ },
543
+ {
544
+ "epoch": 10.75,
545
+ "learning_rate": 0.0006000102759081334,
546
+ "loss": 0.6775,
547
+ "step": 17630
548
+ },
549
+ {
550
+ "epoch": 11.0,
551
+ "learning_rate": 0.0005894774700714176,
552
+ "loss": 0.6859,
553
+ "step": 18040
554
+ },
555
+ {
556
+ "epoch": 11.25,
557
+ "eval_accuracy": 0.7376432358474432,
558
+ "eval_b_acc": 0.5791605614639431,
559
+ "eval_f1": 0.7321535509780166,
560
+ "eval_f1_anger": 0.573612228479485,
561
+ "eval_f1_disgust": 0.4152892561983471,
562
+ "eval_f1_fear": 0.5193253689388616,
563
+ "eval_f1_joy": 0.7545115121344119,
564
+ "eval_f1_neutral": 0.8334647312526087,
565
+ "eval_f1_sadness": 0.6374764595103578,
566
+ "eval_f1_surprise": 0.4521835268103925,
567
+ "eval_loss": 0.7211122512817383,
568
+ "eval_micro_f1": 0.7376432358474432,
569
+ "eval_prec": 0.7321470081981006,
570
+ "eval_prec_anger": 0.6216216216216216,
571
+ "eval_prec_disgust": 0.41188524590163933,
572
+ "eval_prec_fear": 0.5719814241486069,
573
+ "eval_prec_joy": 0.7066880373014717,
574
+ "eval_prec_neutral": 0.8263748390656612,
575
+ "eval_prec_sadness": 0.6676528599605522,
576
+ "eval_prec_surprise": 0.5876436781609196,
577
+ "eval_recall": 0.7376432358474432,
578
+ "eval_recall_anger": 0.5324869305451829,
579
+ "eval_recall_disgust": 0.41875,
580
+ "eval_recall_fear": 0.47554697554697556,
581
+ "eval_recall_joy": 0.809277490405473,
582
+ "eval_recall_neutral": 0.8406773318364674,
583
+ "eval_recall_sadness": 0.6099099099099099,
584
+ "eval_recall_surprise": 0.3674752920035939,
585
+ "eval_runtime": 52.2223,
586
+ "eval_samples_per_second": 447.855,
587
+ "eval_steps_per_second": 13.998,
588
+ "step": 18441
589
+ },
590
+ {
591
+ "epoch": 11.25,
592
+ "learning_rate": 0.0005789446642347017,
593
+ "loss": 0.6627,
594
+ "step": 18450
595
+ },
596
+ {
597
+ "epoch": 11.5,
598
+ "learning_rate": 0.0005684118583979859,
599
+ "loss": 0.672,
600
+ "step": 18860
601
+ },
602
+ {
603
+ "epoch": 11.75,
604
+ "learning_rate": 0.0005578790525612702,
605
+ "loss": 0.6734,
606
+ "step": 19270
607
+ },
608
+ {
609
+ "epoch": 12.0,
610
+ "learning_rate": 0.0005473462467245543,
611
+ "loss": 0.6741,
612
+ "step": 19680
613
+ },
614
+ {
615
+ "epoch": 12.25,
616
+ "learning_rate": 0.0005368391306581719,
617
+ "loss": 0.6542,
618
+ "step": 20090
619
+ },
620
+ {
621
+ "epoch": 12.5,
622
+ "eval_accuracy": 0.7347357619291944,
623
+ "eval_b_acc": 0.5816682643851953,
624
+ "eval_f1": 0.7307386267588237,
625
+ "eval_f1_anger": 0.5803607214428858,
626
+ "eval_f1_disgust": 0.3860021208907741,
627
+ "eval_f1_fear": 0.5130927105449399,
628
+ "eval_f1_joy": 0.7583697234352255,
629
+ "eval_f1_neutral": 0.8282762773891694,
630
+ "eval_f1_sadness": 0.6398268398268399,
631
+ "eval_f1_surprise": 0.4600309437854564,
632
+ "eval_loss": 0.7143167853355408,
633
+ "eval_micro_f1": 0.7347357619291943,
634
+ "eval_prec": 0.7293955660159446,
635
+ "eval_prec_anger": 0.6262975778546713,
636
+ "eval_prec_disgust": 0.3930885529157667,
637
+ "eval_prec_fear": 0.5699685534591195,
638
+ "eval_prec_joy": 0.735760238506198,
639
+ "eval_prec_neutral": 0.820334006239677,
640
+ "eval_prec_sadness": 0.6158333333333333,
641
+ "eval_prec_surprise": 0.5399515738498789,
642
+ "eval_recall": 0.7347357619291944,
643
+ "eval_recall_anger": 0.5407020164301718,
644
+ "eval_recall_disgust": 0.37916666666666665,
645
+ "eval_recall_fear": 0.46653796653796653,
646
+ "eval_recall_joy": 0.7824128149507759,
647
+ "eval_recall_neutral": 0.8363738422677519,
648
+ "eval_recall_sadness": 0.6657657657657657,
649
+ "eval_recall_surprise": 0.40071877807726863,
650
+ "eval_runtime": 52.2156,
651
+ "eval_samples_per_second": 447.912,
652
+ "eval_steps_per_second": 14.0,
653
+ "step": 20490
654
+ },
655
+ {
656
+ "epoch": 12.5,
657
+ "learning_rate": 0.0005263063248214561,
658
+ "loss": 0.6562,
659
+ "step": 20500
660
+ },
661
+ {
662
+ "epoch": 12.75,
663
+ "learning_rate": 0.0005157735189847404,
664
+ "loss": 0.6563,
665
+ "step": 20910
666
+ },
667
+ {
668
+ "epoch": 13.0,
669
+ "learning_rate": 0.0005052407131480245,
670
+ "loss": 0.6569,
671
+ "step": 21320
672
+ },
673
+ {
674
+ "epoch": 13.25,
675
+ "learning_rate": 0.0004947079073113087,
676
+ "loss": 0.6391,
677
+ "step": 21730
678
+ },
679
+ {
680
+ "epoch": 13.5,
681
+ "learning_rate": 0.0004841751014745928,
682
+ "loss": 0.6463,
683
+ "step": 22140
684
+ },
685
+ {
686
+ "epoch": 13.75,
687
+ "eval_accuracy": 0.736916367367881,
688
+ "eval_b_acc": 0.5947417537959515,
689
+ "eval_f1": 0.7353543182480966,
690
+ "eval_f1_anger": 0.5885183911997249,
691
+ "eval_f1_disgust": 0.3927648578811369,
692
+ "eval_f1_fear": 0.5167693360711841,
693
+ "eval_f1_joy": 0.7609590699927343,
694
+ "eval_f1_neutral": 0.833475661827498,
695
+ "eval_f1_sadness": 0.634156820622986,
696
+ "eval_f1_surprise": 0.4865919063871283,
697
+ "eval_loss": 0.7021653056144714,
698
+ "eval_micro_f1": 0.736916367367881,
699
+ "eval_prec": 0.7366478429202861,
700
+ "eval_prec_anger": 0.5452229299363057,
701
+ "eval_prec_disgust": 0.5170068027210885,
702
+ "eval_prec_fear": 0.5519005847953217,
703
+ "eval_prec_joy": 0.7370972786987801,
704
+ "eval_prec_neutral": 0.8455096736933295,
705
+ "eval_prec_sadness": 0.6061601642710472,
706
+ "eval_prec_surprise": 0.5319829424307037,
707
+ "eval_recall": 0.736916367367881,
708
+ "eval_recall_anger": 0.6392830470500374,
709
+ "eval_recall_disgust": 0.31666666666666665,
710
+ "eval_recall_fear": 0.48584298584298585,
711
+ "eval_recall_joy": 0.7864174870682463,
712
+ "eval_recall_neutral": 0.8217793993825429,
713
+ "eval_recall_sadness": 0.6648648648648648,
714
+ "eval_recall_surprise": 0.44833782569631625,
715
+ "eval_runtime": 52.3587,
716
+ "eval_samples_per_second": 446.688,
717
+ "eval_steps_per_second": 13.961,
718
+ "step": 22539
719
+ },
720
+ {
721
+ "epoch": 13.75,
722
+ "learning_rate": 0.00047364229563787706,
723
+ "loss": 0.6502,
724
+ "step": 22550
725
+ },
726
+ {
727
+ "epoch": 14.0,
728
+ "learning_rate": 0.0004631094898011612,
729
+ "loss": 0.6445,
730
+ "step": 22960
731
+ },
732
+ {
733
+ "epoch": 14.25,
734
+ "learning_rate": 0.00045260237373477883,
735
+ "loss": 0.6274,
736
+ "step": 23370
737
+ },
738
+ {
739
+ "epoch": 14.5,
740
+ "learning_rate": 0.000442069567898063,
741
+ "loss": 0.6373,
742
+ "step": 23780
743
+ },
744
+ {
745
+ "epoch": 14.75,
746
+ "learning_rate": 0.0004315367620613472,
747
+ "loss": 0.6333,
748
+ "step": 24190
749
+ },
750
+ {
751
+ "epoch": 15.0,
752
+ "eval_accuracy": 0.7404651958269198,
753
+ "eval_b_acc": 0.598152273776674,
754
+ "eval_f1": 0.7387479859330638,
755
+ "eval_f1_anger": 0.594402837997635,
756
+ "eval_f1_disgust": 0.4209328782707622,
757
+ "eval_f1_fear": 0.5314520870076425,
758
+ "eval_f1_joy": 0.758746779689188,
759
+ "eval_f1_neutral": 0.8385366079274779,
760
+ "eval_f1_sadness": 0.6361726954492415,
761
+ "eval_f1_surprise": 0.4774640911342249,
762
+ "eval_loss": 0.7106355428695679,
763
+ "eval_micro_f1": 0.7404651958269197,
764
+ "eval_prec": 0.7386648012356265,
765
+ "eval_prec_anger": 0.6293823038397329,
766
+ "eval_prec_disgust": 0.46365914786967416,
767
+ "eval_prec_fear": 0.48917748917748916,
768
+ "eval_prec_joy": 0.7557947019867549,
769
+ "eval_prec_neutral": 0.8292169776802049,
770
+ "eval_prec_sadness": 0.660048426150121,
771
+ "eval_prec_surprise": 0.5320088300220751,
772
+ "eval_recall": 0.7404651958269198,
773
+ "eval_recall_anger": 0.5631067961165048,
774
+ "eval_recall_disgust": 0.3854166666666667,
775
+ "eval_recall_fear": 0.5817245817245817,
776
+ "eval_recall_joy": 0.7617220090105122,
777
+ "eval_recall_neutral": 0.8480681074001309,
778
+ "eval_recall_sadness": 0.6139639639639639,
779
+ "eval_recall_surprise": 0.4330637915543576,
780
+ "eval_runtime": 52.1895,
781
+ "eval_samples_per_second": 448.136,
782
+ "eval_steps_per_second": 14.007,
783
+ "step": 24588
784
+ },
785
+ {
786
+ "epoch": 15.0,
787
+ "learning_rate": 0.00042100395622463136,
788
+ "loss": 0.6432,
789
+ "step": 24600
790
+ },
791
+ {
792
+ "epoch": 15.25,
793
+ "learning_rate": 0.00041047115038791553,
794
+ "loss": 0.6199,
795
+ "step": 25010
796
+ },
797
+ {
798
+ "epoch": 15.5,
799
+ "learning_rate": 0.0003999383445511997,
800
+ "loss": 0.6174,
801
+ "step": 25420
802
+ },
803
+ {
804
+ "epoch": 15.75,
805
+ "learning_rate": 0.00038940553871448394,
806
+ "loss": 0.6308,
807
+ "step": 25830
808
+ },
809
+ {
810
+ "epoch": 16.0,
811
+ "learning_rate": 0.00037887273287776806,
812
+ "loss": 0.6184,
813
+ "step": 26240
814
+ },
815
+ {
816
+ "epoch": 16.25,
817
+ "eval_accuracy": 0.7337523516333162,
818
+ "eval_b_acc": 0.6077119026145958,
819
+ "eval_f1": 0.7348413799234946,
820
+ "eval_f1_anger": 0.5841816758026626,
821
+ "eval_f1_disgust": 0.396917148362235,
822
+ "eval_f1_fear": 0.5457218465157306,
823
+ "eval_f1_joy": 0.7586152386974392,
824
+ "eval_f1_neutral": 0.8300653594771242,
825
+ "eval_f1_sadness": 0.6336822074215033,
826
+ "eval_f1_surprise": 0.48513449740443604,
827
+ "eval_loss": 0.719868540763855,
828
+ "eval_micro_f1": 0.7337523516333162,
829
+ "eval_prec": 0.7388716422973962,
830
+ "eval_prec_anger": 0.6139917695473252,
831
+ "eval_prec_disgust": 0.36917562724014336,
832
+ "eval_prec_fear": 0.5024363833243097,
833
+ "eval_prec_joy": 0.7206787805976873,
834
+ "eval_prec_neutral": 0.8534440162071351,
835
+ "eval_prec_sadness": 0.6713709677419355,
836
+ "eval_prec_surprise": 0.510934393638171,
837
+ "eval_recall": 0.7337523516333162,
838
+ "eval_recall_anger": 0.5571321882001493,
839
+ "eval_recall_disgust": 0.42916666666666664,
840
+ "eval_recall_fear": 0.5971685971685972,
841
+ "eval_recall_joy": 0.8007675621558485,
842
+ "eval_recall_neutral": 0.807933389465806,
843
+ "eval_recall_sadness": 0.6,
844
+ "eval_recall_surprise": 0.4618149146451033,
845
+ "eval_runtime": 52.0741,
846
+ "eval_samples_per_second": 449.129,
847
+ "eval_steps_per_second": 14.038,
848
+ "step": 26637
849
+ },
850
+ {
851
+ "epoch": 16.25,
852
+ "learning_rate": 0.00036833992704105224,
853
+ "loss": 0.6063,
854
+ "step": 26650
855
+ },
856
+ {
857
+ "epoch": 16.51,
858
+ "learning_rate": 0.00035780712120433647,
859
+ "loss": 0.6049,
860
+ "step": 27060
861
+ },
862
+ {
863
+ "epoch": 16.76,
864
+ "learning_rate": 0.0003472743153676206,
865
+ "loss": 0.611,
866
+ "step": 27470
867
+ },
868
+ {
869
+ "epoch": 17.01,
870
+ "learning_rate": 0.0003367671993012383,
871
+ "loss": 0.6121,
872
+ "step": 27880
873
+ },
874
+ {
875
+ "epoch": 17.26,
876
+ "learning_rate": 0.0003262343934645224,
877
+ "loss": 0.5916,
878
+ "step": 28290
879
+ },
880
+ {
881
+ "epoch": 17.5,
882
+ "eval_accuracy": 0.736788096459723,
883
+ "eval_b_acc": 0.6002649432660556,
884
+ "eval_f1": 0.7362620116350573,
885
+ "eval_f1_anger": 0.5784499054820416,
886
+ "eval_f1_disgust": 0.39246778989098113,
887
+ "eval_f1_fear": 0.5149572649572649,
888
+ "eval_f1_joy": 0.7636186770428015,
889
+ "eval_f1_neutral": 0.8366824644549764,
890
+ "eval_f1_sadness": 0.6251845602193629,
891
+ "eval_f1_surprise": 0.4932147870846982,
892
+ "eval_loss": 0.7219593524932861,
893
+ "eval_micro_f1": 0.736788096459723,
894
+ "eval_prec": 0.7375843989348974,
895
+ "eval_prec_anger": 0.5857580398162328,
896
+ "eval_prec_disgust": 0.3742911153119093,
897
+ "eval_prec_fear": 0.5765550239234449,
898
+ "eval_prec_joy": 0.7425508434494719,
899
+ "eval_prec_neutral": 0.8478532321582941,
900
+ "eval_prec_sadness": 0.5878619595398651,
901
+ "eval_prec_surprise": 0.5146484375,
902
+ "eval_recall": 0.736788096459723,
903
+ "eval_recall_anger": 0.5713218820014937,
904
+ "eval_recall_disgust": 0.4125,
905
+ "eval_recall_fear": 0.46525096525096526,
906
+ "eval_recall_joy": 0.7859169030535625,
907
+ "eval_recall_neutral": 0.8258022265880812,
908
+ "eval_recall_sadness": 0.6675675675675675,
909
+ "eval_recall_surprise": 0.47349505840071876,
910
+ "eval_runtime": 52.1004,
911
+ "eval_samples_per_second": 448.902,
912
+ "eval_steps_per_second": 14.031,
913
+ "step": 28686
914
+ },
915
+ {
916
+ "epoch": 17.51,
917
+ "learning_rate": 0.0003157015876278066,
918
+ "loss": 0.6019,
919
+ "step": 28700
920
+ },
921
+ {
922
+ "epoch": 17.76,
923
+ "learning_rate": 0.0003051687817910908,
924
+ "loss": 0.5998,
925
+ "step": 29110
926
+ },
927
+ {
928
+ "epoch": 18.01,
929
+ "learning_rate": 0.00029463597595437495,
930
+ "loss": 0.5937,
931
+ "step": 29520
932
+ },
933
+ {
934
+ "epoch": 18.26,
935
+ "learning_rate": 0.0002841031701176591,
936
+ "loss": 0.5805,
937
+ "step": 29930
938
+ },
939
+ {
940
+ "epoch": 18.51,
941
+ "learning_rate": 0.00027357036428094335,
942
+ "loss": 0.5823,
943
+ "step": 30340
944
+ },
945
+ {
946
+ "epoch": 18.75,
947
+ "eval_accuracy": 0.7375577219086711,
948
+ "eval_b_acc": 0.5959965915979468,
949
+ "eval_f1": 0.7364086067599112,
950
+ "eval_f1_anger": 0.5850769838136597,
951
+ "eval_f1_disgust": 0.3828382838283828,
952
+ "eval_f1_fear": 0.5406240345999382,
953
+ "eval_f1_joy": 0.7610117406035773,
954
+ "eval_f1_neutral": 0.8328186612190602,
955
+ "eval_f1_sadness": 0.6353655687923382,
956
+ "eval_f1_surprise": 0.48747591522157996,
957
+ "eval_loss": 0.7228219509124756,
958
+ "eval_micro_f1": 0.737557721908671,
959
+ "eval_prec": 0.7374470363297007,
960
+ "eval_prec_anger": 0.6206030150753769,
961
+ "eval_prec_disgust": 0.40559440559440557,
962
+ "eval_prec_fear": 0.5199049316696376,
963
+ "eval_prec_joy": 0.7209614810391162,
964
+ "eval_prec_neutral": 0.846042471042471,
965
+ "eval_prec_sadness": 0.6598738476467734,
966
+ "eval_prec_surprise": 0.5254413291796469,
967
+ "eval_recall": 0.7375577219086711,
968
+ "eval_recall_anger": 0.5533980582524272,
969
+ "eval_recall_disgust": 0.3625,
970
+ "eval_recall_fear": 0.5630630630630631,
971
+ "eval_recall_joy": 0.8057734023026865,
972
+ "eval_recall_neutral": 0.8200018710824212,
973
+ "eval_recall_sadness": 0.6126126126126126,
974
+ "eval_recall_surprise": 0.4546271338724169,
975
+ "eval_runtime": 52.2209,
976
+ "eval_samples_per_second": 447.867,
977
+ "eval_steps_per_second": 13.998,
978
+ "step": 30735
979
+ },
980
+ {
981
+ "epoch": 18.76,
982
+ "learning_rate": 0.00026306324821456095,
983
+ "loss": 0.5816,
984
+ "step": 30750
985
+ },
986
+ {
987
+ "epoch": 19.01,
988
+ "learning_rate": 0.0002525304423778452,
989
+ "loss": 0.5868,
990
+ "step": 31160
991
+ },
992
+ {
993
+ "epoch": 19.26,
994
+ "learning_rate": 0.00024199763654112933,
995
+ "loss": 0.5698,
996
+ "step": 31570
997
+ },
998
+ {
999
+ "epoch": 19.51,
1000
+ "learning_rate": 0.0002314648307044135,
1001
+ "loss": 0.5696,
1002
+ "step": 31980
1003
+ },
1004
+ {
1005
+ "epoch": 19.76,
1006
+ "learning_rate": 0.0002209320248676977,
1007
+ "loss": 0.5728,
1008
+ "step": 32390
1009
+ },
1010
+ {
1011
+ "epoch": 20.0,
1012
+ "eval_accuracy": 0.7344364631434924,
1013
+ "eval_b_acc": 0.6090238361567221,
1014
+ "eval_f1": 0.7349132691412003,
1015
+ "eval_f1_anger": 0.5850591715976332,
1016
+ "eval_f1_disgust": 0.4091816367265469,
1017
+ "eval_f1_fear": 0.5293729372937294,
1018
+ "eval_f1_joy": 0.760131084645512,
1019
+ "eval_f1_neutral": 0.8284622731614136,
1020
+ "eval_f1_sadness": 0.6423658872077029,
1021
+ "eval_f1_surprise": 0.49303996407723394,
1022
+ "eval_loss": 0.7313103675842285,
1023
+ "eval_micro_f1": 0.7344364631434924,
1024
+ "eval_prec": 0.7365301337731266,
1025
+ "eval_prec_anger": 0.5794871794871795,
1026
+ "eval_prec_disgust": 0.39272030651340994,
1027
+ "eval_prec_fear": 0.5433604336043361,
1028
+ "eval_prec_joy": 0.7295182571340902,
1029
+ "eval_prec_neutral": 0.8461613501121842,
1030
+ "eval_prec_sadness": 0.65406162464986,
1031
+ "eval_prec_surprise": 0.49281867145421904,
1032
+ "eval_recall": 0.7344364631434924,
1033
+ "eval_recall_anger": 0.590739357729649,
1034
+ "eval_recall_disgust": 0.4270833333333333,
1035
+ "eval_recall_fear": 0.5160875160875161,
1036
+ "eval_recall_joy": 0.7934256632738195,
1037
+ "eval_recall_neutral": 0.8114884460660492,
1038
+ "eval_recall_sadness": 0.6310810810810811,
1039
+ "eval_recall_surprise": 0.49326145552560646,
1040
+ "eval_runtime": 53.2758,
1041
+ "eval_samples_per_second": 438.999,
1042
+ "eval_steps_per_second": 13.721,
1043
+ "step": 32784
1044
+ },
1045
+ {
1046
+ "epoch": 20.01,
1047
+ "learning_rate": 0.00021039921903098186,
1048
+ "loss": 0.5687,
1049
+ "step": 32800
1050
+ },
1051
+ {
1052
+ "epoch": 20.26,
1053
+ "learning_rate": 0.0001998921029645995,
1054
+ "loss": 0.5589,
1055
+ "step": 33210
1056
+ },
1057
+ {
1058
+ "epoch": 20.51,
1059
+ "learning_rate": 0.00018935929712788369,
1060
+ "loss": 0.5561,
1061
+ "step": 33620
1062
+ },
1063
+ {
1064
+ "epoch": 20.76,
1065
+ "learning_rate": 0.00017882649129116783,
1066
+ "loss": 0.5543,
1067
+ "step": 34030
1068
+ },
1069
+ {
1070
+ "epoch": 21.01,
1071
+ "learning_rate": 0.00016829368545445204,
1072
+ "loss": 0.5562,
1073
+ "step": 34440
1074
+ },
1075
+ {
1076
+ "epoch": 21.25,
1077
+ "eval_accuracy": 0.7375577219086711,
1078
+ "eval_b_acc": 0.5994500729257896,
1079
+ "eval_f1": 0.7366274505530668,
1080
+ "eval_f1_anger": 0.5735234215885947,
1081
+ "eval_f1_disgust": 0.4092276830491474,
1082
+ "eval_f1_fear": 0.5332476710568583,
1083
+ "eval_f1_joy": 0.764285140239492,
1084
+ "eval_f1_neutral": 0.8349395313681028,
1085
+ "eval_f1_sadness": 0.625414364640884,
1086
+ "eval_f1_surprise": 0.4867469879518072,
1087
+ "eval_loss": 0.7413733005523682,
1088
+ "eval_micro_f1": 0.737557721908671,
1089
+ "eval_prec": 0.7371596076951062,
1090
+ "eval_prec_anger": 0.6308243727598566,
1091
+ "eval_prec_disgust": 0.3945841392649903,
1092
+ "eval_prec_fear": 0.5323925593329057,
1093
+ "eval_prec_joy": 0.7372093023255814,
1094
+ "eval_prec_neutral": 0.8433056589369214,
1095
+ "eval_prec_sadness": 0.613882863340564,
1096
+ "eval_prec_surprise": 0.524948024948025,
1097
+ "eval_recall": 0.7375577219086711,
1098
+ "eval_recall_anger": 0.5257654966392831,
1099
+ "eval_recall_disgust": 0.425,
1100
+ "eval_recall_fear": 0.5341055341055341,
1101
+ "eval_recall_joy": 0.7934256632738195,
1102
+ "eval_recall_neutral": 0.8267377677986716,
1103
+ "eval_recall_sadness": 0.6373873873873874,
1104
+ "eval_recall_surprise": 0.4537286612758311,
1105
+ "eval_runtime": 56.258,
1106
+ "eval_samples_per_second": 415.728,
1107
+ "eval_steps_per_second": 12.994,
1108
+ "step": 34833
1109
+ },
1110
+ {
1111
+ "epoch": 21.26,
1112
+ "learning_rate": 0.00015776087961773621,
1113
+ "loss": 0.5488,
1114
+ "step": 34850
1115
+ },
1116
+ {
1117
+ "epoch": 21.51,
1118
+ "learning_rate": 0.00014722807378102042,
1119
+ "loss": 0.5439,
1120
+ "step": 35260
1121
+ },
1122
+ {
1123
+ "epoch": 21.76,
1124
+ "learning_rate": 0.0001366952679443046,
1125
+ "loss": 0.5457,
1126
+ "step": 35670
1127
+ },
1128
+ {
1129
+ "epoch": 22.01,
1130
+ "learning_rate": 0.00012616246210758874,
1131
+ "loss": 0.5452,
1132
+ "step": 36080
1133
+ },
1134
+ {
1135
+ "epoch": 22.26,
1136
+ "learning_rate": 0.00011562965627087293,
1137
+ "loss": 0.5348,
1138
+ "step": 36490
1139
+ },
1140
+ {
1141
+ "epoch": 22.5,
1142
+ "eval_accuracy": 0.737001881306653,
1143
+ "eval_b_acc": 0.6016861275031393,
1144
+ "eval_f1": 0.7364691206284374,
1145
+ "eval_f1_anger": 0.5868352547415396,
1146
+ "eval_f1_disgust": 0.39747634069400634,
1147
+ "eval_f1_fear": 0.5283998701720221,
1148
+ "eval_f1_joy": 0.763428933607986,
1149
+ "eval_f1_neutral": 0.8322205265165968,
1150
+ "eval_f1_sadness": 0.6372369624885635,
1151
+ "eval_f1_surprise": 0.48638318203535597,
1152
+ "eval_loss": 0.7397785782814026,
1153
+ "eval_micro_f1": 0.7370018813066531,
1154
+ "eval_prec": 0.73744814726634,
1155
+ "eval_prec_anger": 0.5844444444444444,
1156
+ "eval_prec_disgust": 0.4012738853503185,
1157
+ "eval_prec_fear": 0.5330713817943681,
1158
+ "eval_prec_joy": 0.7268064564791069,
1159
+ "eval_prec_neutral": 0.8488179784025683,
1160
+ "eval_prec_sadness": 0.6473048327137546,
1161
+ "eval_prec_surprise": 0.5193877551020408,
1162
+ "eval_recall": 0.737001881306653,
1163
+ "eval_recall_anger": 0.5892457057505601,
1164
+ "eval_recall_disgust": 0.39375,
1165
+ "eval_recall_fear": 0.5238095238095238,
1166
+ "eval_recall_joy": 0.8039379275821792,
1167
+ "eval_recall_neutral": 0.8162597062400598,
1168
+ "eval_recall_sadness": 0.6274774774774775,
1169
+ "eval_recall_surprise": 0.4573225516621743,
1170
+ "eval_runtime": 54.191,
1171
+ "eval_samples_per_second": 431.585,
1172
+ "eval_steps_per_second": 13.489,
1173
+ "step": 36882
1174
+ },
1175
+ {
1176
+ "epoch": 22.51,
1177
+ "learning_rate": 0.00010509685043415712,
1178
+ "loss": 0.5275,
1179
+ "step": 36900
1180
+ },
1181
+ {
1182
+ "epoch": 22.76,
1183
+ "learning_rate": 9.456404459744131e-05,
1184
+ "loss": 0.5375,
1185
+ "step": 37310
1186
+ },
1187
+ {
1188
+ "epoch": 23.01,
1189
+ "learning_rate": 8.403123876072547e-05,
1190
+ "loss": 0.5249,
1191
+ "step": 37720
1192
+ },
1193
+ {
1194
+ "epoch": 23.26,
1195
+ "learning_rate": 7.352412269434311e-05,
1196
+ "loss": 0.5238,
1197
+ "step": 38130
1198
+ },
1199
+ {
1200
+ "epoch": 23.51,
1201
+ "learning_rate": 6.29913168576273e-05,
1202
+ "loss": 0.5202,
1203
+ "step": 38540
1204
+ },
1205
+ {
1206
+ "epoch": 23.75,
1207
+ "eval_accuracy": 0.7389259449290234,
1208
+ "eval_b_acc": 0.6013150721994661,
1209
+ "eval_f1": 0.738139896893294,
1210
+ "eval_f1_anger": 0.5870646766169154,
1211
+ "eval_f1_disgust": 0.3953488372093023,
1212
+ "eval_f1_fear": 0.5318255250403877,
1213
+ "eval_f1_joy": 0.7646298092466861,
1214
+ "eval_f1_neutral": 0.8346932037919161,
1215
+ "eval_f1_sadness": 0.6373725934314837,
1216
+ "eval_f1_surprise": 0.4868667917448405,
1217
+ "eval_loss": 0.7423349618911743,
1218
+ "eval_micro_f1": 0.7389259449290234,
1219
+ "eval_prec": 0.7379281962820202,
1220
+ "eval_prec_anger": 0.6020408163265306,
1221
+ "eval_prec_disgust": 0.4012875536480687,
1222
+ "eval_prec_fear": 0.5340687865022713,
1223
+ "eval_prec_joy": 0.7414955322150807,
1224
+ "eval_prec_neutral": 0.8416397184706106,
1225
+ "eval_prec_sadness": 0.6410022779043281,
1226
+ "eval_prec_surprise": 0.5093228655544652,
1227
+ "eval_recall": 0.7389259449290234,
1228
+ "eval_recall_anger": 0.5728155339805825,
1229
+ "eval_recall_disgust": 0.38958333333333334,
1230
+ "eval_recall_fear": 0.5296010296010296,
1231
+ "eval_recall_joy": 0.7892541298181212,
1232
+ "eval_recall_neutral": 0.8278604172513799,
1233
+ "eval_recall_sadness": 0.6337837837837837,
1234
+ "eval_recall_surprise": 0.46630727762803237,
1235
+ "eval_runtime": 54.5174,
1236
+ "eval_samples_per_second": 429.001,
1237
+ "eval_steps_per_second": 13.409,
1238
+ "step": 38931
1239
+ },
1240
+ {
1241
+ "epoch": 23.76,
1242
+ "learning_rate": 5.245851102091148e-05,
1243
+ "loss": 0.5257,
1244
+ "step": 38950
1245
+ },
1246
+ {
1247
+ "epoch": 24.01,
1248
+ "learning_rate": 4.1925705184195654e-05,
1249
+ "loss": 0.5132,
1250
+ "step": 39360
1251
+ },
1252
+ {
1253
+ "epoch": 24.26,
1254
+ "learning_rate": 3.139289934747983e-05,
1255
+ "loss": 0.5149,
1256
+ "step": 39770
1257
+ },
1258
+ {
1259
+ "epoch": 24.51,
1260
+ "learning_rate": 2.0860093510764016e-05,
1261
+ "loss": 0.5047,
1262
+ "step": 40180
1263
+ },
1264
+ {
1265
+ "epoch": 24.76,
1266
+ "learning_rate": 1.0352977444381647e-05,
1267
+ "loss": 0.52,
1268
+ "step": 40590
1269
+ },
1270
+ {
1271
+ "epoch": 24.99,
1272
+ "step": 40975,
1273
+ "total_flos": 1.2293541171245507e+18,
1274
+ "train_loss": 0.6590935281284558,
1275
+ "train_runtime": 28744.9325,
1276
+ "train_samples_per_second": 182.51,
1277
+ "train_steps_per_second": 1.425
1278
+ }
1279
+ ],
1280
+ "logging_steps": 410,
1281
+ "max_steps": 40975,
1282
+ "num_train_epochs": 25,
1283
+ "save_steps": 4098,
1284
+ "total_flos": 1.2293541171245507e+18,
1285
+ "trial_name": null,
1286
+ "trial_params": null
1287
+ }