File size: 3,791 Bytes
614ef3c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
{
  "best_metric": null,
  "best_model_checkpoint": null,
  "epoch": 90.99029126213593,
  "eval_steps": 500,
  "global_step": 2343,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 3.15,
      "learning_rate": 0.000324,
      "loss": 1.6248,
      "step": 81
    },
    {
      "epoch": 6.29,
      "learning_rate": 0.000648,
      "loss": 1.5109,
      "step": 162
    },
    {
      "epoch": 9.44,
      "learning_rate": 0.000972,
      "loss": 1.4155,
      "step": 243
    },
    {
      "epoch": 12.58,
      "learning_rate": 0.0009671111111111112,
      "loss": 1.328,
      "step": 324
    },
    {
      "epoch": 15.73,
      "learning_rate": 0.0009311111111111112,
      "loss": 1.2665,
      "step": 405
    },
    {
      "epoch": 18.87,
      "learning_rate": 0.0008951111111111111,
      "loss": 1.2178,
      "step": 486
    },
    {
      "epoch": 22.02,
      "learning_rate": 0.0008591111111111112,
      "loss": 1.1829,
      "step": 567
    },
    {
      "epoch": 25.17,
      "learning_rate": 0.0008231111111111112,
      "loss": 1.1523,
      "step": 648
    },
    {
      "epoch": 28.31,
      "learning_rate": 0.0007871111111111111,
      "loss": 1.1296,
      "step": 729
    },
    {
      "epoch": 31.46,
      "learning_rate": 0.000751111111111111,
      "loss": 1.1084,
      "step": 810
    },
    {
      "epoch": 34.6,
      "learning_rate": 0.0007151111111111111,
      "loss": 1.0855,
      "step": 891
    },
    {
      "epoch": 37.75,
      "learning_rate": 0.0006791111111111111,
      "loss": 1.0708,
      "step": 972
    },
    {
      "epoch": 40.89,
      "learning_rate": 0.0006431111111111111,
      "loss": 1.0536,
      "step": 1053
    },
    {
      "epoch": 44.04,
      "learning_rate": 0.0006071111111111112,
      "loss": 1.0359,
      "step": 1134
    },
    {
      "epoch": 47.18,
      "learning_rate": 0.0005711111111111111,
      "loss": 1.0246,
      "step": 1215
    },
    {
      "epoch": 50.33,
      "learning_rate": 0.0005351111111111111,
      "loss": 1.0132,
      "step": 1296
    },
    {
      "epoch": 53.48,
      "learning_rate": 0.0004991111111111111,
      "loss": 1.0013,
      "step": 1377
    },
    {
      "epoch": 56.62,
      "learning_rate": 0.0004631111111111111,
      "loss": 0.9878,
      "step": 1458
    },
    {
      "epoch": 59.77,
      "learning_rate": 0.0004271111111111111,
      "loss": 0.9766,
      "step": 1539
    },
    {
      "epoch": 62.91,
      "learning_rate": 0.0003911111111111111,
      "loss": 0.9643,
      "step": 1620
    },
    {
      "epoch": 66.06,
      "learning_rate": 0.0003551111111111111,
      "loss": 0.9538,
      "step": 1701
    },
    {
      "epoch": 69.2,
      "learning_rate": 0.0003191111111111111,
      "loss": 0.9486,
      "step": 1782
    },
    {
      "epoch": 72.35,
      "learning_rate": 0.0002831111111111111,
      "loss": 0.9382,
      "step": 1863
    },
    {
      "epoch": 75.5,
      "learning_rate": 0.00024711111111111114,
      "loss": 0.9255,
      "step": 1944
    },
    {
      "epoch": 78.64,
      "learning_rate": 0.0002111111111111111,
      "loss": 0.9153,
      "step": 2025
    },
    {
      "epoch": 81.79,
      "learning_rate": 0.0001751111111111111,
      "loss": 0.9069,
      "step": 2106
    },
    {
      "epoch": 84.93,
      "learning_rate": 0.0001391111111111111,
      "loss": 0.8996,
      "step": 2187
    },
    {
      "epoch": 88.08,
      "learning_rate": 0.00010311111111111111,
      "loss": 0.888,
      "step": 2268
    }
  ],
  "logging_steps": 81,
  "max_steps": 2500,
  "num_train_epochs": 100,
  "save_steps": 500,
  "total_flos": 2.6232265876392903e+19,
  "trial_name": null,
  "trial_params": null
}