abdiharyadi
commited on
Training in progress, epoch 15
Browse files- last-checkpoint/model.safetensors +1 -1
- last-checkpoint/optimizer.pt +1 -1
- last-checkpoint/rng_state.pth +1 -1
- last-checkpoint/scheduler.pt +1 -1
- last-checkpoint/trainer_state.json +5 -3515
- model.safetensors +1 -1
- run-2024-10-28T15:22:08+00:00.log +0 -0
- val_outputs/val_generated_predictions_14521373.txt +0 -0
last-checkpoint/model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1575259780
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b35a3bc2ac180ed070b42029c9a9dd327a1a9559e81df276f329c07eb21d04fc
|
3 |
size 1575259780
|
last-checkpoint/optimizer.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3150397656
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ba79276529acf6b7fcac21ab5be5fc5756c900e697415161ce71f7759f9fa8e0
|
3 |
size 3150397656
|
last-checkpoint/rng_state.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14244
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2e52ca2f5b1048c2984d9cb01ff8bc5c06ec7e6e1ac850eb54ef8fe7147dcf65
|
3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1064
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:df05e968bcb6cb9f8c607bdaf90fbac1131121f0efd29f2e5e7bc42c79c2d577
|
3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
{
|
2 |
-
"best_metric":
|
3 |
-
"best_model_checkpoint": "/kaggle/tmp/amr-tst-indo/AMRBART-id/fine-tune/../outputs/mbart-en-id-smaller-fted/checkpoint-
|
4 |
-
"epoch":
|
5 |
"eval_steps": 500,
|
6 |
-
"global_step":
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
@@ -12891,3516 +12891,6 @@
|
|
12891 |
"eval_samples_per_second": 2.165,
|
12892 |
"eval_steps_per_second": 1.083,
|
12893 |
"step": 42562
|
12894 |
-
},
|
12895 |
-
{
|
12896 |
-
"epoch": 11.0054270577594,
|
12897 |
-
"learning_rate": 1.3131725657980034e-07,
|
12898 |
-
"loss": 2.4441,
|
12899 |
-
"step": 42580
|
12900 |
-
},
|
12901 |
-
{
|
12902 |
-
"epoch": 11.010595684196925,
|
12903 |
-
"learning_rate": 1.31284843770258e-07,
|
12904 |
-
"loss": 2.5027,
|
12905 |
-
"step": 42600
|
12906 |
-
},
|
12907 |
-
{
|
12908 |
-
"epoch": 11.015764310634449,
|
12909 |
-
"learning_rate": 1.3125243096071566e-07,
|
12910 |
-
"loss": 2.4709,
|
12911 |
-
"step": 42620
|
12912 |
-
},
|
12913 |
-
{
|
12914 |
-
"epoch": 11.020932937071974,
|
12915 |
-
"learning_rate": 1.3122001815117335e-07,
|
12916 |
-
"loss": 2.4576,
|
12917 |
-
"step": 42640
|
12918 |
-
},
|
12919 |
-
{
|
12920 |
-
"epoch": 11.026101563509497,
|
12921 |
-
"learning_rate": 1.31187605341631e-07,
|
12922 |
-
"loss": 2.4395,
|
12923 |
-
"step": 42660
|
12924 |
-
},
|
12925 |
-
{
|
12926 |
-
"epoch": 11.031270189947021,
|
12927 |
-
"learning_rate": 1.3115519253208867e-07,
|
12928 |
-
"loss": 2.504,
|
12929 |
-
"step": 42680
|
12930 |
-
},
|
12931 |
-
{
|
12932 |
-
"epoch": 11.036438816384546,
|
12933 |
-
"learning_rate": 1.3112277972254634e-07,
|
12934 |
-
"loss": 2.5521,
|
12935 |
-
"step": 42700
|
12936 |
-
},
|
12937 |
-
{
|
12938 |
-
"epoch": 11.04160744282207,
|
12939 |
-
"learning_rate": 1.31090366913004e-07,
|
12940 |
-
"loss": 2.5074,
|
12941 |
-
"step": 42720
|
12942 |
-
},
|
12943 |
-
{
|
12944 |
-
"epoch": 11.046776069259595,
|
12945 |
-
"learning_rate": 1.3105795410346169e-07,
|
12946 |
-
"loss": 2.5079,
|
12947 |
-
"step": 42740
|
12948 |
-
},
|
12949 |
-
{
|
12950 |
-
"epoch": 11.051944695697118,
|
12951 |
-
"learning_rate": 1.3102554129391935e-07,
|
12952 |
-
"loss": 2.5337,
|
12953 |
-
"step": 42760
|
12954 |
-
},
|
12955 |
-
{
|
12956 |
-
"epoch": 11.057113322134644,
|
12957 |
-
"learning_rate": 1.30993128484377e-07,
|
12958 |
-
"loss": 2.4815,
|
12959 |
-
"step": 42780
|
12960 |
-
},
|
12961 |
-
{
|
12962 |
-
"epoch": 11.062281948572167,
|
12963 |
-
"learning_rate": 1.309607156748347e-07,
|
12964 |
-
"loss": 2.4897,
|
12965 |
-
"step": 42800
|
12966 |
-
},
|
12967 |
-
{
|
12968 |
-
"epoch": 11.06745057500969,
|
12969 |
-
"learning_rate": 1.3092830286529236e-07,
|
12970 |
-
"loss": 2.5184,
|
12971 |
-
"step": 42820
|
12972 |
-
},
|
12973 |
-
{
|
12974 |
-
"epoch": 11.072619201447216,
|
12975 |
-
"learning_rate": 1.3089589005575002e-07,
|
12976 |
-
"loss": 2.459,
|
12977 |
-
"step": 42840
|
12978 |
-
},
|
12979 |
-
{
|
12980 |
-
"epoch": 11.07778782788474,
|
12981 |
-
"learning_rate": 1.308634772462077e-07,
|
12982 |
-
"loss": 2.4623,
|
12983 |
-
"step": 42860
|
12984 |
-
},
|
12985 |
-
{
|
12986 |
-
"epoch": 11.082956454322265,
|
12987 |
-
"learning_rate": 1.3083106443666535e-07,
|
12988 |
-
"loss": 2.4422,
|
12989 |
-
"step": 42880
|
12990 |
-
},
|
12991 |
-
{
|
12992 |
-
"epoch": 11.088125080759788,
|
12993 |
-
"learning_rate": 1.3079865162712303e-07,
|
12994 |
-
"loss": 2.5212,
|
12995 |
-
"step": 42900
|
12996 |
-
},
|
12997 |
-
{
|
12998 |
-
"epoch": 11.093293707197311,
|
12999 |
-
"learning_rate": 1.307662388175807e-07,
|
13000 |
-
"loss": 2.4923,
|
13001 |
-
"step": 42920
|
13002 |
-
},
|
13003 |
-
{
|
13004 |
-
"epoch": 11.098462333634837,
|
13005 |
-
"learning_rate": 1.3073382600803836e-07,
|
13006 |
-
"loss": 2.5076,
|
13007 |
-
"step": 42940
|
13008 |
-
},
|
13009 |
-
{
|
13010 |
-
"epoch": 11.10363096007236,
|
13011 |
-
"learning_rate": 1.3070141319849605e-07,
|
13012 |
-
"loss": 2.5291,
|
13013 |
-
"step": 42960
|
13014 |
-
},
|
13015 |
-
{
|
13016 |
-
"epoch": 11.108799586509885,
|
13017 |
-
"learning_rate": 1.306690003889537e-07,
|
13018 |
-
"loss": 2.4459,
|
13019 |
-
"step": 42980
|
13020 |
-
},
|
13021 |
-
{
|
13022 |
-
"epoch": 11.113968212947409,
|
13023 |
-
"learning_rate": 1.3063658757941137e-07,
|
13024 |
-
"loss": 2.4522,
|
13025 |
-
"step": 43000
|
13026 |
-
},
|
13027 |
-
{
|
13028 |
-
"epoch": 11.119136839384934,
|
13029 |
-
"learning_rate": 1.3060417476986906e-07,
|
13030 |
-
"loss": 2.4752,
|
13031 |
-
"step": 43020
|
13032 |
-
},
|
13033 |
-
{
|
13034 |
-
"epoch": 11.124305465822458,
|
13035 |
-
"learning_rate": 1.3057176196032672e-07,
|
13036 |
-
"loss": 2.4877,
|
13037 |
-
"step": 43040
|
13038 |
-
},
|
13039 |
-
{
|
13040 |
-
"epoch": 11.129474092259981,
|
13041 |
-
"learning_rate": 1.3053934915078438e-07,
|
13042 |
-
"loss": 2.4969,
|
13043 |
-
"step": 43060
|
13044 |
-
},
|
13045 |
-
{
|
13046 |
-
"epoch": 11.134642718697506,
|
13047 |
-
"learning_rate": 1.3050693634124207e-07,
|
13048 |
-
"loss": 2.5331,
|
13049 |
-
"step": 43080
|
13050 |
-
},
|
13051 |
-
{
|
13052 |
-
"epoch": 11.13981134513503,
|
13053 |
-
"learning_rate": 1.304745235316997e-07,
|
13054 |
-
"loss": 2.4214,
|
13055 |
-
"step": 43100
|
13056 |
-
},
|
13057 |
-
{
|
13058 |
-
"epoch": 11.144979971572555,
|
13059 |
-
"learning_rate": 1.304421107221574e-07,
|
13060 |
-
"loss": 2.4795,
|
13061 |
-
"step": 43120
|
13062 |
-
},
|
13063 |
-
{
|
13064 |
-
"epoch": 11.150148598010079,
|
13065 |
-
"learning_rate": 1.3040969791261506e-07,
|
13066 |
-
"loss": 2.491,
|
13067 |
-
"step": 43140
|
13068 |
-
},
|
13069 |
-
{
|
13070 |
-
"epoch": 11.155317224447604,
|
13071 |
-
"learning_rate": 1.3037728510307272e-07,
|
13072 |
-
"loss": 2.4823,
|
13073 |
-
"step": 43160
|
13074 |
-
},
|
13075 |
-
{
|
13076 |
-
"epoch": 11.160485850885127,
|
13077 |
-
"learning_rate": 1.303448722935304e-07,
|
13078 |
-
"loss": 2.5129,
|
13079 |
-
"step": 43180
|
13080 |
-
},
|
13081 |
-
{
|
13082 |
-
"epoch": 11.16565447732265,
|
13083 |
-
"learning_rate": 1.3031245948398807e-07,
|
13084 |
-
"loss": 2.4543,
|
13085 |
-
"step": 43200
|
13086 |
-
},
|
13087 |
-
{
|
13088 |
-
"epoch": 11.170823103760176,
|
13089 |
-
"learning_rate": 1.3028004667444573e-07,
|
13090 |
-
"loss": 2.4735,
|
13091 |
-
"step": 43220
|
13092 |
-
},
|
13093 |
-
{
|
13094 |
-
"epoch": 11.1759917301977,
|
13095 |
-
"learning_rate": 1.3024763386490342e-07,
|
13096 |
-
"loss": 2.5127,
|
13097 |
-
"step": 43240
|
13098 |
-
},
|
13099 |
-
{
|
13100 |
-
"epoch": 11.181160356635225,
|
13101 |
-
"learning_rate": 1.3021522105536105e-07,
|
13102 |
-
"loss": 2.4045,
|
13103 |
-
"step": 43260
|
13104 |
-
},
|
13105 |
-
{
|
13106 |
-
"epoch": 11.186328983072748,
|
13107 |
-
"learning_rate": 1.3018280824581874e-07,
|
13108 |
-
"loss": 2.4599,
|
13109 |
-
"step": 43280
|
13110 |
-
},
|
13111 |
-
{
|
13112 |
-
"epoch": 11.191497609510273,
|
13113 |
-
"learning_rate": 1.301503954362764e-07,
|
13114 |
-
"loss": 2.4875,
|
13115 |
-
"step": 43300
|
13116 |
-
},
|
13117 |
-
{
|
13118 |
-
"epoch": 11.196666235947797,
|
13119 |
-
"learning_rate": 1.3011798262673407e-07,
|
13120 |
-
"loss": 2.4523,
|
13121 |
-
"step": 43320
|
13122 |
-
},
|
13123 |
-
{
|
13124 |
-
"epoch": 11.20183486238532,
|
13125 |
-
"learning_rate": 1.3008556981719175e-07,
|
13126 |
-
"loss": 2.47,
|
13127 |
-
"step": 43340
|
13128 |
-
},
|
13129 |
-
{
|
13130 |
-
"epoch": 11.207003488822846,
|
13131 |
-
"learning_rate": 1.3005315700764942e-07,
|
13132 |
-
"loss": 2.464,
|
13133 |
-
"step": 43360
|
13134 |
-
},
|
13135 |
-
{
|
13136 |
-
"epoch": 11.212172115260369,
|
13137 |
-
"learning_rate": 1.3002074419810708e-07,
|
13138 |
-
"loss": 2.5223,
|
13139 |
-
"step": 43380
|
13140 |
-
},
|
13141 |
-
{
|
13142 |
-
"epoch": 11.217340741697894,
|
13143 |
-
"learning_rate": 1.2998833138856477e-07,
|
13144 |
-
"loss": 2.4749,
|
13145 |
-
"step": 43400
|
13146 |
-
},
|
13147 |
-
{
|
13148 |
-
"epoch": 11.222509368135418,
|
13149 |
-
"learning_rate": 1.2995591857902243e-07,
|
13150 |
-
"loss": 2.5268,
|
13151 |
-
"step": 43420
|
13152 |
-
},
|
13153 |
-
{
|
13154 |
-
"epoch": 11.227677994572943,
|
13155 |
-
"learning_rate": 1.299235057694801e-07,
|
13156 |
-
"loss": 2.4837,
|
13157 |
-
"step": 43440
|
13158 |
-
},
|
13159 |
-
{
|
13160 |
-
"epoch": 11.232846621010466,
|
13161 |
-
"learning_rate": 1.2989109295993778e-07,
|
13162 |
-
"loss": 2.4285,
|
13163 |
-
"step": 43460
|
13164 |
-
},
|
13165 |
-
{
|
13166 |
-
"epoch": 11.23801524744799,
|
13167 |
-
"learning_rate": 1.2985868015039541e-07,
|
13168 |
-
"loss": 2.4918,
|
13169 |
-
"step": 43480
|
13170 |
-
},
|
13171 |
-
{
|
13172 |
-
"epoch": 11.243183873885515,
|
13173 |
-
"learning_rate": 1.298262673408531e-07,
|
13174 |
-
"loss": 2.4585,
|
13175 |
-
"step": 43500
|
13176 |
-
},
|
13177 |
-
{
|
13178 |
-
"epoch": 11.248352500323039,
|
13179 |
-
"learning_rate": 1.2979385453131076e-07,
|
13180 |
-
"loss": 2.5156,
|
13181 |
-
"step": 43520
|
13182 |
-
},
|
13183 |
-
{
|
13184 |
-
"epoch": 11.253521126760564,
|
13185 |
-
"learning_rate": 1.2976144172176843e-07,
|
13186 |
-
"loss": 2.5031,
|
13187 |
-
"step": 43540
|
13188 |
-
},
|
13189 |
-
{
|
13190 |
-
"epoch": 11.258689753198087,
|
13191 |
-
"learning_rate": 1.2972902891222611e-07,
|
13192 |
-
"loss": 2.456,
|
13193 |
-
"step": 43560
|
13194 |
-
},
|
13195 |
-
{
|
13196 |
-
"epoch": 11.263858379635613,
|
13197 |
-
"learning_rate": 1.2969661610268378e-07,
|
13198 |
-
"loss": 2.5083,
|
13199 |
-
"step": 43580
|
13200 |
-
},
|
13201 |
-
{
|
13202 |
-
"epoch": 11.269027006073136,
|
13203 |
-
"learning_rate": 1.2966420329314144e-07,
|
13204 |
-
"loss": 2.5241,
|
13205 |
-
"step": 43600
|
13206 |
-
},
|
13207 |
-
{
|
13208 |
-
"epoch": 11.27419563251066,
|
13209 |
-
"learning_rate": 1.2963179048359913e-07,
|
13210 |
-
"loss": 2.4751,
|
13211 |
-
"step": 43620
|
13212 |
-
},
|
13213 |
-
{
|
13214 |
-
"epoch": 11.279364258948185,
|
13215 |
-
"learning_rate": 1.295993776740568e-07,
|
13216 |
-
"loss": 2.4695,
|
13217 |
-
"step": 43640
|
13218 |
-
},
|
13219 |
-
{
|
13220 |
-
"epoch": 11.284532885385708,
|
13221 |
-
"learning_rate": 1.2956696486451445e-07,
|
13222 |
-
"loss": 2.5578,
|
13223 |
-
"step": 43660
|
13224 |
-
},
|
13225 |
-
{
|
13226 |
-
"epoch": 11.289701511823234,
|
13227 |
-
"learning_rate": 1.2953455205497214e-07,
|
13228 |
-
"loss": 2.4649,
|
13229 |
-
"step": 43680
|
13230 |
-
},
|
13231 |
-
{
|
13232 |
-
"epoch": 11.294870138260757,
|
13233 |
-
"learning_rate": 1.2950213924542977e-07,
|
13234 |
-
"loss": 2.4747,
|
13235 |
-
"step": 43700
|
13236 |
-
},
|
13237 |
-
{
|
13238 |
-
"epoch": 11.300038764698282,
|
13239 |
-
"learning_rate": 1.2946972643588746e-07,
|
13240 |
-
"loss": 2.4679,
|
13241 |
-
"step": 43720
|
13242 |
-
},
|
13243 |
-
{
|
13244 |
-
"epoch": 11.305207391135806,
|
13245 |
-
"learning_rate": 1.2943731362634512e-07,
|
13246 |
-
"loss": 2.4683,
|
13247 |
-
"step": 43740
|
13248 |
-
},
|
13249 |
-
{
|
13250 |
-
"epoch": 11.31037601757333,
|
13251 |
-
"learning_rate": 1.2940490081680279e-07,
|
13252 |
-
"loss": 2.4857,
|
13253 |
-
"step": 43760
|
13254 |
-
},
|
13255 |
-
{
|
13256 |
-
"epoch": 11.315544644010854,
|
13257 |
-
"learning_rate": 1.2937248800726047e-07,
|
13258 |
-
"loss": 2.426,
|
13259 |
-
"step": 43780
|
13260 |
-
},
|
13261 |
-
{
|
13262 |
-
"epoch": 11.320713270448378,
|
13263 |
-
"learning_rate": 1.2934007519771814e-07,
|
13264 |
-
"loss": 2.4733,
|
13265 |
-
"step": 43800
|
13266 |
-
},
|
13267 |
-
{
|
13268 |
-
"epoch": 11.325881896885903,
|
13269 |
-
"learning_rate": 1.293076623881758e-07,
|
13270 |
-
"loss": 2.4812,
|
13271 |
-
"step": 43820
|
13272 |
-
},
|
13273 |
-
{
|
13274 |
-
"epoch": 11.331050523323427,
|
13275 |
-
"learning_rate": 1.2927524957863349e-07,
|
13276 |
-
"loss": 2.4528,
|
13277 |
-
"step": 43840
|
13278 |
-
},
|
13279 |
-
{
|
13280 |
-
"epoch": 11.33621914976095,
|
13281 |
-
"learning_rate": 1.2924283676909112e-07,
|
13282 |
-
"loss": 2.494,
|
13283 |
-
"step": 43860
|
13284 |
-
},
|
13285 |
-
{
|
13286 |
-
"epoch": 11.341387776198475,
|
13287 |
-
"learning_rate": 1.292104239595488e-07,
|
13288 |
-
"loss": 2.4795,
|
13289 |
-
"step": 43880
|
13290 |
-
},
|
13291 |
-
{
|
13292 |
-
"epoch": 11.346556402635999,
|
13293 |
-
"learning_rate": 1.2917801115000647e-07,
|
13294 |
-
"loss": 2.4816,
|
13295 |
-
"step": 43900
|
13296 |
-
},
|
13297 |
-
{
|
13298 |
-
"epoch": 11.351725029073524,
|
13299 |
-
"learning_rate": 1.2914559834046413e-07,
|
13300 |
-
"loss": 2.5282,
|
13301 |
-
"step": 43920
|
13302 |
-
},
|
13303 |
-
{
|
13304 |
-
"epoch": 11.356893655511048,
|
13305 |
-
"learning_rate": 1.2911318553092182e-07,
|
13306 |
-
"loss": 2.4849,
|
13307 |
-
"step": 43940
|
13308 |
-
},
|
13309 |
-
{
|
13310 |
-
"epoch": 11.362062281948573,
|
13311 |
-
"learning_rate": 1.2908077272137948e-07,
|
13312 |
-
"loss": 2.4246,
|
13313 |
-
"step": 43960
|
13314 |
-
},
|
13315 |
-
{
|
13316 |
-
"epoch": 11.367230908386096,
|
13317 |
-
"learning_rate": 1.2904835991183715e-07,
|
13318 |
-
"loss": 2.5491,
|
13319 |
-
"step": 43980
|
13320 |
-
},
|
13321 |
-
{
|
13322 |
-
"epoch": 11.372399534823622,
|
13323 |
-
"learning_rate": 1.2901594710229483e-07,
|
13324 |
-
"loss": 2.4584,
|
13325 |
-
"step": 44000
|
13326 |
-
},
|
13327 |
-
{
|
13328 |
-
"epoch": 11.377568161261145,
|
13329 |
-
"learning_rate": 1.289835342927525e-07,
|
13330 |
-
"loss": 2.5351,
|
13331 |
-
"step": 44020
|
13332 |
-
},
|
13333 |
-
{
|
13334 |
-
"epoch": 11.382736787698668,
|
13335 |
-
"learning_rate": 1.2895112148321016e-07,
|
13336 |
-
"loss": 2.5108,
|
13337 |
-
"step": 44040
|
13338 |
-
},
|
13339 |
-
{
|
13340 |
-
"epoch": 11.387905414136194,
|
13341 |
-
"learning_rate": 1.2891870867366782e-07,
|
13342 |
-
"loss": 2.5328,
|
13343 |
-
"step": 44060
|
13344 |
-
},
|
13345 |
-
{
|
13346 |
-
"epoch": 11.393074040573717,
|
13347 |
-
"learning_rate": 1.2888629586412548e-07,
|
13348 |
-
"loss": 2.4501,
|
13349 |
-
"step": 44080
|
13350 |
-
},
|
13351 |
-
{
|
13352 |
-
"epoch": 11.398242667011242,
|
13353 |
-
"learning_rate": 1.2885388305458317e-07,
|
13354 |
-
"loss": 2.434,
|
13355 |
-
"step": 44100
|
13356 |
-
},
|
13357 |
-
{
|
13358 |
-
"epoch": 11.403411293448766,
|
13359 |
-
"learning_rate": 1.2882147024504083e-07,
|
13360 |
-
"loss": 2.4713,
|
13361 |
-
"step": 44120
|
13362 |
-
},
|
13363 |
-
{
|
13364 |
-
"epoch": 11.40857991988629,
|
13365 |
-
"learning_rate": 1.287890574354985e-07,
|
13366 |
-
"loss": 2.5093,
|
13367 |
-
"step": 44140
|
13368 |
-
},
|
13369 |
-
{
|
13370 |
-
"epoch": 11.413748546323815,
|
13371 |
-
"learning_rate": 1.2875664462595618e-07,
|
13372 |
-
"loss": 2.5031,
|
13373 |
-
"step": 44160
|
13374 |
-
},
|
13375 |
-
{
|
13376 |
-
"epoch": 11.418917172761338,
|
13377 |
-
"learning_rate": 1.2872423181641384e-07,
|
13378 |
-
"loss": 2.4712,
|
13379 |
-
"step": 44180
|
13380 |
-
},
|
13381 |
-
{
|
13382 |
-
"epoch": 11.424085799198863,
|
13383 |
-
"learning_rate": 1.286918190068715e-07,
|
13384 |
-
"loss": 2.4963,
|
13385 |
-
"step": 44200
|
13386 |
-
},
|
13387 |
-
{
|
13388 |
-
"epoch": 11.429254425636387,
|
13389 |
-
"learning_rate": 1.2865940619732917e-07,
|
13390 |
-
"loss": 2.4866,
|
13391 |
-
"step": 44220
|
13392 |
-
},
|
13393 |
-
{
|
13394 |
-
"epoch": 11.434423052073912,
|
13395 |
-
"learning_rate": 1.2862699338778686e-07,
|
13396 |
-
"loss": 2.4665,
|
13397 |
-
"step": 44240
|
13398 |
-
},
|
13399 |
-
{
|
13400 |
-
"epoch": 11.439591678511436,
|
13401 |
-
"learning_rate": 1.2859458057824452e-07,
|
13402 |
-
"loss": 2.4525,
|
13403 |
-
"step": 44260
|
13404 |
-
},
|
13405 |
-
{
|
13406 |
-
"epoch": 11.444760304948959,
|
13407 |
-
"learning_rate": 1.2856216776870218e-07,
|
13408 |
-
"loss": 2.5271,
|
13409 |
-
"step": 44280
|
13410 |
-
},
|
13411 |
-
{
|
13412 |
-
"epoch": 11.449928931386484,
|
13413 |
-
"learning_rate": 1.2852975495915984e-07,
|
13414 |
-
"loss": 2.4676,
|
13415 |
-
"step": 44300
|
13416 |
-
},
|
13417 |
-
{
|
13418 |
-
"epoch": 11.455097557824008,
|
13419 |
-
"learning_rate": 1.2849734214961753e-07,
|
13420 |
-
"loss": 2.4996,
|
13421 |
-
"step": 44320
|
13422 |
-
},
|
13423 |
-
{
|
13424 |
-
"epoch": 11.460266184261533,
|
13425 |
-
"learning_rate": 1.284649293400752e-07,
|
13426 |
-
"loss": 2.4901,
|
13427 |
-
"step": 44340
|
13428 |
-
},
|
13429 |
-
{
|
13430 |
-
"epoch": 11.465434810699056,
|
13431 |
-
"learning_rate": 1.2843251653053285e-07,
|
13432 |
-
"loss": 2.411,
|
13433 |
-
"step": 44360
|
13434 |
-
},
|
13435 |
-
{
|
13436 |
-
"epoch": 11.470603437136582,
|
13437 |
-
"learning_rate": 1.2840010372099054e-07,
|
13438 |
-
"loss": 2.4688,
|
13439 |
-
"step": 44380
|
13440 |
-
},
|
13441 |
-
{
|
13442 |
-
"epoch": 11.475772063574105,
|
13443 |
-
"learning_rate": 1.283676909114482e-07,
|
13444 |
-
"loss": 2.4458,
|
13445 |
-
"step": 44400
|
13446 |
-
},
|
13447 |
-
{
|
13448 |
-
"epoch": 11.480940690011629,
|
13449 |
-
"learning_rate": 1.2833527810190587e-07,
|
13450 |
-
"loss": 2.4871,
|
13451 |
-
"step": 44420
|
13452 |
-
},
|
13453 |
-
{
|
13454 |
-
"epoch": 11.486109316449154,
|
13455 |
-
"learning_rate": 1.2830286529236353e-07,
|
13456 |
-
"loss": 2.4476,
|
13457 |
-
"step": 44440
|
13458 |
-
},
|
13459 |
-
{
|
13460 |
-
"epoch": 11.491277942886677,
|
13461 |
-
"learning_rate": 1.282704524828212e-07,
|
13462 |
-
"loss": 2.4647,
|
13463 |
-
"step": 44460
|
13464 |
-
},
|
13465 |
-
{
|
13466 |
-
"epoch": 11.496446569324203,
|
13467 |
-
"learning_rate": 1.2823803967327888e-07,
|
13468 |
-
"loss": 2.4887,
|
13469 |
-
"step": 44480
|
13470 |
-
},
|
13471 |
-
{
|
13472 |
-
"epoch": 11.501615195761726,
|
13473 |
-
"learning_rate": 1.2820562686373654e-07,
|
13474 |
-
"loss": 2.4835,
|
13475 |
-
"step": 44500
|
13476 |
-
},
|
13477 |
-
{
|
13478 |
-
"epoch": 11.506783822199251,
|
13479 |
-
"learning_rate": 1.281732140541942e-07,
|
13480 |
-
"loss": 2.4435,
|
13481 |
-
"step": 44520
|
13482 |
-
},
|
13483 |
-
{
|
13484 |
-
"epoch": 11.511952448636775,
|
13485 |
-
"learning_rate": 1.281408012446519e-07,
|
13486 |
-
"loss": 2.4716,
|
13487 |
-
"step": 44540
|
13488 |
-
},
|
13489 |
-
{
|
13490 |
-
"epoch": 11.517121075074298,
|
13491 |
-
"learning_rate": 1.2810838843510955e-07,
|
13492 |
-
"loss": 2.457,
|
13493 |
-
"step": 44560
|
13494 |
-
},
|
13495 |
-
{
|
13496 |
-
"epoch": 11.522289701511824,
|
13497 |
-
"learning_rate": 1.2807597562556721e-07,
|
13498 |
-
"loss": 2.4465,
|
13499 |
-
"step": 44580
|
13500 |
-
},
|
13501 |
-
{
|
13502 |
-
"epoch": 11.527458327949347,
|
13503 |
-
"learning_rate": 1.2804356281602488e-07,
|
13504 |
-
"loss": 2.4804,
|
13505 |
-
"step": 44600
|
13506 |
-
},
|
13507 |
-
{
|
13508 |
-
"epoch": 11.532626954386872,
|
13509 |
-
"learning_rate": 1.2801115000648256e-07,
|
13510 |
-
"loss": 2.5168,
|
13511 |
-
"step": 44620
|
13512 |
-
},
|
13513 |
-
{
|
13514 |
-
"epoch": 11.537795580824396,
|
13515 |
-
"learning_rate": 1.2797873719694023e-07,
|
13516 |
-
"loss": 2.4759,
|
13517 |
-
"step": 44640
|
13518 |
-
},
|
13519 |
-
{
|
13520 |
-
"epoch": 11.542964207261921,
|
13521 |
-
"learning_rate": 1.279463243873979e-07,
|
13522 |
-
"loss": 2.4929,
|
13523 |
-
"step": 44660
|
13524 |
-
},
|
13525 |
-
{
|
13526 |
-
"epoch": 11.548132833699444,
|
13527 |
-
"learning_rate": 1.2791391157785555e-07,
|
13528 |
-
"loss": 2.4136,
|
13529 |
-
"step": 44680
|
13530 |
-
},
|
13531 |
-
{
|
13532 |
-
"epoch": 11.553301460136968,
|
13533 |
-
"learning_rate": 1.2788149876831324e-07,
|
13534 |
-
"loss": 2.4497,
|
13535 |
-
"step": 44700
|
13536 |
-
},
|
13537 |
-
{
|
13538 |
-
"epoch": 11.558470086574493,
|
13539 |
-
"learning_rate": 1.278490859587709e-07,
|
13540 |
-
"loss": 2.4912,
|
13541 |
-
"step": 44720
|
13542 |
-
},
|
13543 |
-
{
|
13544 |
-
"epoch": 11.563638713012017,
|
13545 |
-
"learning_rate": 1.2781667314922856e-07,
|
13546 |
-
"loss": 2.5089,
|
13547 |
-
"step": 44740
|
13548 |
-
},
|
13549 |
-
{
|
13550 |
-
"epoch": 11.568807339449542,
|
13551 |
-
"learning_rate": 1.2778426033968622e-07,
|
13552 |
-
"loss": 2.4516,
|
13553 |
-
"step": 44760
|
13554 |
-
},
|
13555 |
-
{
|
13556 |
-
"epoch": 11.573975965887065,
|
13557 |
-
"learning_rate": 1.277518475301439e-07,
|
13558 |
-
"loss": 2.5023,
|
13559 |
-
"step": 44780
|
13560 |
-
},
|
13561 |
-
{
|
13562 |
-
"epoch": 11.579144592324589,
|
13563 |
-
"learning_rate": 1.2771943472060157e-07,
|
13564 |
-
"loss": 2.4883,
|
13565 |
-
"step": 44800
|
13566 |
-
},
|
13567 |
-
{
|
13568 |
-
"epoch": 11.584313218762114,
|
13569 |
-
"learning_rate": 1.2768702191105924e-07,
|
13570 |
-
"loss": 2.4885,
|
13571 |
-
"step": 44820
|
13572 |
-
},
|
13573 |
-
{
|
13574 |
-
"epoch": 11.589481845199638,
|
13575 |
-
"learning_rate": 1.2765460910151692e-07,
|
13576 |
-
"loss": 2.5421,
|
13577 |
-
"step": 44840
|
13578 |
-
},
|
13579 |
-
{
|
13580 |
-
"epoch": 11.594650471637163,
|
13581 |
-
"learning_rate": 1.2762219629197459e-07,
|
13582 |
-
"loss": 2.4938,
|
13583 |
-
"step": 44860
|
13584 |
-
},
|
13585 |
-
{
|
13586 |
-
"epoch": 11.599819098074686,
|
13587 |
-
"learning_rate": 1.2758978348243225e-07,
|
13588 |
-
"loss": 2.4712,
|
13589 |
-
"step": 44880
|
13590 |
-
},
|
13591 |
-
{
|
13592 |
-
"epoch": 11.604987724512211,
|
13593 |
-
"learning_rate": 1.275573706728899e-07,
|
13594 |
-
"loss": 2.5227,
|
13595 |
-
"step": 44900
|
13596 |
-
},
|
13597 |
-
{
|
13598 |
-
"epoch": 11.610156350949735,
|
13599 |
-
"learning_rate": 1.275249578633476e-07,
|
13600 |
-
"loss": 2.4463,
|
13601 |
-
"step": 44920
|
13602 |
-
},
|
13603 |
-
{
|
13604 |
-
"epoch": 11.61532497738726,
|
13605 |
-
"learning_rate": 1.2749254505380526e-07,
|
13606 |
-
"loss": 2.4526,
|
13607 |
-
"step": 44940
|
13608 |
-
},
|
13609 |
-
{
|
13610 |
-
"epoch": 11.620493603824784,
|
13611 |
-
"learning_rate": 1.2746013224426292e-07,
|
13612 |
-
"loss": 2.4763,
|
13613 |
-
"step": 44960
|
13614 |
-
},
|
13615 |
-
{
|
13616 |
-
"epoch": 11.625662230262307,
|
13617 |
-
"learning_rate": 1.2742771943472058e-07,
|
13618 |
-
"loss": 2.5575,
|
13619 |
-
"step": 44980
|
13620 |
-
},
|
13621 |
-
{
|
13622 |
-
"epoch": 11.630830856699832,
|
13623 |
-
"learning_rate": 1.2739530662517827e-07,
|
13624 |
-
"loss": 2.5079,
|
13625 |
-
"step": 45000
|
13626 |
-
},
|
13627 |
-
{
|
13628 |
-
"epoch": 11.635999483137356,
|
13629 |
-
"learning_rate": 1.2736289381563593e-07,
|
13630 |
-
"loss": 2.4868,
|
13631 |
-
"step": 45020
|
13632 |
-
},
|
13633 |
-
{
|
13634 |
-
"epoch": 11.641168109574881,
|
13635 |
-
"learning_rate": 1.273304810060936e-07,
|
13636 |
-
"loss": 2.4547,
|
13637 |
-
"step": 45040
|
13638 |
-
},
|
13639 |
-
{
|
13640 |
-
"epoch": 11.646336736012405,
|
13641 |
-
"learning_rate": 1.2729806819655126e-07,
|
13642 |
-
"loss": 2.4594,
|
13643 |
-
"step": 45060
|
13644 |
-
},
|
13645 |
-
{
|
13646 |
-
"epoch": 11.651505362449928,
|
13647 |
-
"learning_rate": 1.2726565538700895e-07,
|
13648 |
-
"loss": 2.495,
|
13649 |
-
"step": 45080
|
13650 |
-
},
|
13651 |
-
{
|
13652 |
-
"epoch": 11.656673988887453,
|
13653 |
-
"learning_rate": 1.272332425774666e-07,
|
13654 |
-
"loss": 2.5174,
|
13655 |
-
"step": 45100
|
13656 |
-
},
|
13657 |
-
{
|
13658 |
-
"epoch": 11.661842615324977,
|
13659 |
-
"learning_rate": 1.2720082976792427e-07,
|
13660 |
-
"loss": 2.4223,
|
13661 |
-
"step": 45120
|
13662 |
-
},
|
13663 |
-
{
|
13664 |
-
"epoch": 11.667011241762502,
|
13665 |
-
"learning_rate": 1.2716841695838196e-07,
|
13666 |
-
"loss": 2.4523,
|
13667 |
-
"step": 45140
|
13668 |
-
},
|
13669 |
-
{
|
13670 |
-
"epoch": 11.672179868200026,
|
13671 |
-
"learning_rate": 1.2713600414883962e-07,
|
13672 |
-
"loss": 2.4271,
|
13673 |
-
"step": 45160
|
13674 |
-
},
|
13675 |
-
{
|
13676 |
-
"epoch": 11.67734849463755,
|
13677 |
-
"learning_rate": 1.2710359133929728e-07,
|
13678 |
-
"loss": 2.5235,
|
13679 |
-
"step": 45180
|
13680 |
-
},
|
13681 |
-
{
|
13682 |
-
"epoch": 11.682517121075074,
|
13683 |
-
"learning_rate": 1.2707117852975494e-07,
|
13684 |
-
"loss": 2.4666,
|
13685 |
-
"step": 45200
|
13686 |
-
},
|
13687 |
-
{
|
13688 |
-
"epoch": 11.687685747512598,
|
13689 |
-
"learning_rate": 1.2703876572021263e-07,
|
13690 |
-
"loss": 2.4413,
|
13691 |
-
"step": 45220
|
13692 |
-
},
|
13693 |
-
{
|
13694 |
-
"epoch": 11.692854373950123,
|
13695 |
-
"learning_rate": 1.270063529106703e-07,
|
13696 |
-
"loss": 2.422,
|
13697 |
-
"step": 45240
|
13698 |
-
},
|
13699 |
-
{
|
13700 |
-
"epoch": 11.698023000387646,
|
13701 |
-
"learning_rate": 1.2697394010112796e-07,
|
13702 |
-
"loss": 2.5057,
|
13703 |
-
"step": 45260
|
13704 |
-
},
|
13705 |
-
{
|
13706 |
-
"epoch": 11.703191626825172,
|
13707 |
-
"learning_rate": 1.2694152729158562e-07,
|
13708 |
-
"loss": 2.5153,
|
13709 |
-
"step": 45280
|
13710 |
-
},
|
13711 |
-
{
|
13712 |
-
"epoch": 11.708360253262695,
|
13713 |
-
"learning_rate": 1.269091144820433e-07,
|
13714 |
-
"loss": 2.4894,
|
13715 |
-
"step": 45300
|
13716 |
-
},
|
13717 |
-
{
|
13718 |
-
"epoch": 11.71352887970022,
|
13719 |
-
"learning_rate": 1.2687670167250097e-07,
|
13720 |
-
"loss": 2.5211,
|
13721 |
-
"step": 45320
|
13722 |
-
},
|
13723 |
-
{
|
13724 |
-
"epoch": 11.718697506137744,
|
13725 |
-
"learning_rate": 1.2684428886295863e-07,
|
13726 |
-
"loss": 2.4449,
|
13727 |
-
"step": 45340
|
13728 |
-
},
|
13729 |
-
{
|
13730 |
-
"epoch": 11.723866132575267,
|
13731 |
-
"learning_rate": 1.268118760534163e-07,
|
13732 |
-
"loss": 2.4859,
|
13733 |
-
"step": 45360
|
13734 |
-
},
|
13735 |
-
{
|
13736 |
-
"epoch": 11.729034759012793,
|
13737 |
-
"learning_rate": 1.2677946324387398e-07,
|
13738 |
-
"loss": 2.4578,
|
13739 |
-
"step": 45380
|
13740 |
-
},
|
13741 |
-
{
|
13742 |
-
"epoch": 11.734203385450316,
|
13743 |
-
"learning_rate": 1.2674705043433164e-07,
|
13744 |
-
"loss": 2.4715,
|
13745 |
-
"step": 45400
|
13746 |
-
},
|
13747 |
-
{
|
13748 |
-
"epoch": 11.739372011887841,
|
13749 |
-
"learning_rate": 1.267146376247893e-07,
|
13750 |
-
"loss": 2.4258,
|
13751 |
-
"step": 45420
|
13752 |
-
},
|
13753 |
-
{
|
13754 |
-
"epoch": 11.744540638325365,
|
13755 |
-
"learning_rate": 1.26682224815247e-07,
|
13756 |
-
"loss": 2.4635,
|
13757 |
-
"step": 45440
|
13758 |
-
},
|
13759 |
-
{
|
13760 |
-
"epoch": 11.74970926476289,
|
13761 |
-
"learning_rate": 1.2664981200570465e-07,
|
13762 |
-
"loss": 2.4803,
|
13763 |
-
"step": 45460
|
13764 |
-
},
|
13765 |
-
{
|
13766 |
-
"epoch": 11.754877891200413,
|
13767 |
-
"learning_rate": 1.2661739919616232e-07,
|
13768 |
-
"loss": 2.4333,
|
13769 |
-
"step": 45480
|
13770 |
-
},
|
13771 |
-
{
|
13772 |
-
"epoch": 11.760046517637937,
|
13773 |
-
"learning_rate": 1.2658498638661998e-07,
|
13774 |
-
"loss": 2.5486,
|
13775 |
-
"step": 45500
|
13776 |
-
},
|
13777 |
-
{
|
13778 |
-
"epoch": 11.765215144075462,
|
13779 |
-
"learning_rate": 1.2655257357707767e-07,
|
13780 |
-
"loss": 2.4803,
|
13781 |
-
"step": 45520
|
13782 |
-
},
|
13783 |
-
{
|
13784 |
-
"epoch": 11.770383770512986,
|
13785 |
-
"learning_rate": 1.2652016076753533e-07,
|
13786 |
-
"loss": 2.4958,
|
13787 |
-
"step": 45540
|
13788 |
-
},
|
13789 |
-
{
|
13790 |
-
"epoch": 11.775552396950511,
|
13791 |
-
"learning_rate": 1.26487747957993e-07,
|
13792 |
-
"loss": 2.4918,
|
13793 |
-
"step": 45560
|
13794 |
-
},
|
13795 |
-
{
|
13796 |
-
"epoch": 11.780721023388034,
|
13797 |
-
"learning_rate": 1.2645533514845065e-07,
|
13798 |
-
"loss": 2.513,
|
13799 |
-
"step": 45580
|
13800 |
-
},
|
13801 |
-
{
|
13802 |
-
"epoch": 11.78588964982556,
|
13803 |
-
"learning_rate": 1.2642292233890834e-07,
|
13804 |
-
"loss": 2.4897,
|
13805 |
-
"step": 45600
|
13806 |
-
},
|
13807 |
-
{
|
13808 |
-
"epoch": 11.791058276263083,
|
13809 |
-
"learning_rate": 1.26390509529366e-07,
|
13810 |
-
"loss": 2.4468,
|
13811 |
-
"step": 45620
|
13812 |
-
},
|
13813 |
-
{
|
13814 |
-
"epoch": 11.796226902700607,
|
13815 |
-
"learning_rate": 1.2635809671982366e-07,
|
13816 |
-
"loss": 2.4928,
|
13817 |
-
"step": 45640
|
13818 |
-
},
|
13819 |
-
{
|
13820 |
-
"epoch": 11.801395529138132,
|
13821 |
-
"learning_rate": 1.2632568391028132e-07,
|
13822 |
-
"loss": 2.4793,
|
13823 |
-
"step": 45660
|
13824 |
-
},
|
13825 |
-
{
|
13826 |
-
"epoch": 11.806564155575655,
|
13827 |
-
"learning_rate": 1.26293271100739e-07,
|
13828 |
-
"loss": 2.4788,
|
13829 |
-
"step": 45680
|
13830 |
-
},
|
13831 |
-
{
|
13832 |
-
"epoch": 11.81173278201318,
|
13833 |
-
"learning_rate": 1.2626085829119668e-07,
|
13834 |
-
"loss": 2.4913,
|
13835 |
-
"step": 45700
|
13836 |
-
},
|
13837 |
-
{
|
13838 |
-
"epoch": 11.816901408450704,
|
13839 |
-
"learning_rate": 1.2622844548165434e-07,
|
13840 |
-
"loss": 2.4966,
|
13841 |
-
"step": 45720
|
13842 |
-
},
|
13843 |
-
{
|
13844 |
-
"epoch": 11.82207003488823,
|
13845 |
-
"learning_rate": 1.2619603267211203e-07,
|
13846 |
-
"loss": 2.4954,
|
13847 |
-
"step": 45740
|
13848 |
-
},
|
13849 |
-
{
|
13850 |
-
"epoch": 11.827238661325753,
|
13851 |
-
"learning_rate": 1.261636198625697e-07,
|
13852 |
-
"loss": 2.4608,
|
13853 |
-
"step": 45760
|
13854 |
-
},
|
13855 |
-
{
|
13856 |
-
"epoch": 11.832407287763276,
|
13857 |
-
"learning_rate": 1.2613120705302735e-07,
|
13858 |
-
"loss": 2.4901,
|
13859 |
-
"step": 45780
|
13860 |
-
},
|
13861 |
-
{
|
13862 |
-
"epoch": 11.837575914200801,
|
13863 |
-
"learning_rate": 1.26098794243485e-07,
|
13864 |
-
"loss": 2.5462,
|
13865 |
-
"step": 45800
|
13866 |
-
},
|
13867 |
-
{
|
13868 |
-
"epoch": 11.842744540638325,
|
13869 |
-
"learning_rate": 1.260663814339427e-07,
|
13870 |
-
"loss": 2.5217,
|
13871 |
-
"step": 45820
|
13872 |
-
},
|
13873 |
-
{
|
13874 |
-
"epoch": 11.84791316707585,
|
13875 |
-
"learning_rate": 1.2603396862440036e-07,
|
13876 |
-
"loss": 2.4889,
|
13877 |
-
"step": 45840
|
13878 |
-
},
|
13879 |
-
{
|
13880 |
-
"epoch": 11.853081793513374,
|
13881 |
-
"learning_rate": 1.2600155581485802e-07,
|
13882 |
-
"loss": 2.5021,
|
13883 |
-
"step": 45860
|
13884 |
-
},
|
13885 |
-
{
|
13886 |
-
"epoch": 11.858250419950899,
|
13887 |
-
"learning_rate": 1.2596914300531568e-07,
|
13888 |
-
"loss": 2.5027,
|
13889 |
-
"step": 45880
|
13890 |
-
},
|
13891 |
-
{
|
13892 |
-
"epoch": 11.863419046388422,
|
13893 |
-
"learning_rate": 1.2593673019577337e-07,
|
13894 |
-
"loss": 2.4681,
|
13895 |
-
"step": 45900
|
13896 |
-
},
|
13897 |
-
{
|
13898 |
-
"epoch": 11.868587672825946,
|
13899 |
-
"learning_rate": 1.2590431738623104e-07,
|
13900 |
-
"loss": 2.409,
|
13901 |
-
"step": 45920
|
13902 |
-
},
|
13903 |
-
{
|
13904 |
-
"epoch": 11.873756299263471,
|
13905 |
-
"learning_rate": 1.258719045766887e-07,
|
13906 |
-
"loss": 2.4426,
|
13907 |
-
"step": 45940
|
13908 |
-
},
|
13909 |
-
{
|
13910 |
-
"epoch": 11.878924925700995,
|
13911 |
-
"learning_rate": 1.2583949176714636e-07,
|
13912 |
-
"loss": 2.4285,
|
13913 |
-
"step": 45960
|
13914 |
-
},
|
13915 |
-
{
|
13916 |
-
"epoch": 11.88409355213852,
|
13917 |
-
"learning_rate": 1.2580707895760405e-07,
|
13918 |
-
"loss": 2.4452,
|
13919 |
-
"step": 45980
|
13920 |
-
},
|
13921 |
-
{
|
13922 |
-
"epoch": 11.889262178576043,
|
13923 |
-
"learning_rate": 1.257746661480617e-07,
|
13924 |
-
"loss": 2.4816,
|
13925 |
-
"step": 46000
|
13926 |
-
},
|
13927 |
-
{
|
13928 |
-
"epoch": 11.894430805013567,
|
13929 |
-
"learning_rate": 1.2574225333851937e-07,
|
13930 |
-
"loss": 2.4585,
|
13931 |
-
"step": 46020
|
13932 |
-
},
|
13933 |
-
{
|
13934 |
-
"epoch": 11.899599431451092,
|
13935 |
-
"learning_rate": 1.2570984052897706e-07,
|
13936 |
-
"loss": 2.4171,
|
13937 |
-
"step": 46040
|
13938 |
-
},
|
13939 |
-
{
|
13940 |
-
"epoch": 11.904768057888615,
|
13941 |
-
"learning_rate": 1.2567742771943472e-07,
|
13942 |
-
"loss": 2.4597,
|
13943 |
-
"step": 46060
|
13944 |
-
},
|
13945 |
-
{
|
13946 |
-
"epoch": 11.90993668432614,
|
13947 |
-
"learning_rate": 1.2564501490989238e-07,
|
13948 |
-
"loss": 2.4195,
|
13949 |
-
"step": 46080
|
13950 |
-
},
|
13951 |
-
{
|
13952 |
-
"epoch": 11.915105310763664,
|
13953 |
-
"learning_rate": 1.2561260210035004e-07,
|
13954 |
-
"loss": 2.5034,
|
13955 |
-
"step": 46100
|
13956 |
-
},
|
13957 |
-
{
|
13958 |
-
"epoch": 11.92027393720119,
|
13959 |
-
"learning_rate": 1.2558018929080773e-07,
|
13960 |
-
"loss": 2.4407,
|
13961 |
-
"step": 46120
|
13962 |
-
},
|
13963 |
-
{
|
13964 |
-
"epoch": 11.925442563638713,
|
13965 |
-
"learning_rate": 1.255477764812654e-07,
|
13966 |
-
"loss": 2.4813,
|
13967 |
-
"step": 46140
|
13968 |
-
},
|
13969 |
-
{
|
13970 |
-
"epoch": 11.930611190076238,
|
13971 |
-
"learning_rate": 1.2551536367172306e-07,
|
13972 |
-
"loss": 2.4599,
|
13973 |
-
"step": 46160
|
13974 |
-
},
|
13975 |
-
{
|
13976 |
-
"epoch": 11.935779816513762,
|
13977 |
-
"learning_rate": 1.2548295086218072e-07,
|
13978 |
-
"loss": 2.4641,
|
13979 |
-
"step": 46180
|
13980 |
-
},
|
13981 |
-
{
|
13982 |
-
"epoch": 11.940948442951285,
|
13983 |
-
"learning_rate": 1.254505380526384e-07,
|
13984 |
-
"loss": 2.5262,
|
13985 |
-
"step": 46200
|
13986 |
-
},
|
13987 |
-
{
|
13988 |
-
"epoch": 11.94611706938881,
|
13989 |
-
"learning_rate": 1.2541812524309607e-07,
|
13990 |
-
"loss": 2.4518,
|
13991 |
-
"step": 46220
|
13992 |
-
},
|
13993 |
-
{
|
13994 |
-
"epoch": 11.951285695826334,
|
13995 |
-
"learning_rate": 1.2538571243355373e-07,
|
13996 |
-
"loss": 2.5018,
|
13997 |
-
"step": 46240
|
13998 |
-
},
|
13999 |
-
{
|
14000 |
-
"epoch": 11.956454322263859,
|
14001 |
-
"learning_rate": 1.253532996240114e-07,
|
14002 |
-
"loss": 2.4634,
|
14003 |
-
"step": 46260
|
14004 |
-
},
|
14005 |
-
{
|
14006 |
-
"epoch": 11.961622948701383,
|
14007 |
-
"learning_rate": 1.2532088681446908e-07,
|
14008 |
-
"loss": 2.4889,
|
14009 |
-
"step": 46280
|
14010 |
-
},
|
14011 |
-
{
|
14012 |
-
"epoch": 11.966791575138906,
|
14013 |
-
"learning_rate": 1.2528847400492674e-07,
|
14014 |
-
"loss": 2.4577,
|
14015 |
-
"step": 46300
|
14016 |
-
},
|
14017 |
-
{
|
14018 |
-
"epoch": 11.971960201576431,
|
14019 |
-
"learning_rate": 1.252560611953844e-07,
|
14020 |
-
"loss": 2.4812,
|
14021 |
-
"step": 46320
|
14022 |
-
},
|
14023 |
-
{
|
14024 |
-
"epoch": 11.977128828013955,
|
14025 |
-
"learning_rate": 1.252236483858421e-07,
|
14026 |
-
"loss": 2.4448,
|
14027 |
-
"step": 46340
|
14028 |
-
},
|
14029 |
-
{
|
14030 |
-
"epoch": 11.98229745445148,
|
14031 |
-
"learning_rate": 1.2519123557629975e-07,
|
14032 |
-
"loss": 2.4594,
|
14033 |
-
"step": 46360
|
14034 |
-
},
|
14035 |
-
{
|
14036 |
-
"epoch": 11.987466080889003,
|
14037 |
-
"learning_rate": 1.2515882276675742e-07,
|
14038 |
-
"loss": 2.4473,
|
14039 |
-
"step": 46380
|
14040 |
-
},
|
14041 |
-
{
|
14042 |
-
"epoch": 11.992634707326529,
|
14043 |
-
"learning_rate": 1.2512640995721508e-07,
|
14044 |
-
"loss": 2.4831,
|
14045 |
-
"step": 46400
|
14046 |
-
},
|
14047 |
-
{
|
14048 |
-
"epoch": 11.997803333764052,
|
14049 |
-
"learning_rate": 1.2509399714767277e-07,
|
14050 |
-
"loss": 2.4706,
|
14051 |
-
"step": 46420
|
14052 |
-
},
|
14053 |
-
{
|
14054 |
-
"epoch": 11.999870784339063,
|
14055 |
-
"eval_bleu": 12.51,
|
14056 |
-
"eval_gen_len": 37.4146,
|
14057 |
-
"eval_loss": 2.4478776454925537,
|
14058 |
-
"eval_runtime": 807.7318,
|
14059 |
-
"eval_samples_per_second": 2.132,
|
14060 |
-
"eval_steps_per_second": 1.066,
|
14061 |
-
"step": 46428
|
14062 |
-
},
|
14063 |
-
{
|
14064 |
-
"epoch": 12.002971960201576,
|
14065 |
-
"learning_rate": 1.2506158433813043e-07,
|
14066 |
-
"loss": 2.4527,
|
14067 |
-
"step": 46440
|
14068 |
-
},
|
14069 |
-
{
|
14070 |
-
"epoch": 12.008140586639101,
|
14071 |
-
"learning_rate": 1.250291715285881e-07,
|
14072 |
-
"loss": 2.436,
|
14073 |
-
"step": 46460
|
14074 |
-
},
|
14075 |
-
{
|
14076 |
-
"epoch": 12.013309213076624,
|
14077 |
-
"learning_rate": 1.2499675871904575e-07,
|
14078 |
-
"loss": 2.4514,
|
14079 |
-
"step": 46480
|
14080 |
-
},
|
14081 |
-
{
|
14082 |
-
"epoch": 12.01847783951415,
|
14083 |
-
"learning_rate": 1.2496434590950344e-07,
|
14084 |
-
"loss": 2.4921,
|
14085 |
-
"step": 46500
|
14086 |
-
},
|
14087 |
-
{
|
14088 |
-
"epoch": 12.023646465951673,
|
14089 |
-
"learning_rate": 1.249319330999611e-07,
|
14090 |
-
"loss": 2.4637,
|
14091 |
-
"step": 46520
|
14092 |
-
},
|
14093 |
-
{
|
14094 |
-
"epoch": 12.028815092389198,
|
14095 |
-
"learning_rate": 1.2489952029041876e-07,
|
14096 |
-
"loss": 2.4924,
|
14097 |
-
"step": 46540
|
14098 |
-
},
|
14099 |
-
{
|
14100 |
-
"epoch": 12.033983718826722,
|
14101 |
-
"learning_rate": 1.2486710748087643e-07,
|
14102 |
-
"loss": 2.4788,
|
14103 |
-
"step": 46560
|
14104 |
-
},
|
14105 |
-
{
|
14106 |
-
"epoch": 12.039152345264245,
|
14107 |
-
"learning_rate": 1.2483469467133411e-07,
|
14108 |
-
"loss": 2.5291,
|
14109 |
-
"step": 46580
|
14110 |
-
},
|
14111 |
-
{
|
14112 |
-
"epoch": 12.04432097170177,
|
14113 |
-
"learning_rate": 1.2480228186179178e-07,
|
14114 |
-
"loss": 2.4419,
|
14115 |
-
"step": 46600
|
14116 |
-
},
|
14117 |
-
{
|
14118 |
-
"epoch": 12.049489598139294,
|
14119 |
-
"learning_rate": 1.2476986905224944e-07,
|
14120 |
-
"loss": 2.4768,
|
14121 |
-
"step": 46620
|
14122 |
-
},
|
14123 |
-
{
|
14124 |
-
"epoch": 12.05465822457682,
|
14125 |
-
"learning_rate": 1.2473745624270713e-07,
|
14126 |
-
"loss": 2.444,
|
14127 |
-
"step": 46640
|
14128 |
-
},
|
14129 |
-
{
|
14130 |
-
"epoch": 12.059826851014343,
|
14131 |
-
"learning_rate": 1.247050434331648e-07,
|
14132 |
-
"loss": 2.5193,
|
14133 |
-
"step": 46660
|
14134 |
-
},
|
14135 |
-
{
|
14136 |
-
"epoch": 12.064995477451868,
|
14137 |
-
"learning_rate": 1.2467263062362245e-07,
|
14138 |
-
"loss": 2.5019,
|
14139 |
-
"step": 46680
|
14140 |
-
},
|
14141 |
-
{
|
14142 |
-
"epoch": 12.070164103889391,
|
14143 |
-
"learning_rate": 1.246402178140801e-07,
|
14144 |
-
"loss": 2.465,
|
14145 |
-
"step": 46700
|
14146 |
-
},
|
14147 |
-
{
|
14148 |
-
"epoch": 12.075332730326915,
|
14149 |
-
"learning_rate": 1.246078050045378e-07,
|
14150 |
-
"loss": 2.4508,
|
14151 |
-
"step": 46720
|
14152 |
-
},
|
14153 |
-
{
|
14154 |
-
"epoch": 12.08050135676444,
|
14155 |
-
"learning_rate": 1.2457539219499546e-07,
|
14156 |
-
"loss": 2.4784,
|
14157 |
-
"step": 46740
|
14158 |
-
},
|
14159 |
-
{
|
14160 |
-
"epoch": 12.085669983201964,
|
14161 |
-
"learning_rate": 1.2454297938545312e-07,
|
14162 |
-
"loss": 2.5013,
|
14163 |
-
"step": 46760
|
14164 |
-
},
|
14165 |
-
{
|
14166 |
-
"epoch": 12.090838609639489,
|
14167 |
-
"learning_rate": 1.2451056657591079e-07,
|
14168 |
-
"loss": 2.3997,
|
14169 |
-
"step": 46780
|
14170 |
-
},
|
14171 |
-
{
|
14172 |
-
"epoch": 12.096007236077012,
|
14173 |
-
"learning_rate": 1.2447815376636847e-07,
|
14174 |
-
"loss": 2.4475,
|
14175 |
-
"step": 46800
|
14176 |
-
},
|
14177 |
-
{
|
14178 |
-
"epoch": 12.101175862514538,
|
14179 |
-
"learning_rate": 1.2444574095682614e-07,
|
14180 |
-
"loss": 2.4564,
|
14181 |
-
"step": 46820
|
14182 |
-
},
|
14183 |
-
{
|
14184 |
-
"epoch": 12.106344488952061,
|
14185 |
-
"learning_rate": 1.244133281472838e-07,
|
14186 |
-
"loss": 2.5028,
|
14187 |
-
"step": 46840
|
14188 |
-
},
|
14189 |
-
{
|
14190 |
-
"epoch": 12.111513115389585,
|
14191 |
-
"learning_rate": 1.2438091533774146e-07,
|
14192 |
-
"loss": 2.4969,
|
14193 |
-
"step": 46860
|
14194 |
-
},
|
14195 |
-
{
|
14196 |
-
"epoch": 12.11668174182711,
|
14197 |
-
"learning_rate": 1.2434850252819915e-07,
|
14198 |
-
"loss": 2.4488,
|
14199 |
-
"step": 46880
|
14200 |
-
},
|
14201 |
-
{
|
14202 |
-
"epoch": 12.121850368264633,
|
14203 |
-
"learning_rate": 1.243160897186568e-07,
|
14204 |
-
"loss": 2.4779,
|
14205 |
-
"step": 46900
|
14206 |
-
},
|
14207 |
-
{
|
14208 |
-
"epoch": 12.127018994702158,
|
14209 |
-
"learning_rate": 1.2428367690911447e-07,
|
14210 |
-
"loss": 2.4605,
|
14211 |
-
"step": 46920
|
14212 |
-
},
|
14213 |
-
{
|
14214 |
-
"epoch": 12.132187621139682,
|
14215 |
-
"learning_rate": 1.2425126409957216e-07,
|
14216 |
-
"loss": 2.4592,
|
14217 |
-
"step": 46940
|
14218 |
-
},
|
14219 |
-
{
|
14220 |
-
"epoch": 12.137356247577207,
|
14221 |
-
"learning_rate": 1.2421885129002982e-07,
|
14222 |
-
"loss": 2.5115,
|
14223 |
-
"step": 46960
|
14224 |
-
},
|
14225 |
-
{
|
14226 |
-
"epoch": 12.14252487401473,
|
14227 |
-
"learning_rate": 1.2418643848048748e-07,
|
14228 |
-
"loss": 2.494,
|
14229 |
-
"step": 46980
|
14230 |
-
},
|
14231 |
-
{
|
14232 |
-
"epoch": 12.147693500452254,
|
14233 |
-
"learning_rate": 1.2415402567094515e-07,
|
14234 |
-
"loss": 2.4545,
|
14235 |
-
"step": 47000
|
14236 |
-
},
|
14237 |
-
{
|
14238 |
-
"epoch": 12.15286212688978,
|
14239 |
-
"learning_rate": 1.2412161286140283e-07,
|
14240 |
-
"loss": 2.5211,
|
14241 |
-
"step": 47020
|
14242 |
-
},
|
14243 |
-
{
|
14244 |
-
"epoch": 12.158030753327303,
|
14245 |
-
"learning_rate": 1.240892000518605e-07,
|
14246 |
-
"loss": 2.4661,
|
14247 |
-
"step": 47040
|
14248 |
-
},
|
14249 |
-
{
|
14250 |
-
"epoch": 12.163199379764828,
|
14251 |
-
"learning_rate": 1.2405678724231816e-07,
|
14252 |
-
"loss": 2.4593,
|
14253 |
-
"step": 47060
|
14254 |
-
},
|
14255 |
-
{
|
14256 |
-
"epoch": 12.168368006202352,
|
14257 |
-
"learning_rate": 1.2402437443277582e-07,
|
14258 |
-
"loss": 2.555,
|
14259 |
-
"step": 47080
|
14260 |
-
},
|
14261 |
-
{
|
14262 |
-
"epoch": 12.173536632639877,
|
14263 |
-
"learning_rate": 1.239919616232335e-07,
|
14264 |
-
"loss": 2.5006,
|
14265 |
-
"step": 47100
|
14266 |
-
},
|
14267 |
-
{
|
14268 |
-
"epoch": 12.1787052590774,
|
14269 |
-
"learning_rate": 1.2395954881369117e-07,
|
14270 |
-
"loss": 2.4607,
|
14271 |
-
"step": 47120
|
14272 |
-
},
|
14273 |
-
{
|
14274 |
-
"epoch": 12.183873885514924,
|
14275 |
-
"learning_rate": 1.2392713600414883e-07,
|
14276 |
-
"loss": 2.4595,
|
14277 |
-
"step": 47140
|
14278 |
-
},
|
14279 |
-
{
|
14280 |
-
"epoch": 12.189042511952449,
|
14281 |
-
"learning_rate": 1.238947231946065e-07,
|
14282 |
-
"loss": 2.444,
|
14283 |
-
"step": 47160
|
14284 |
-
},
|
14285 |
-
{
|
14286 |
-
"epoch": 12.194211138389973,
|
14287 |
-
"learning_rate": 1.2386231038506418e-07,
|
14288 |
-
"loss": 2.4657,
|
14289 |
-
"step": 47180
|
14290 |
-
},
|
14291 |
-
{
|
14292 |
-
"epoch": 12.199379764827498,
|
14293 |
-
"learning_rate": 1.2382989757552184e-07,
|
14294 |
-
"loss": 2.454,
|
14295 |
-
"step": 47200
|
14296 |
-
},
|
14297 |
-
{
|
14298 |
-
"epoch": 12.204548391265021,
|
14299 |
-
"learning_rate": 1.237974847659795e-07,
|
14300 |
-
"loss": 2.4939,
|
14301 |
-
"step": 47220
|
14302 |
-
},
|
14303 |
-
{
|
14304 |
-
"epoch": 12.209717017702545,
|
14305 |
-
"learning_rate": 1.237650719564372e-07,
|
14306 |
-
"loss": 2.4465,
|
14307 |
-
"step": 47240
|
14308 |
-
},
|
14309 |
-
{
|
14310 |
-
"epoch": 12.21488564414007,
|
14311 |
-
"learning_rate": 1.2373265914689486e-07,
|
14312 |
-
"loss": 2.5464,
|
14313 |
-
"step": 47260
|
14314 |
-
},
|
14315 |
-
{
|
14316 |
-
"epoch": 12.220054270577593,
|
14317 |
-
"learning_rate": 1.2370024633735252e-07,
|
14318 |
-
"loss": 2.4521,
|
14319 |
-
"step": 47280
|
14320 |
-
},
|
14321 |
-
{
|
14322 |
-
"epoch": 12.225222897015119,
|
14323 |
-
"learning_rate": 1.2366783352781018e-07,
|
14324 |
-
"loss": 2.5132,
|
14325 |
-
"step": 47300
|
14326 |
-
},
|
14327 |
-
{
|
14328 |
-
"epoch": 12.230391523452642,
|
14329 |
-
"learning_rate": 1.2363542071826787e-07,
|
14330 |
-
"loss": 2.43,
|
14331 |
-
"step": 47320
|
14332 |
-
},
|
14333 |
-
{
|
14334 |
-
"epoch": 12.235560149890167,
|
14335 |
-
"learning_rate": 1.2360300790872553e-07,
|
14336 |
-
"loss": 2.4295,
|
14337 |
-
"step": 47340
|
14338 |
-
},
|
14339 |
-
{
|
14340 |
-
"epoch": 12.24072877632769,
|
14341 |
-
"learning_rate": 1.235705950991832e-07,
|
14342 |
-
"loss": 2.5134,
|
14343 |
-
"step": 47360
|
14344 |
-
},
|
14345 |
-
{
|
14346 |
-
"epoch": 12.245897402765214,
|
14347 |
-
"learning_rate": 1.2353818228964085e-07,
|
14348 |
-
"loss": 2.461,
|
14349 |
-
"step": 47380
|
14350 |
-
},
|
14351 |
-
{
|
14352 |
-
"epoch": 12.25106602920274,
|
14353 |
-
"learning_rate": 1.2350576948009854e-07,
|
14354 |
-
"loss": 2.4942,
|
14355 |
-
"step": 47400
|
14356 |
-
},
|
14357 |
-
{
|
14358 |
-
"epoch": 12.256234655640263,
|
14359 |
-
"learning_rate": 1.234733566705562e-07,
|
14360 |
-
"loss": 2.4607,
|
14361 |
-
"step": 47420
|
14362 |
-
},
|
14363 |
-
{
|
14364 |
-
"epoch": 12.261403282077788,
|
14365 |
-
"learning_rate": 1.2344094386101387e-07,
|
14366 |
-
"loss": 2.4634,
|
14367 |
-
"step": 47440
|
14368 |
-
},
|
14369 |
-
{
|
14370 |
-
"epoch": 12.266571908515312,
|
14371 |
-
"learning_rate": 1.2340853105147153e-07,
|
14372 |
-
"loss": 2.4448,
|
14373 |
-
"step": 47460
|
14374 |
-
},
|
14375 |
-
{
|
14376 |
-
"epoch": 12.271740534952837,
|
14377 |
-
"learning_rate": 1.2337611824192922e-07,
|
14378 |
-
"loss": 2.4803,
|
14379 |
-
"step": 47480
|
14380 |
-
},
|
14381 |
-
{
|
14382 |
-
"epoch": 12.27690916139036,
|
14383 |
-
"learning_rate": 1.2334370543238688e-07,
|
14384 |
-
"loss": 2.4154,
|
14385 |
-
"step": 47500
|
14386 |
-
},
|
14387 |
-
{
|
14388 |
-
"epoch": 12.282077787827884,
|
14389 |
-
"learning_rate": 1.2331129262284454e-07,
|
14390 |
-
"loss": 2.5181,
|
14391 |
-
"step": 47520
|
14392 |
-
},
|
14393 |
-
{
|
14394 |
-
"epoch": 12.28724641426541,
|
14395 |
-
"learning_rate": 1.2327887981330223e-07,
|
14396 |
-
"loss": 2.5086,
|
14397 |
-
"step": 47540
|
14398 |
-
},
|
14399 |
-
{
|
14400 |
-
"epoch": 12.292415040702933,
|
14401 |
-
"learning_rate": 1.232464670037599e-07,
|
14402 |
-
"loss": 2.4487,
|
14403 |
-
"step": 47560
|
14404 |
-
},
|
14405 |
-
{
|
14406 |
-
"epoch": 12.297583667140458,
|
14407 |
-
"learning_rate": 1.2321405419421755e-07,
|
14408 |
-
"loss": 2.4109,
|
14409 |
-
"step": 47580
|
14410 |
-
},
|
14411 |
-
{
|
14412 |
-
"epoch": 12.302752293577981,
|
14413 |
-
"learning_rate": 1.2318164138467521e-07,
|
14414 |
-
"loss": 2.4675,
|
14415 |
-
"step": 47600
|
14416 |
-
},
|
14417 |
-
{
|
14418 |
-
"epoch": 12.307920920015507,
|
14419 |
-
"learning_rate": 1.2314922857513288e-07,
|
14420 |
-
"loss": 2.4326,
|
14421 |
-
"step": 47620
|
14422 |
-
},
|
14423 |
-
{
|
14424 |
-
"epoch": 12.31308954645303,
|
14425 |
-
"learning_rate": 1.2311681576559056e-07,
|
14426 |
-
"loss": 2.441,
|
14427 |
-
"step": 47640
|
14428 |
-
},
|
14429 |
-
{
|
14430 |
-
"epoch": 12.318258172890554,
|
14431 |
-
"learning_rate": 1.2308440295604823e-07,
|
14432 |
-
"loss": 2.4154,
|
14433 |
-
"step": 47660
|
14434 |
-
},
|
14435 |
-
{
|
14436 |
-
"epoch": 12.323426799328079,
|
14437 |
-
"learning_rate": 1.230519901465059e-07,
|
14438 |
-
"loss": 2.4491,
|
14439 |
-
"step": 47680
|
14440 |
-
},
|
14441 |
-
{
|
14442 |
-
"epoch": 12.328595425765602,
|
14443 |
-
"learning_rate": 1.2301957733696358e-07,
|
14444 |
-
"loss": 2.4764,
|
14445 |
-
"step": 47700
|
14446 |
-
},
|
14447 |
-
{
|
14448 |
-
"epoch": 12.333764052203128,
|
14449 |
-
"learning_rate": 1.2298716452742124e-07,
|
14450 |
-
"loss": 2.4786,
|
14451 |
-
"step": 47720
|
14452 |
-
},
|
14453 |
-
{
|
14454 |
-
"epoch": 12.338932678640651,
|
14455 |
-
"learning_rate": 1.229547517178789e-07,
|
14456 |
-
"loss": 2.4033,
|
14457 |
-
"step": 47740
|
14458 |
-
},
|
14459 |
-
{
|
14460 |
-
"epoch": 12.344101305078176,
|
14461 |
-
"learning_rate": 1.2292233890833656e-07,
|
14462 |
-
"loss": 2.4742,
|
14463 |
-
"step": 47760
|
14464 |
-
},
|
14465 |
-
{
|
14466 |
-
"epoch": 12.3492699315157,
|
14467 |
-
"learning_rate": 1.2288992609879422e-07,
|
14468 |
-
"loss": 2.5316,
|
14469 |
-
"step": 47780
|
14470 |
-
},
|
14471 |
-
{
|
14472 |
-
"epoch": 12.354438557953223,
|
14473 |
-
"learning_rate": 1.228575132892519e-07,
|
14474 |
-
"loss": 2.4717,
|
14475 |
-
"step": 47800
|
14476 |
-
},
|
14477 |
-
{
|
14478 |
-
"epoch": 12.359607184390748,
|
14479 |
-
"learning_rate": 1.2282510047970957e-07,
|
14480 |
-
"loss": 2.4518,
|
14481 |
-
"step": 47820
|
14482 |
-
},
|
14483 |
-
{
|
14484 |
-
"epoch": 12.364775810828272,
|
14485 |
-
"learning_rate": 1.2279268767016724e-07,
|
14486 |
-
"loss": 2.4747,
|
14487 |
-
"step": 47840
|
14488 |
-
},
|
14489 |
-
{
|
14490 |
-
"epoch": 12.369944437265797,
|
14491 |
-
"learning_rate": 1.2276027486062492e-07,
|
14492 |
-
"loss": 2.4391,
|
14493 |
-
"step": 47860
|
14494 |
-
},
|
14495 |
-
{
|
14496 |
-
"epoch": 12.37511306370332,
|
14497 |
-
"learning_rate": 1.2272786205108259e-07,
|
14498 |
-
"loss": 2.4076,
|
14499 |
-
"step": 47880
|
14500 |
-
},
|
14501 |
-
{
|
14502 |
-
"epoch": 12.380281690140846,
|
14503 |
-
"learning_rate": 1.2269544924154025e-07,
|
14504 |
-
"loss": 2.4275,
|
14505 |
-
"step": 47900
|
14506 |
-
},
|
14507 |
-
{
|
14508 |
-
"epoch": 12.38545031657837,
|
14509 |
-
"learning_rate": 1.2266303643199794e-07,
|
14510 |
-
"loss": 2.4593,
|
14511 |
-
"step": 47920
|
14512 |
-
},
|
14513 |
-
{
|
14514 |
-
"epoch": 12.390618943015893,
|
14515 |
-
"learning_rate": 1.226306236224556e-07,
|
14516 |
-
"loss": 2.4666,
|
14517 |
-
"step": 47940
|
14518 |
-
},
|
14519 |
-
{
|
14520 |
-
"epoch": 12.395787569453418,
|
14521 |
-
"learning_rate": 1.2259821081291326e-07,
|
14522 |
-
"loss": 2.4939,
|
14523 |
-
"step": 47960
|
14524 |
-
},
|
14525 |
-
{
|
14526 |
-
"epoch": 12.400956195890942,
|
14527 |
-
"learning_rate": 1.2256579800337092e-07,
|
14528 |
-
"loss": 2.4933,
|
14529 |
-
"step": 47980
|
14530 |
-
},
|
14531 |
-
{
|
14532 |
-
"epoch": 12.406124822328467,
|
14533 |
-
"learning_rate": 1.2253338519382858e-07,
|
14534 |
-
"loss": 2.4298,
|
14535 |
-
"step": 48000
|
14536 |
-
},
|
14537 |
-
{
|
14538 |
-
"epoch": 12.41129344876599,
|
14539 |
-
"learning_rate": 1.2250097238428627e-07,
|
14540 |
-
"loss": 2.4923,
|
14541 |
-
"step": 48020
|
14542 |
-
},
|
14543 |
-
{
|
14544 |
-
"epoch": 12.416462075203516,
|
14545 |
-
"learning_rate": 1.2246855957474393e-07,
|
14546 |
-
"loss": 2.4368,
|
14547 |
-
"step": 48040
|
14548 |
-
},
|
14549 |
-
{
|
14550 |
-
"epoch": 12.421630701641039,
|
14551 |
-
"learning_rate": 1.224361467652016e-07,
|
14552 |
-
"loss": 2.4882,
|
14553 |
-
"step": 48060
|
14554 |
-
},
|
14555 |
-
{
|
14556 |
-
"epoch": 12.426799328078562,
|
14557 |
-
"learning_rate": 1.2240373395565928e-07,
|
14558 |
-
"loss": 2.4417,
|
14559 |
-
"step": 48080
|
14560 |
-
},
|
14561 |
-
{
|
14562 |
-
"epoch": 12.431967954516088,
|
14563 |
-
"learning_rate": 1.2237132114611695e-07,
|
14564 |
-
"loss": 2.4881,
|
14565 |
-
"step": 48100
|
14566 |
-
},
|
14567 |
-
{
|
14568 |
-
"epoch": 12.437136580953611,
|
14569 |
-
"learning_rate": 1.223389083365746e-07,
|
14570 |
-
"loss": 2.4633,
|
14571 |
-
"step": 48120
|
14572 |
-
},
|
14573 |
-
{
|
14574 |
-
"epoch": 12.442305207391136,
|
14575 |
-
"learning_rate": 1.223064955270323e-07,
|
14576 |
-
"loss": 2.4914,
|
14577 |
-
"step": 48140
|
14578 |
-
},
|
14579 |
-
{
|
14580 |
-
"epoch": 12.44747383382866,
|
14581 |
-
"learning_rate": 1.2227408271748993e-07,
|
14582 |
-
"loss": 2.4236,
|
14583 |
-
"step": 48160
|
14584 |
-
},
|
14585 |
-
{
|
14586 |
-
"epoch": 12.452642460266183,
|
14587 |
-
"learning_rate": 1.2224166990794762e-07,
|
14588 |
-
"loss": 2.4603,
|
14589 |
-
"step": 48180
|
14590 |
-
},
|
14591 |
-
{
|
14592 |
-
"epoch": 12.457811086703709,
|
14593 |
-
"learning_rate": 1.2220925709840528e-07,
|
14594 |
-
"loss": 2.5019,
|
14595 |
-
"step": 48200
|
14596 |
-
},
|
14597 |
-
{
|
14598 |
-
"epoch": 12.462979713141232,
|
14599 |
-
"learning_rate": 1.2217684428886294e-07,
|
14600 |
-
"loss": 2.4368,
|
14601 |
-
"step": 48220
|
14602 |
-
},
|
14603 |
-
{
|
14604 |
-
"epoch": 12.468148339578757,
|
14605 |
-
"learning_rate": 1.2214443147932063e-07,
|
14606 |
-
"loss": 2.5019,
|
14607 |
-
"step": 48240
|
14608 |
-
},
|
14609 |
-
{
|
14610 |
-
"epoch": 12.47331696601628,
|
14611 |
-
"learning_rate": 1.221120186697783e-07,
|
14612 |
-
"loss": 2.4857,
|
14613 |
-
"step": 48260
|
14614 |
-
},
|
14615 |
-
{
|
14616 |
-
"epoch": 12.478485592453806,
|
14617 |
-
"learning_rate": 1.2207960586023596e-07,
|
14618 |
-
"loss": 2.4496,
|
14619 |
-
"step": 48280
|
14620 |
-
},
|
14621 |
-
{
|
14622 |
-
"epoch": 12.48365421889133,
|
14623 |
-
"learning_rate": 1.2204719305069364e-07,
|
14624 |
-
"loss": 2.5045,
|
14625 |
-
"step": 48300
|
14626 |
-
},
|
14627 |
-
{
|
14628 |
-
"epoch": 12.488822845328853,
|
14629 |
-
"learning_rate": 1.2201478024115128e-07,
|
14630 |
-
"loss": 2.4589,
|
14631 |
-
"step": 48320
|
14632 |
-
},
|
14633 |
-
{
|
14634 |
-
"epoch": 12.493991471766378,
|
14635 |
-
"learning_rate": 1.2198236743160897e-07,
|
14636 |
-
"loss": 2.4778,
|
14637 |
-
"step": 48340
|
14638 |
-
},
|
14639 |
-
{
|
14640 |
-
"epoch": 12.499160098203902,
|
14641 |
-
"learning_rate": 1.2194995462206663e-07,
|
14642 |
-
"loss": 2.4297,
|
14643 |
-
"step": 48360
|
14644 |
-
},
|
14645 |
-
{
|
14646 |
-
"epoch": 12.504328724641427,
|
14647 |
-
"learning_rate": 1.219175418125243e-07,
|
14648 |
-
"loss": 2.395,
|
14649 |
-
"step": 48380
|
14650 |
-
},
|
14651 |
-
{
|
14652 |
-
"epoch": 12.50949735107895,
|
14653 |
-
"learning_rate": 1.2188512900298198e-07,
|
14654 |
-
"loss": 2.4515,
|
14655 |
-
"step": 48400
|
14656 |
-
},
|
14657 |
-
{
|
14658 |
-
"epoch": 12.514665977516476,
|
14659 |
-
"learning_rate": 1.2185271619343964e-07,
|
14660 |
-
"loss": 2.4912,
|
14661 |
-
"step": 48420
|
14662 |
-
},
|
14663 |
-
{
|
14664 |
-
"epoch": 12.519834603954,
|
14665 |
-
"learning_rate": 1.218203033838973e-07,
|
14666 |
-
"loss": 2.4681,
|
14667 |
-
"step": 48440
|
14668 |
-
},
|
14669 |
-
{
|
14670 |
-
"epoch": 12.525003230391523,
|
14671 |
-
"learning_rate": 1.21787890574355e-07,
|
14672 |
-
"loss": 2.4598,
|
14673 |
-
"step": 48460
|
14674 |
-
},
|
14675 |
-
{
|
14676 |
-
"epoch": 12.530171856829048,
|
14677 |
-
"learning_rate": 1.2175547776481265e-07,
|
14678 |
-
"loss": 2.4792,
|
14679 |
-
"step": 48480
|
14680 |
-
},
|
14681 |
-
{
|
14682 |
-
"epoch": 12.535340483266571,
|
14683 |
-
"learning_rate": 1.2172306495527032e-07,
|
14684 |
-
"loss": 2.4331,
|
14685 |
-
"step": 48500
|
14686 |
-
},
|
14687 |
-
{
|
14688 |
-
"epoch": 12.540509109704097,
|
14689 |
-
"learning_rate": 1.21690652145728e-07,
|
14690 |
-
"loss": 2.4537,
|
14691 |
-
"step": 48520
|
14692 |
-
},
|
14693 |
-
{
|
14694 |
-
"epoch": 12.54567773614162,
|
14695 |
-
"learning_rate": 1.2165823933618564e-07,
|
14696 |
-
"loss": 2.5384,
|
14697 |
-
"step": 48540
|
14698 |
-
},
|
14699 |
-
{
|
14700 |
-
"epoch": 12.550846362579145,
|
14701 |
-
"learning_rate": 1.2162582652664333e-07,
|
14702 |
-
"loss": 2.5251,
|
14703 |
-
"step": 48560
|
14704 |
-
},
|
14705 |
-
{
|
14706 |
-
"epoch": 12.556014989016669,
|
14707 |
-
"learning_rate": 1.21593413717101e-07,
|
14708 |
-
"loss": 2.4403,
|
14709 |
-
"step": 48580
|
14710 |
-
},
|
14711 |
-
{
|
14712 |
-
"epoch": 12.561183615454192,
|
14713 |
-
"learning_rate": 1.2156100090755865e-07,
|
14714 |
-
"loss": 2.4652,
|
14715 |
-
"step": 48600
|
14716 |
-
},
|
14717 |
-
{
|
14718 |
-
"epoch": 12.566352241891718,
|
14719 |
-
"learning_rate": 1.2152858809801634e-07,
|
14720 |
-
"loss": 2.5051,
|
14721 |
-
"step": 48620
|
14722 |
-
},
|
14723 |
-
{
|
14724 |
-
"epoch": 12.571520868329241,
|
14725 |
-
"learning_rate": 1.21496175288474e-07,
|
14726 |
-
"loss": 2.4572,
|
14727 |
-
"step": 48640
|
14728 |
-
},
|
14729 |
-
{
|
14730 |
-
"epoch": 12.576689494766766,
|
14731 |
-
"learning_rate": 1.2146376247893166e-07,
|
14732 |
-
"loss": 2.4891,
|
14733 |
-
"step": 48660
|
14734 |
-
},
|
14735 |
-
{
|
14736 |
-
"epoch": 12.58185812120429,
|
14737 |
-
"learning_rate": 1.2143134966938935e-07,
|
14738 |
-
"loss": 2.5186,
|
14739 |
-
"step": 48680
|
14740 |
-
},
|
14741 |
-
{
|
14742 |
-
"epoch": 12.587026747641815,
|
14743 |
-
"learning_rate": 1.21398936859847e-07,
|
14744 |
-
"loss": 2.4751,
|
14745 |
-
"step": 48700
|
14746 |
-
},
|
14747 |
-
{
|
14748 |
-
"epoch": 12.592195374079338,
|
14749 |
-
"learning_rate": 1.2136652405030468e-07,
|
14750 |
-
"loss": 2.4716,
|
14751 |
-
"step": 48720
|
14752 |
-
},
|
14753 |
-
{
|
14754 |
-
"epoch": 12.597364000516862,
|
14755 |
-
"learning_rate": 1.2133411124076234e-07,
|
14756 |
-
"loss": 2.4849,
|
14757 |
-
"step": 48740
|
14758 |
-
},
|
14759 |
-
{
|
14760 |
-
"epoch": 12.602532626954387,
|
14761 |
-
"learning_rate": 1.2130169843122e-07,
|
14762 |
-
"loss": 2.4891,
|
14763 |
-
"step": 48760
|
14764 |
-
},
|
14765 |
-
{
|
14766 |
-
"epoch": 12.60770125339191,
|
14767 |
-
"learning_rate": 1.212692856216777e-07,
|
14768 |
-
"loss": 2.4776,
|
14769 |
-
"step": 48780
|
14770 |
-
},
|
14771 |
-
{
|
14772 |
-
"epoch": 12.612869879829436,
|
14773 |
-
"learning_rate": 1.2123687281213535e-07,
|
14774 |
-
"loss": 2.5003,
|
14775 |
-
"step": 48800
|
14776 |
-
},
|
14777 |
-
{
|
14778 |
-
"epoch": 12.61803850626696,
|
14779 |
-
"learning_rate": 1.21204460002593e-07,
|
14780 |
-
"loss": 2.4541,
|
14781 |
-
"step": 48820
|
14782 |
-
},
|
14783 |
-
{
|
14784 |
-
"epoch": 12.623207132704485,
|
14785 |
-
"learning_rate": 1.211720471930507e-07,
|
14786 |
-
"loss": 2.3995,
|
14787 |
-
"step": 48840
|
14788 |
-
},
|
14789 |
-
{
|
14790 |
-
"epoch": 12.628375759142008,
|
14791 |
-
"learning_rate": 1.2113963438350836e-07,
|
14792 |
-
"loss": 2.4991,
|
14793 |
-
"step": 48860
|
14794 |
-
},
|
14795 |
-
{
|
14796 |
-
"epoch": 12.633544385579532,
|
14797 |
-
"learning_rate": 1.2110722157396602e-07,
|
14798 |
-
"loss": 2.4393,
|
14799 |
-
"step": 48880
|
14800 |
-
},
|
14801 |
-
{
|
14802 |
-
"epoch": 12.638713012017057,
|
14803 |
-
"learning_rate": 1.2107480876442369e-07,
|
14804 |
-
"loss": 2.4266,
|
14805 |
-
"step": 48900
|
14806 |
-
},
|
14807 |
-
{
|
14808 |
-
"epoch": 12.64388163845458,
|
14809 |
-
"learning_rate": 1.2104239595488135e-07,
|
14810 |
-
"loss": 2.4504,
|
14811 |
-
"step": 48920
|
14812 |
-
},
|
14813 |
-
{
|
14814 |
-
"epoch": 12.649050264892105,
|
14815 |
-
"learning_rate": 1.2100998314533904e-07,
|
14816 |
-
"loss": 2.4426,
|
14817 |
-
"step": 48940
|
14818 |
-
},
|
14819 |
-
{
|
14820 |
-
"epoch": 12.654218891329629,
|
14821 |
-
"learning_rate": 1.209775703357967e-07,
|
14822 |
-
"loss": 2.4279,
|
14823 |
-
"step": 48960
|
14824 |
-
},
|
14825 |
-
{
|
14826 |
-
"epoch": 12.659387517767154,
|
14827 |
-
"learning_rate": 1.2094515752625436e-07,
|
14828 |
-
"loss": 2.4973,
|
14829 |
-
"step": 48980
|
14830 |
-
},
|
14831 |
-
{
|
14832 |
-
"epoch": 12.664556144204678,
|
14833 |
-
"learning_rate": 1.2091274471671205e-07,
|
14834 |
-
"loss": 2.4447,
|
14835 |
-
"step": 49000
|
14836 |
-
},
|
14837 |
-
{
|
14838 |
-
"epoch": 12.669724770642201,
|
14839 |
-
"learning_rate": 1.208803319071697e-07,
|
14840 |
-
"loss": 2.497,
|
14841 |
-
"step": 49020
|
14842 |
-
},
|
14843 |
-
{
|
14844 |
-
"epoch": 12.674893397079726,
|
14845 |
-
"learning_rate": 1.2084791909762737e-07,
|
14846 |
-
"loss": 2.4223,
|
14847 |
-
"step": 49040
|
14848 |
-
},
|
14849 |
-
{
|
14850 |
-
"epoch": 12.68006202351725,
|
14851 |
-
"learning_rate": 1.2081550628808503e-07,
|
14852 |
-
"loss": 2.447,
|
14853 |
-
"step": 49060
|
14854 |
-
},
|
14855 |
-
{
|
14856 |
-
"epoch": 12.685230649954775,
|
14857 |
-
"learning_rate": 1.2078309347854272e-07,
|
14858 |
-
"loss": 2.4652,
|
14859 |
-
"step": 49080
|
14860 |
-
},
|
14861 |
-
{
|
14862 |
-
"epoch": 12.690399276392299,
|
14863 |
-
"learning_rate": 1.2075068066900038e-07,
|
14864 |
-
"loss": 2.4167,
|
14865 |
-
"step": 49100
|
14866 |
-
},
|
14867 |
-
{
|
14868 |
-
"epoch": 12.695567902829822,
|
14869 |
-
"learning_rate": 1.2071826785945805e-07,
|
14870 |
-
"loss": 2.4103,
|
14871 |
-
"step": 49120
|
14872 |
-
},
|
14873 |
-
{
|
14874 |
-
"epoch": 12.700736529267347,
|
14875 |
-
"learning_rate": 1.206858550499157e-07,
|
14876 |
-
"loss": 2.4225,
|
14877 |
-
"step": 49140
|
14878 |
-
},
|
14879 |
-
{
|
14880 |
-
"epoch": 12.70590515570487,
|
14881 |
-
"learning_rate": 1.206534422403734e-07,
|
14882 |
-
"loss": 2.4668,
|
14883 |
-
"step": 49160
|
14884 |
-
},
|
14885 |
-
{
|
14886 |
-
"epoch": 12.711073782142396,
|
14887 |
-
"learning_rate": 1.2062102943083106e-07,
|
14888 |
-
"loss": 2.4861,
|
14889 |
-
"step": 49180
|
14890 |
-
},
|
14891 |
-
{
|
14892 |
-
"epoch": 12.71624240857992,
|
14893 |
-
"learning_rate": 1.2058861662128872e-07,
|
14894 |
-
"loss": 2.433,
|
14895 |
-
"step": 49200
|
14896 |
-
},
|
14897 |
-
{
|
14898 |
-
"epoch": 12.721411035017445,
|
14899 |
-
"learning_rate": 1.205562038117464e-07,
|
14900 |
-
"loss": 2.4395,
|
14901 |
-
"step": 49220
|
14902 |
-
},
|
14903 |
-
{
|
14904 |
-
"epoch": 12.726579661454968,
|
14905 |
-
"learning_rate": 1.2052379100220407e-07,
|
14906 |
-
"loss": 2.4702,
|
14907 |
-
"step": 49240
|
14908 |
-
},
|
14909 |
-
{
|
14910 |
-
"epoch": 12.731748287892493,
|
14911 |
-
"learning_rate": 1.2049137819266173e-07,
|
14912 |
-
"loss": 2.4585,
|
14913 |
-
"step": 49260
|
14914 |
-
},
|
14915 |
-
{
|
14916 |
-
"epoch": 12.736916914330017,
|
14917 |
-
"learning_rate": 1.204589653831194e-07,
|
14918 |
-
"loss": 2.414,
|
14919 |
-
"step": 49280
|
14920 |
-
},
|
14921 |
-
{
|
14922 |
-
"epoch": 12.74208554076754,
|
14923 |
-
"learning_rate": 1.2042655257357706e-07,
|
14924 |
-
"loss": 2.4743,
|
14925 |
-
"step": 49300
|
14926 |
-
},
|
14927 |
-
{
|
14928 |
-
"epoch": 12.747254167205066,
|
14929 |
-
"learning_rate": 1.2039413976403474e-07,
|
14930 |
-
"loss": 2.4963,
|
14931 |
-
"step": 49320
|
14932 |
-
},
|
14933 |
-
{
|
14934 |
-
"epoch": 12.75242279364259,
|
14935 |
-
"learning_rate": 1.203617269544924e-07,
|
14936 |
-
"loss": 2.4157,
|
14937 |
-
"step": 49340
|
14938 |
-
},
|
14939 |
-
{
|
14940 |
-
"epoch": 12.757591420080114,
|
14941 |
-
"learning_rate": 1.2032931414495007e-07,
|
14942 |
-
"loss": 2.4355,
|
14943 |
-
"step": 49360
|
14944 |
-
},
|
14945 |
-
{
|
14946 |
-
"epoch": 12.762760046517638,
|
14947 |
-
"learning_rate": 1.2029690133540776e-07,
|
14948 |
-
"loss": 2.4174,
|
14949 |
-
"step": 49380
|
14950 |
-
},
|
14951 |
-
{
|
14952 |
-
"epoch": 12.767928672955161,
|
14953 |
-
"learning_rate": 1.2026448852586542e-07,
|
14954 |
-
"loss": 2.4903,
|
14955 |
-
"step": 49400
|
14956 |
-
},
|
14957 |
-
{
|
14958 |
-
"epoch": 12.773097299392687,
|
14959 |
-
"learning_rate": 1.2023207571632308e-07,
|
14960 |
-
"loss": 2.4285,
|
14961 |
-
"step": 49420
|
14962 |
-
},
|
14963 |
-
{
|
14964 |
-
"epoch": 12.77826592583021,
|
14965 |
-
"learning_rate": 1.2019966290678074e-07,
|
14966 |
-
"loss": 2.5139,
|
14967 |
-
"step": 49440
|
14968 |
-
},
|
14969 |
-
{
|
14970 |
-
"epoch": 12.783434552267735,
|
14971 |
-
"learning_rate": 1.2016725009723843e-07,
|
14972 |
-
"loss": 2.4306,
|
14973 |
-
"step": 49460
|
14974 |
-
},
|
14975 |
-
{
|
14976 |
-
"epoch": 12.788603178705259,
|
14977 |
-
"learning_rate": 1.201348372876961e-07,
|
14978 |
-
"loss": 2.4208,
|
14979 |
-
"step": 49480
|
14980 |
-
},
|
14981 |
-
{
|
14982 |
-
"epoch": 12.793771805142784,
|
14983 |
-
"learning_rate": 1.2010242447815375e-07,
|
14984 |
-
"loss": 2.4154,
|
14985 |
-
"step": 49500
|
14986 |
-
},
|
14987 |
-
{
|
14988 |
-
"epoch": 12.798940431580307,
|
14989 |
-
"learning_rate": 1.2007001166861142e-07,
|
14990 |
-
"loss": 2.4862,
|
14991 |
-
"step": 49520
|
14992 |
-
},
|
14993 |
-
{
|
14994 |
-
"epoch": 12.804109058017831,
|
14995 |
-
"learning_rate": 1.200375988590691e-07,
|
14996 |
-
"loss": 2.4184,
|
14997 |
-
"step": 49540
|
14998 |
-
},
|
14999 |
-
{
|
15000 |
-
"epoch": 12.809277684455356,
|
15001 |
-
"learning_rate": 1.2000518604952677e-07,
|
15002 |
-
"loss": 2.374,
|
15003 |
-
"step": 49560
|
15004 |
-
},
|
15005 |
-
{
|
15006 |
-
"epoch": 12.81444631089288,
|
15007 |
-
"learning_rate": 1.1997277323998443e-07,
|
15008 |
-
"loss": 2.4885,
|
15009 |
-
"step": 49580
|
15010 |
-
},
|
15011 |
-
{
|
15012 |
-
"epoch": 12.819614937330405,
|
15013 |
-
"learning_rate": 1.199403604304421e-07,
|
15014 |
-
"loss": 2.4429,
|
15015 |
-
"step": 49600
|
15016 |
-
},
|
15017 |
-
{
|
15018 |
-
"epoch": 12.824783563767928,
|
15019 |
-
"learning_rate": 1.1990794762089978e-07,
|
15020 |
-
"loss": 2.3834,
|
15021 |
-
"step": 49620
|
15022 |
-
},
|
15023 |
-
{
|
15024 |
-
"epoch": 12.829952190205454,
|
15025 |
-
"learning_rate": 1.1987553481135744e-07,
|
15026 |
-
"loss": 2.415,
|
15027 |
-
"step": 49640
|
15028 |
-
},
|
15029 |
-
{
|
15030 |
-
"epoch": 12.835120816642977,
|
15031 |
-
"learning_rate": 1.198431220018151e-07,
|
15032 |
-
"loss": 2.5206,
|
15033 |
-
"step": 49660
|
15034 |
-
},
|
15035 |
-
{
|
15036 |
-
"epoch": 12.8402894430805,
|
15037 |
-
"learning_rate": 1.198107091922728e-07,
|
15038 |
-
"loss": 2.4883,
|
15039 |
-
"step": 49680
|
15040 |
-
},
|
15041 |
-
{
|
15042 |
-
"epoch": 12.845458069518026,
|
15043 |
-
"learning_rate": 1.1977829638273045e-07,
|
15044 |
-
"loss": 2.4729,
|
15045 |
-
"step": 49700
|
15046 |
-
},
|
15047 |
-
{
|
15048 |
-
"epoch": 12.85062669595555,
|
15049 |
-
"learning_rate": 1.1974588357318811e-07,
|
15050 |
-
"loss": 2.3815,
|
15051 |
-
"step": 49720
|
15052 |
-
},
|
15053 |
-
{
|
15054 |
-
"epoch": 12.855795322393075,
|
15055 |
-
"learning_rate": 1.1971347076364578e-07,
|
15056 |
-
"loss": 2.453,
|
15057 |
-
"step": 49740
|
15058 |
-
},
|
15059 |
-
{
|
15060 |
-
"epoch": 12.860963948830598,
|
15061 |
-
"learning_rate": 1.1968105795410346e-07,
|
15062 |
-
"loss": 2.4414,
|
15063 |
-
"step": 49760
|
15064 |
-
},
|
15065 |
-
{
|
15066 |
-
"epoch": 12.866132575268123,
|
15067 |
-
"learning_rate": 1.1964864514456113e-07,
|
15068 |
-
"loss": 2.5048,
|
15069 |
-
"step": 49780
|
15070 |
-
},
|
15071 |
-
{
|
15072 |
-
"epoch": 12.871301201705647,
|
15073 |
-
"learning_rate": 1.196162323350188e-07,
|
15074 |
-
"loss": 2.4836,
|
15075 |
-
"step": 49800
|
15076 |
-
},
|
15077 |
-
{
|
15078 |
-
"epoch": 12.87646982814317,
|
15079 |
-
"learning_rate": 1.1958381952547645e-07,
|
15080 |
-
"loss": 2.4018,
|
15081 |
-
"step": 49820
|
15082 |
-
},
|
15083 |
-
{
|
15084 |
-
"epoch": 12.881638454580695,
|
15085 |
-
"learning_rate": 1.1955140671593414e-07,
|
15086 |
-
"loss": 2.4562,
|
15087 |
-
"step": 49840
|
15088 |
-
},
|
15089 |
-
{
|
15090 |
-
"epoch": 12.886807081018219,
|
15091 |
-
"learning_rate": 1.195189939063918e-07,
|
15092 |
-
"loss": 2.4431,
|
15093 |
-
"step": 49860
|
15094 |
-
},
|
15095 |
-
{
|
15096 |
-
"epoch": 12.891975707455744,
|
15097 |
-
"learning_rate": 1.1948658109684946e-07,
|
15098 |
-
"loss": 2.4554,
|
15099 |
-
"step": 49880
|
15100 |
-
},
|
15101 |
-
{
|
15102 |
-
"epoch": 12.897144333893268,
|
15103 |
-
"learning_rate": 1.1945416828730712e-07,
|
15104 |
-
"loss": 2.4341,
|
15105 |
-
"step": 49900
|
15106 |
-
},
|
15107 |
-
{
|
15108 |
-
"epoch": 12.902312960330793,
|
15109 |
-
"learning_rate": 1.194217554777648e-07,
|
15110 |
-
"loss": 2.4189,
|
15111 |
-
"step": 49920
|
15112 |
-
},
|
15113 |
-
{
|
15114 |
-
"epoch": 12.907481586768316,
|
15115 |
-
"learning_rate": 1.1938934266822247e-07,
|
15116 |
-
"loss": 2.4739,
|
15117 |
-
"step": 49940
|
15118 |
-
},
|
15119 |
-
{
|
15120 |
-
"epoch": 12.91265021320584,
|
15121 |
-
"learning_rate": 1.1935692985868014e-07,
|
15122 |
-
"loss": 2.3941,
|
15123 |
-
"step": 49960
|
15124 |
-
},
|
15125 |
-
{
|
15126 |
-
"epoch": 12.917818839643365,
|
15127 |
-
"learning_rate": 1.1932451704913782e-07,
|
15128 |
-
"loss": 2.4357,
|
15129 |
-
"step": 49980
|
15130 |
-
},
|
15131 |
-
{
|
15132 |
-
"epoch": 12.922987466080889,
|
15133 |
-
"learning_rate": 1.1929210423959549e-07,
|
15134 |
-
"loss": 2.4676,
|
15135 |
-
"step": 50000
|
15136 |
-
},
|
15137 |
-
{
|
15138 |
-
"epoch": 12.928156092518414,
|
15139 |
-
"learning_rate": 1.1925969143005315e-07,
|
15140 |
-
"loss": 2.4551,
|
15141 |
-
"step": 50020
|
15142 |
-
},
|
15143 |
-
{
|
15144 |
-
"epoch": 12.933324718955937,
|
15145 |
-
"learning_rate": 1.192272786205108e-07,
|
15146 |
-
"loss": 2.4811,
|
15147 |
-
"step": 50040
|
15148 |
-
},
|
15149 |
-
{
|
15150 |
-
"epoch": 12.938493345393463,
|
15151 |
-
"learning_rate": 1.191948658109685e-07,
|
15152 |
-
"loss": 2.4513,
|
15153 |
-
"step": 50060
|
15154 |
-
},
|
15155 |
-
{
|
15156 |
-
"epoch": 12.943661971830986,
|
15157 |
-
"learning_rate": 1.1916245300142616e-07,
|
15158 |
-
"loss": 2.4607,
|
15159 |
-
"step": 50080
|
15160 |
-
},
|
15161 |
-
{
|
15162 |
-
"epoch": 12.94883059826851,
|
15163 |
-
"learning_rate": 1.1913004019188382e-07,
|
15164 |
-
"loss": 2.482,
|
15165 |
-
"step": 50100
|
15166 |
-
},
|
15167 |
-
{
|
15168 |
-
"epoch": 12.953999224706035,
|
15169 |
-
"learning_rate": 1.1909762738234151e-07,
|
15170 |
-
"loss": 2.4465,
|
15171 |
-
"step": 50120
|
15172 |
-
},
|
15173 |
-
{
|
15174 |
-
"epoch": 12.959167851143558,
|
15175 |
-
"learning_rate": 1.1906521457279917e-07,
|
15176 |
-
"loss": 2.4672,
|
15177 |
-
"step": 50140
|
15178 |
-
},
|
15179 |
-
{
|
15180 |
-
"epoch": 12.964336477581083,
|
15181 |
-
"learning_rate": 1.1903280176325683e-07,
|
15182 |
-
"loss": 2.421,
|
15183 |
-
"step": 50160
|
15184 |
-
},
|
15185 |
-
{
|
15186 |
-
"epoch": 12.969505104018607,
|
15187 |
-
"learning_rate": 1.190003889537145e-07,
|
15188 |
-
"loss": 2.4271,
|
15189 |
-
"step": 50180
|
15190 |
-
},
|
15191 |
-
{
|
15192 |
-
"epoch": 12.974673730456132,
|
15193 |
-
"learning_rate": 1.1896797614417218e-07,
|
15194 |
-
"loss": 2.4226,
|
15195 |
-
"step": 50200
|
15196 |
-
},
|
15197 |
-
{
|
15198 |
-
"epoch": 12.979842356893656,
|
15199 |
-
"learning_rate": 1.1893556333462985e-07,
|
15200 |
-
"loss": 2.4162,
|
15201 |
-
"step": 50220
|
15202 |
-
},
|
15203 |
-
{
|
15204 |
-
"epoch": 12.985010983331179,
|
15205 |
-
"learning_rate": 1.1890315052508751e-07,
|
15206 |
-
"loss": 2.509,
|
15207 |
-
"step": 50240
|
15208 |
-
},
|
15209 |
-
{
|
15210 |
-
"epoch": 12.990179609768704,
|
15211 |
-
"learning_rate": 1.1887073771554518e-07,
|
15212 |
-
"loss": 2.4553,
|
15213 |
-
"step": 50260
|
15214 |
-
},
|
15215 |
-
{
|
15216 |
-
"epoch": 12.995348236206228,
|
15217 |
-
"learning_rate": 1.1883832490600286e-07,
|
15218 |
-
"loss": 2.4558,
|
15219 |
-
"step": 50280
|
15220 |
-
},
|
15221 |
-
{
|
15222 |
-
"epoch": 13.0,
|
15223 |
-
"eval_bleu": 12.8144,
|
15224 |
-
"eval_gen_len": 37.2979,
|
15225 |
-
"eval_loss": 2.433004140853882,
|
15226 |
-
"eval_runtime": 806.1399,
|
15227 |
-
"eval_samples_per_second": 2.136,
|
15228 |
-
"eval_steps_per_second": 1.068,
|
15229 |
-
"step": 50298
|
15230 |
-
},
|
15231 |
-
{
|
15232 |
-
"epoch": 13.000516862643753,
|
15233 |
-
"learning_rate": 1.1880591209646052e-07,
|
15234 |
-
"loss": 2.4458,
|
15235 |
-
"step": 50300
|
15236 |
-
},
|
15237 |
-
{
|
15238 |
-
"epoch": 13.005685489081277,
|
15239 |
-
"learning_rate": 1.1877349928691818e-07,
|
15240 |
-
"loss": 2.4137,
|
15241 |
-
"step": 50320
|
15242 |
-
},
|
15243 |
-
{
|
15244 |
-
"epoch": 13.0108541155188,
|
15245 |
-
"learning_rate": 1.1874108647737586e-07,
|
15246 |
-
"loss": 2.4151,
|
15247 |
-
"step": 50340
|
15248 |
-
},
|
15249 |
-
{
|
15250 |
-
"epoch": 13.016022741956325,
|
15251 |
-
"learning_rate": 1.1870867366783353e-07,
|
15252 |
-
"loss": 2.4124,
|
15253 |
-
"step": 50360
|
15254 |
-
},
|
15255 |
-
{
|
15256 |
-
"epoch": 13.021191368393849,
|
15257 |
-
"learning_rate": 1.1867626085829119e-07,
|
15258 |
-
"loss": 2.4259,
|
15259 |
-
"step": 50380
|
15260 |
-
},
|
15261 |
-
{
|
15262 |
-
"epoch": 13.026359994831374,
|
15263 |
-
"learning_rate": 1.1864384804874886e-07,
|
15264 |
-
"loss": 2.4933,
|
15265 |
-
"step": 50400
|
15266 |
-
},
|
15267 |
-
{
|
15268 |
-
"epoch": 13.031528621268897,
|
15269 |
-
"learning_rate": 1.1861143523920653e-07,
|
15270 |
-
"loss": 2.4579,
|
15271 |
-
"step": 50420
|
15272 |
-
},
|
15273 |
-
{
|
15274 |
-
"epoch": 13.036697247706423,
|
15275 |
-
"learning_rate": 1.185790224296642e-07,
|
15276 |
-
"loss": 2.4766,
|
15277 |
-
"step": 50440
|
15278 |
-
},
|
15279 |
-
{
|
15280 |
-
"epoch": 13.041865874143946,
|
15281 |
-
"learning_rate": 1.1854660962012187e-07,
|
15282 |
-
"loss": 2.4438,
|
15283 |
-
"step": 50460
|
15284 |
-
},
|
15285 |
-
{
|
15286 |
-
"epoch": 13.04703450058147,
|
15287 |
-
"learning_rate": 1.1851419681057953e-07,
|
15288 |
-
"loss": 2.4372,
|
15289 |
-
"step": 50480
|
15290 |
-
},
|
15291 |
-
{
|
15292 |
-
"epoch": 13.052203127018995,
|
15293 |
-
"learning_rate": 1.184817840010372e-07,
|
15294 |
-
"loss": 2.4097,
|
15295 |
-
"step": 50500
|
15296 |
-
},
|
15297 |
-
{
|
15298 |
-
"epoch": 13.057371753456518,
|
15299 |
-
"learning_rate": 1.1844937119149488e-07,
|
15300 |
-
"loss": 2.4199,
|
15301 |
-
"step": 50520
|
15302 |
-
},
|
15303 |
-
{
|
15304 |
-
"epoch": 13.062540379894044,
|
15305 |
-
"learning_rate": 1.1841695838195254e-07,
|
15306 |
-
"loss": 2.477,
|
15307 |
-
"step": 50540
|
15308 |
-
},
|
15309 |
-
{
|
15310 |
-
"epoch": 13.067709006331567,
|
15311 |
-
"learning_rate": 1.183845455724102e-07,
|
15312 |
-
"loss": 2.3863,
|
15313 |
-
"step": 50560
|
15314 |
-
},
|
15315 |
-
{
|
15316 |
-
"epoch": 13.072877632769092,
|
15317 |
-
"learning_rate": 1.1835213276286788e-07,
|
15318 |
-
"loss": 2.4797,
|
15319 |
-
"step": 50580
|
15320 |
-
},
|
15321 |
-
{
|
15322 |
-
"epoch": 13.078046259206616,
|
15323 |
-
"learning_rate": 1.1831971995332555e-07,
|
15324 |
-
"loss": 2.4302,
|
15325 |
-
"step": 50600
|
15326 |
-
},
|
15327 |
-
{
|
15328 |
-
"epoch": 13.08321488564414,
|
15329 |
-
"learning_rate": 1.1828730714378322e-07,
|
15330 |
-
"loss": 2.4693,
|
15331 |
-
"step": 50620
|
15332 |
-
},
|
15333 |
-
{
|
15334 |
-
"epoch": 13.088383512081665,
|
15335 |
-
"learning_rate": 1.1825489433424088e-07,
|
15336 |
-
"loss": 2.5014,
|
15337 |
-
"step": 50640
|
15338 |
-
},
|
15339 |
-
{
|
15340 |
-
"epoch": 13.093552138519188,
|
15341 |
-
"learning_rate": 1.1822248152469855e-07,
|
15342 |
-
"loss": 2.4204,
|
15343 |
-
"step": 50660
|
15344 |
-
},
|
15345 |
-
{
|
15346 |
-
"epoch": 13.098720764956713,
|
15347 |
-
"learning_rate": 1.1819006871515623e-07,
|
15348 |
-
"loss": 2.4063,
|
15349 |
-
"step": 50680
|
15350 |
-
},
|
15351 |
-
{
|
15352 |
-
"epoch": 13.103889391394237,
|
15353 |
-
"learning_rate": 1.1815765590561389e-07,
|
15354 |
-
"loss": 2.4893,
|
15355 |
-
"step": 50700
|
15356 |
-
},
|
15357 |
-
{
|
15358 |
-
"epoch": 13.109058017831762,
|
15359 |
-
"learning_rate": 1.1812524309607155e-07,
|
15360 |
-
"loss": 2.4443,
|
15361 |
-
"step": 50720
|
15362 |
-
},
|
15363 |
-
{
|
15364 |
-
"epoch": 13.114226644269285,
|
15365 |
-
"learning_rate": 1.1809283028652923e-07,
|
15366 |
-
"loss": 2.4188,
|
15367 |
-
"step": 50740
|
15368 |
-
},
|
15369 |
-
{
|
15370 |
-
"epoch": 13.119395270706809,
|
15371 |
-
"learning_rate": 1.180604174769869e-07,
|
15372 |
-
"loss": 2.4566,
|
15373 |
-
"step": 50760
|
15374 |
-
},
|
15375 |
-
{
|
15376 |
-
"epoch": 13.124563897144334,
|
15377 |
-
"learning_rate": 1.1802800466744456e-07,
|
15378 |
-
"loss": 2.4187,
|
15379 |
-
"step": 50780
|
15380 |
-
},
|
15381 |
-
{
|
15382 |
-
"epoch": 13.129732523581858,
|
15383 |
-
"learning_rate": 1.1799559185790222e-07,
|
15384 |
-
"loss": 2.4418,
|
15385 |
-
"step": 50800
|
15386 |
-
},
|
15387 |
-
{
|
15388 |
-
"epoch": 13.134901150019383,
|
15389 |
-
"learning_rate": 1.179631790483599e-07,
|
15390 |
-
"loss": 2.4705,
|
15391 |
-
"step": 50820
|
15392 |
-
},
|
15393 |
-
{
|
15394 |
-
"epoch": 13.140069776456906,
|
15395 |
-
"learning_rate": 1.1793076623881758e-07,
|
15396 |
-
"loss": 2.4966,
|
15397 |
-
"step": 50840
|
15398 |
-
},
|
15399 |
-
{
|
15400 |
-
"epoch": 13.145238402894432,
|
15401 |
-
"learning_rate": 1.1789835342927524e-07,
|
15402 |
-
"loss": 2.4615,
|
15403 |
-
"step": 50860
|
15404 |
-
},
|
15405 |
-
{
|
15406 |
-
"epoch": 13.150407029331955,
|
15407 |
-
"learning_rate": 1.1786594061973291e-07,
|
15408 |
-
"loss": 2.4513,
|
15409 |
-
"step": 50880
|
15410 |
-
},
|
15411 |
-
{
|
15412 |
-
"epoch": 13.155575655769479,
|
15413 |
-
"learning_rate": 1.1783352781019059e-07,
|
15414 |
-
"loss": 2.4163,
|
15415 |
-
"step": 50900
|
15416 |
-
},
|
15417 |
-
{
|
15418 |
-
"epoch": 13.160744282207004,
|
15419 |
-
"learning_rate": 1.1780111500064825e-07,
|
15420 |
-
"loss": 2.457,
|
15421 |
-
"step": 50920
|
15422 |
-
},
|
15423 |
-
{
|
15424 |
-
"epoch": 13.165912908644527,
|
15425 |
-
"learning_rate": 1.1776870219110591e-07,
|
15426 |
-
"loss": 2.4637,
|
15427 |
-
"step": 50940
|
15428 |
-
},
|
15429 |
-
{
|
15430 |
-
"epoch": 13.171081535082052,
|
15431 |
-
"learning_rate": 1.1773628938156359e-07,
|
15432 |
-
"loss": 2.4972,
|
15433 |
-
"step": 50960
|
15434 |
-
},
|
15435 |
-
{
|
15436 |
-
"epoch": 13.176250161519576,
|
15437 |
-
"learning_rate": 1.1770387657202126e-07,
|
15438 |
-
"loss": 2.4628,
|
15439 |
-
"step": 50980
|
15440 |
-
},
|
15441 |
-
{
|
15442 |
-
"epoch": 13.181418787957101,
|
15443 |
-
"learning_rate": 1.1767146376247892e-07,
|
15444 |
-
"loss": 2.404,
|
15445 |
-
"step": 51000
|
15446 |
-
},
|
15447 |
-
{
|
15448 |
-
"epoch": 13.186587414394625,
|
15449 |
-
"learning_rate": 1.1763905095293658e-07,
|
15450 |
-
"loss": 2.4146,
|
15451 |
-
"step": 51020
|
15452 |
-
},
|
15453 |
-
{
|
15454 |
-
"epoch": 13.191756040832148,
|
15455 |
-
"learning_rate": 1.1760663814339426e-07,
|
15456 |
-
"loss": 2.4528,
|
15457 |
-
"step": 51040
|
15458 |
-
},
|
15459 |
-
{
|
15460 |
-
"epoch": 13.196924667269673,
|
15461 |
-
"learning_rate": 1.1757422533385193e-07,
|
15462 |
-
"loss": 2.4843,
|
15463 |
-
"step": 51060
|
15464 |
-
},
|
15465 |
-
{
|
15466 |
-
"epoch": 13.202093293707197,
|
15467 |
-
"learning_rate": 1.175418125243096e-07,
|
15468 |
-
"loss": 2.4556,
|
15469 |
-
"step": 51080
|
15470 |
-
},
|
15471 |
-
{
|
15472 |
-
"epoch": 13.207261920144722,
|
15473 |
-
"learning_rate": 1.1750939971476726e-07,
|
15474 |
-
"loss": 2.4307,
|
15475 |
-
"step": 51100
|
15476 |
-
},
|
15477 |
-
{
|
15478 |
-
"epoch": 13.212430546582246,
|
15479 |
-
"learning_rate": 1.1747698690522493e-07,
|
15480 |
-
"loss": 2.4733,
|
15481 |
-
"step": 51120
|
15482 |
-
},
|
15483 |
-
{
|
15484 |
-
"epoch": 13.21759917301977,
|
15485 |
-
"learning_rate": 1.1744457409568261e-07,
|
15486 |
-
"loss": 2.4003,
|
15487 |
-
"step": 51140
|
15488 |
-
},
|
15489 |
-
{
|
15490 |
-
"epoch": 13.222767799457294,
|
15491 |
-
"learning_rate": 1.1741216128614027e-07,
|
15492 |
-
"loss": 2.4786,
|
15493 |
-
"step": 51160
|
15494 |
-
},
|
15495 |
-
{
|
15496 |
-
"epoch": 13.227936425894818,
|
15497 |
-
"learning_rate": 1.1737974847659795e-07,
|
15498 |
-
"loss": 2.4754,
|
15499 |
-
"step": 51180
|
15500 |
-
},
|
15501 |
-
{
|
15502 |
-
"epoch": 13.233105052332343,
|
15503 |
-
"learning_rate": 1.1734733566705561e-07,
|
15504 |
-
"loss": 2.4504,
|
15505 |
-
"step": 51200
|
15506 |
-
},
|
15507 |
-
{
|
15508 |
-
"epoch": 13.238273678769866,
|
15509 |
-
"learning_rate": 1.1731492285751328e-07,
|
15510 |
-
"loss": 2.3622,
|
15511 |
-
"step": 51220
|
15512 |
-
},
|
15513 |
-
{
|
15514 |
-
"epoch": 13.243442305207392,
|
15515 |
-
"learning_rate": 1.1728251004797094e-07,
|
15516 |
-
"loss": 2.4765,
|
15517 |
-
"step": 51240
|
15518 |
-
},
|
15519 |
-
{
|
15520 |
-
"epoch": 13.248610931644915,
|
15521 |
-
"learning_rate": 1.1725009723842862e-07,
|
15522 |
-
"loss": 2.5202,
|
15523 |
-
"step": 51260
|
15524 |
-
},
|
15525 |
-
{
|
15526 |
-
"epoch": 13.253779558082439,
|
15527 |
-
"learning_rate": 1.1721768442888628e-07,
|
15528 |
-
"loss": 2.4383,
|
15529 |
-
"step": 51280
|
15530 |
-
},
|
15531 |
-
{
|
15532 |
-
"epoch": 13.258948184519964,
|
15533 |
-
"learning_rate": 1.1718527161934396e-07,
|
15534 |
-
"loss": 2.4267,
|
15535 |
-
"step": 51300
|
15536 |
-
},
|
15537 |
-
{
|
15538 |
-
"epoch": 13.264116810957487,
|
15539 |
-
"learning_rate": 1.1715285880980162e-07,
|
15540 |
-
"loss": 2.4442,
|
15541 |
-
"step": 51320
|
15542 |
-
},
|
15543 |
-
{
|
15544 |
-
"epoch": 13.269285437395013,
|
15545 |
-
"learning_rate": 1.171204460002593e-07,
|
15546 |
-
"loss": 2.4723,
|
15547 |
-
"step": 51340
|
15548 |
-
},
|
15549 |
-
{
|
15550 |
-
"epoch": 13.274454063832536,
|
15551 |
-
"learning_rate": 1.1708803319071696e-07,
|
15552 |
-
"loss": 2.4552,
|
15553 |
-
"step": 51360
|
15554 |
-
},
|
15555 |
-
{
|
15556 |
-
"epoch": 13.279622690270061,
|
15557 |
-
"learning_rate": 1.1705562038117463e-07,
|
15558 |
-
"loss": 2.4261,
|
15559 |
-
"step": 51380
|
15560 |
-
},
|
15561 |
-
{
|
15562 |
-
"epoch": 13.284791316707585,
|
15563 |
-
"learning_rate": 1.1702320757163229e-07,
|
15564 |
-
"loss": 2.3728,
|
15565 |
-
"step": 51400
|
15566 |
-
},
|
15567 |
-
{
|
15568 |
-
"epoch": 13.28995994314511,
|
15569 |
-
"learning_rate": 1.1699079476208997e-07,
|
15570 |
-
"loss": 2.4434,
|
15571 |
-
"step": 51420
|
15572 |
-
},
|
15573 |
-
{
|
15574 |
-
"epoch": 13.295128569582634,
|
15575 |
-
"learning_rate": 1.1695838195254763e-07,
|
15576 |
-
"loss": 2.433,
|
15577 |
-
"step": 51440
|
15578 |
-
},
|
15579 |
-
{
|
15580 |
-
"epoch": 13.300297196020157,
|
15581 |
-
"learning_rate": 1.169259691430053e-07,
|
15582 |
-
"loss": 2.4091,
|
15583 |
-
"step": 51460
|
15584 |
-
},
|
15585 |
-
{
|
15586 |
-
"epoch": 13.305465822457682,
|
15587 |
-
"learning_rate": 1.1689355633346298e-07,
|
15588 |
-
"loss": 2.418,
|
15589 |
-
"step": 51480
|
15590 |
-
},
|
15591 |
-
{
|
15592 |
-
"epoch": 13.310634448895206,
|
15593 |
-
"learning_rate": 1.1686114352392064e-07,
|
15594 |
-
"loss": 2.4226,
|
15595 |
-
"step": 51500
|
15596 |
-
},
|
15597 |
-
{
|
15598 |
-
"epoch": 13.315803075332731,
|
15599 |
-
"learning_rate": 1.168287307143783e-07,
|
15600 |
-
"loss": 2.4603,
|
15601 |
-
"step": 51520
|
15602 |
-
},
|
15603 |
-
{
|
15604 |
-
"epoch": 13.320971701770254,
|
15605 |
-
"learning_rate": 1.1679631790483598e-07,
|
15606 |
-
"loss": 2.4939,
|
15607 |
-
"step": 51540
|
15608 |
-
},
|
15609 |
-
{
|
15610 |
-
"epoch": 13.326140328207778,
|
15611 |
-
"learning_rate": 1.1676390509529365e-07,
|
15612 |
-
"loss": 2.444,
|
15613 |
-
"step": 51560
|
15614 |
-
},
|
15615 |
-
{
|
15616 |
-
"epoch": 13.331308954645303,
|
15617 |
-
"learning_rate": 1.1673149228575132e-07,
|
15618 |
-
"loss": 2.5078,
|
15619 |
-
"step": 51580
|
15620 |
-
},
|
15621 |
-
{
|
15622 |
-
"epoch": 13.336477581082827,
|
15623 |
-
"learning_rate": 1.1669907947620898e-07,
|
15624 |
-
"loss": 2.5158,
|
15625 |
-
"step": 51600
|
15626 |
-
},
|
15627 |
-
{
|
15628 |
-
"epoch": 13.341646207520352,
|
15629 |
-
"learning_rate": 1.1666666666666665e-07,
|
15630 |
-
"loss": 2.5183,
|
15631 |
-
"step": 51620
|
15632 |
-
},
|
15633 |
-
{
|
15634 |
-
"epoch": 13.346814833957875,
|
15635 |
-
"learning_rate": 1.1663425385712433e-07,
|
15636 |
-
"loss": 2.3932,
|
15637 |
-
"step": 51640
|
15638 |
-
},
|
15639 |
-
{
|
15640 |
-
"epoch": 13.3519834603954,
|
15641 |
-
"learning_rate": 1.1660184104758199e-07,
|
15642 |
-
"loss": 2.413,
|
15643 |
-
"step": 51660
|
15644 |
-
},
|
15645 |
-
{
|
15646 |
-
"epoch": 13.357152086832924,
|
15647 |
-
"learning_rate": 1.1656942823803965e-07,
|
15648 |
-
"loss": 2.4563,
|
15649 |
-
"step": 51680
|
15650 |
-
},
|
15651 |
-
{
|
15652 |
-
"epoch": 13.362320713270448,
|
15653 |
-
"learning_rate": 1.1653701542849733e-07,
|
15654 |
-
"loss": 2.4141,
|
15655 |
-
"step": 51700
|
15656 |
-
},
|
15657 |
-
{
|
15658 |
-
"epoch": 13.367489339707973,
|
15659 |
-
"learning_rate": 1.16504602618955e-07,
|
15660 |
-
"loss": 2.486,
|
15661 |
-
"step": 51720
|
15662 |
-
},
|
15663 |
-
{
|
15664 |
-
"epoch": 13.372657966145496,
|
15665 |
-
"learning_rate": 1.1647218980941266e-07,
|
15666 |
-
"loss": 2.444,
|
15667 |
-
"step": 51740
|
15668 |
-
},
|
15669 |
-
{
|
15670 |
-
"epoch": 13.377826592583022,
|
15671 |
-
"learning_rate": 1.1643977699987033e-07,
|
15672 |
-
"loss": 2.4608,
|
15673 |
-
"step": 51760
|
15674 |
-
},
|
15675 |
-
{
|
15676 |
-
"epoch": 13.382995219020545,
|
15677 |
-
"learning_rate": 1.1640736419032801e-07,
|
15678 |
-
"loss": 2.4173,
|
15679 |
-
"step": 51780
|
15680 |
-
},
|
15681 |
-
{
|
15682 |
-
"epoch": 13.38816384545807,
|
15683 |
-
"learning_rate": 1.1637495138078568e-07,
|
15684 |
-
"loss": 2.4247,
|
15685 |
-
"step": 51800
|
15686 |
-
},
|
15687 |
-
{
|
15688 |
-
"epoch": 13.393332471895594,
|
15689 |
-
"learning_rate": 1.1634253857124334e-07,
|
15690 |
-
"loss": 2.4383,
|
15691 |
-
"step": 51820
|
15692 |
-
},
|
15693 |
-
{
|
15694 |
-
"epoch": 13.398501098333117,
|
15695 |
-
"learning_rate": 1.16310125761701e-07,
|
15696 |
-
"loss": 2.4455,
|
15697 |
-
"step": 51840
|
15698 |
-
},
|
15699 |
-
{
|
15700 |
-
"epoch": 13.403669724770642,
|
15701 |
-
"learning_rate": 1.1627771295215869e-07,
|
15702 |
-
"loss": 2.4771,
|
15703 |
-
"step": 51860
|
15704 |
-
},
|
15705 |
-
{
|
15706 |
-
"epoch": 13.408838351208166,
|
15707 |
-
"learning_rate": 1.1624530014261635e-07,
|
15708 |
-
"loss": 2.4686,
|
15709 |
-
"step": 51880
|
15710 |
-
},
|
15711 |
-
{
|
15712 |
-
"epoch": 13.414006977645691,
|
15713 |
-
"learning_rate": 1.1621288733307401e-07,
|
15714 |
-
"loss": 2.4417,
|
15715 |
-
"step": 51900
|
15716 |
-
},
|
15717 |
-
{
|
15718 |
-
"epoch": 13.419175604083215,
|
15719 |
-
"learning_rate": 1.1618047452353167e-07,
|
15720 |
-
"loss": 2.4765,
|
15721 |
-
"step": 51920
|
15722 |
-
},
|
15723 |
-
{
|
15724 |
-
"epoch": 13.42434423052074,
|
15725 |
-
"learning_rate": 1.1614806171398936e-07,
|
15726 |
-
"loss": 2.4501,
|
15727 |
-
"step": 51940
|
15728 |
-
},
|
15729 |
-
{
|
15730 |
-
"epoch": 13.429512856958263,
|
15731 |
-
"learning_rate": 1.1611564890444702e-07,
|
15732 |
-
"loss": 2.467,
|
15733 |
-
"step": 51960
|
15734 |
-
},
|
15735 |
-
{
|
15736 |
-
"epoch": 13.434681483395787,
|
15737 |
-
"learning_rate": 1.1608323609490469e-07,
|
15738 |
-
"loss": 2.4404,
|
15739 |
-
"step": 51980
|
15740 |
-
},
|
15741 |
-
{
|
15742 |
-
"epoch": 13.439850109833312,
|
15743 |
-
"learning_rate": 1.1605082328536235e-07,
|
15744 |
-
"loss": 2.4413,
|
15745 |
-
"step": 52000
|
15746 |
-
},
|
15747 |
-
{
|
15748 |
-
"epoch": 13.445018736270836,
|
15749 |
-
"learning_rate": 1.1601841047582004e-07,
|
15750 |
-
"loss": 2.4529,
|
15751 |
-
"step": 52020
|
15752 |
-
},
|
15753 |
-
{
|
15754 |
-
"epoch": 13.45018736270836,
|
15755 |
-
"learning_rate": 1.159859976662777e-07,
|
15756 |
-
"loss": 2.4719,
|
15757 |
-
"step": 52040
|
15758 |
-
},
|
15759 |
-
{
|
15760 |
-
"epoch": 13.455355989145884,
|
15761 |
-
"learning_rate": 1.1595358485673536e-07,
|
15762 |
-
"loss": 2.3729,
|
15763 |
-
"step": 52060
|
15764 |
-
},
|
15765 |
-
{
|
15766 |
-
"epoch": 13.46052461558341,
|
15767 |
-
"learning_rate": 1.1592117204719305e-07,
|
15768 |
-
"loss": 2.4296,
|
15769 |
-
"step": 52080
|
15770 |
-
},
|
15771 |
-
{
|
15772 |
-
"epoch": 13.465693242020933,
|
15773 |
-
"learning_rate": 1.1588875923765071e-07,
|
15774 |
-
"loss": 2.4939,
|
15775 |
-
"step": 52100
|
15776 |
-
},
|
15777 |
-
{
|
15778 |
-
"epoch": 13.470861868458456,
|
15779 |
-
"learning_rate": 1.1585634642810837e-07,
|
15780 |
-
"loss": 2.4214,
|
15781 |
-
"step": 52120
|
15782 |
-
},
|
15783 |
-
{
|
15784 |
-
"epoch": 13.476030494895982,
|
15785 |
-
"learning_rate": 1.1582393361856603e-07,
|
15786 |
-
"loss": 2.4435,
|
15787 |
-
"step": 52140
|
15788 |
-
},
|
15789 |
-
{
|
15790 |
-
"epoch": 13.481199121333505,
|
15791 |
-
"learning_rate": 1.1579152080902372e-07,
|
15792 |
-
"loss": 2.4724,
|
15793 |
-
"step": 52160
|
15794 |
-
},
|
15795 |
-
{
|
15796 |
-
"epoch": 13.48636774777103,
|
15797 |
-
"learning_rate": 1.1575910799948138e-07,
|
15798 |
-
"loss": 2.4219,
|
15799 |
-
"step": 52180
|
15800 |
-
},
|
15801 |
-
{
|
15802 |
-
"epoch": 13.491536374208554,
|
15803 |
-
"learning_rate": 1.1572669518993905e-07,
|
15804 |
-
"loss": 2.4166,
|
15805 |
-
"step": 52200
|
15806 |
-
},
|
15807 |
-
{
|
15808 |
-
"epoch": 13.49670500064608,
|
15809 |
-
"learning_rate": 1.1569428238039671e-07,
|
15810 |
-
"loss": 2.456,
|
15811 |
-
"step": 52220
|
15812 |
-
},
|
15813 |
-
{
|
15814 |
-
"epoch": 13.501873627083603,
|
15815 |
-
"learning_rate": 1.156618695708544e-07,
|
15816 |
-
"loss": 2.4335,
|
15817 |
-
"step": 52240
|
15818 |
-
},
|
15819 |
-
{
|
15820 |
-
"epoch": 13.507042253521126,
|
15821 |
-
"learning_rate": 1.1562945676131206e-07,
|
15822 |
-
"loss": 2.4549,
|
15823 |
-
"step": 52260
|
15824 |
-
},
|
15825 |
-
{
|
15826 |
-
"epoch": 13.512210879958651,
|
15827 |
-
"learning_rate": 1.1559704395176972e-07,
|
15828 |
-
"loss": 2.4274,
|
15829 |
-
"step": 52280
|
15830 |
-
},
|
15831 |
-
{
|
15832 |
-
"epoch": 13.517379506396175,
|
15833 |
-
"learning_rate": 1.1556463114222738e-07,
|
15834 |
-
"loss": 2.4509,
|
15835 |
-
"step": 52300
|
15836 |
-
},
|
15837 |
-
{
|
15838 |
-
"epoch": 13.5225481328337,
|
15839 |
-
"learning_rate": 1.1553221833268507e-07,
|
15840 |
-
"loss": 2.4466,
|
15841 |
-
"step": 52320
|
15842 |
-
},
|
15843 |
-
{
|
15844 |
-
"epoch": 13.527716759271224,
|
15845 |
-
"learning_rate": 1.1549980552314273e-07,
|
15846 |
-
"loss": 2.486,
|
15847 |
-
"step": 52340
|
15848 |
-
},
|
15849 |
-
{
|
15850 |
-
"epoch": 13.532885385708749,
|
15851 |
-
"learning_rate": 1.1546739271360039e-07,
|
15852 |
-
"loss": 2.4423,
|
15853 |
-
"step": 52360
|
15854 |
-
},
|
15855 |
-
{
|
15856 |
-
"epoch": 13.538054012146272,
|
15857 |
-
"learning_rate": 1.1543497990405808e-07,
|
15858 |
-
"loss": 2.3907,
|
15859 |
-
"step": 52380
|
15860 |
-
},
|
15861 |
-
{
|
15862 |
-
"epoch": 13.543222638583796,
|
15863 |
-
"learning_rate": 1.1540256709451574e-07,
|
15864 |
-
"loss": 2.4323,
|
15865 |
-
"step": 52400
|
15866 |
-
},
|
15867 |
-
{
|
15868 |
-
"epoch": 13.548391265021321,
|
15869 |
-
"learning_rate": 1.153701542849734e-07,
|
15870 |
-
"loss": 2.4375,
|
15871 |
-
"step": 52420
|
15872 |
-
},
|
15873 |
-
{
|
15874 |
-
"epoch": 13.553559891458844,
|
15875 |
-
"learning_rate": 1.1533774147543107e-07,
|
15876 |
-
"loss": 2.4197,
|
15877 |
-
"step": 52440
|
15878 |
-
},
|
15879 |
-
{
|
15880 |
-
"epoch": 13.55872851789637,
|
15881 |
-
"learning_rate": 1.1530532866588876e-07,
|
15882 |
-
"loss": 2.4902,
|
15883 |
-
"step": 52460
|
15884 |
-
},
|
15885 |
-
{
|
15886 |
-
"epoch": 13.563897144333893,
|
15887 |
-
"learning_rate": 1.1527291585634642e-07,
|
15888 |
-
"loss": 2.4065,
|
15889 |
-
"step": 52480
|
15890 |
-
},
|
15891 |
-
{
|
15892 |
-
"epoch": 13.569065770771417,
|
15893 |
-
"learning_rate": 1.1524050304680408e-07,
|
15894 |
-
"loss": 2.4606,
|
15895 |
-
"step": 52500
|
15896 |
-
},
|
15897 |
-
{
|
15898 |
-
"epoch": 13.574234397208942,
|
15899 |
-
"learning_rate": 1.1520809023726174e-07,
|
15900 |
-
"loss": 2.4576,
|
15901 |
-
"step": 52520
|
15902 |
-
},
|
15903 |
-
{
|
15904 |
-
"epoch": 13.579403023646465,
|
15905 |
-
"learning_rate": 1.1517567742771943e-07,
|
15906 |
-
"loss": 2.4596,
|
15907 |
-
"step": 52540
|
15908 |
-
},
|
15909 |
-
{
|
15910 |
-
"epoch": 13.58457165008399,
|
15911 |
-
"learning_rate": 1.1514326461817709e-07,
|
15912 |
-
"loss": 2.4678,
|
15913 |
-
"step": 52560
|
15914 |
-
},
|
15915 |
-
{
|
15916 |
-
"epoch": 13.589740276521514,
|
15917 |
-
"learning_rate": 1.1511085180863475e-07,
|
15918 |
-
"loss": 2.4333,
|
15919 |
-
"step": 52580
|
15920 |
-
},
|
15921 |
-
{
|
15922 |
-
"epoch": 13.59490890295904,
|
15923 |
-
"learning_rate": 1.1507843899909241e-07,
|
15924 |
-
"loss": 2.4255,
|
15925 |
-
"step": 52600
|
15926 |
-
},
|
15927 |
-
{
|
15928 |
-
"epoch": 13.600077529396563,
|
15929 |
-
"learning_rate": 1.150460261895501e-07,
|
15930 |
-
"loss": 2.414,
|
15931 |
-
"step": 52620
|
15932 |
-
},
|
15933 |
-
{
|
15934 |
-
"epoch": 13.605246155834088,
|
15935 |
-
"learning_rate": 1.1501361338000777e-07,
|
15936 |
-
"loss": 2.4667,
|
15937 |
-
"step": 52640
|
15938 |
-
},
|
15939 |
-
{
|
15940 |
-
"epoch": 13.610414782271612,
|
15941 |
-
"learning_rate": 1.1498120057046543e-07,
|
15942 |
-
"loss": 2.4574,
|
15943 |
-
"step": 52660
|
15944 |
-
},
|
15945 |
-
{
|
15946 |
-
"epoch": 13.615583408709135,
|
15947 |
-
"learning_rate": 1.1494878776092312e-07,
|
15948 |
-
"loss": 2.4472,
|
15949 |
-
"step": 52680
|
15950 |
-
},
|
15951 |
-
{
|
15952 |
-
"epoch": 13.62075203514666,
|
15953 |
-
"learning_rate": 1.1491637495138078e-07,
|
15954 |
-
"loss": 2.4838,
|
15955 |
-
"step": 52700
|
15956 |
-
},
|
15957 |
-
{
|
15958 |
-
"epoch": 13.625920661584184,
|
15959 |
-
"learning_rate": 1.1488396214183844e-07,
|
15960 |
-
"loss": 2.4244,
|
15961 |
-
"step": 52720
|
15962 |
-
},
|
15963 |
-
{
|
15964 |
-
"epoch": 13.631089288021709,
|
15965 |
-
"learning_rate": 1.148515493322961e-07,
|
15966 |
-
"loss": 2.4486,
|
15967 |
-
"step": 52740
|
15968 |
-
},
|
15969 |
-
{
|
15970 |
-
"epoch": 13.636257914459232,
|
15971 |
-
"learning_rate": 1.1481913652275379e-07,
|
15972 |
-
"loss": 2.4595,
|
15973 |
-
"step": 52760
|
15974 |
-
},
|
15975 |
-
{
|
15976 |
-
"epoch": 13.641426540896756,
|
15977 |
-
"learning_rate": 1.1478672371321145e-07,
|
15978 |
-
"loss": 2.4267,
|
15979 |
-
"step": 52780
|
15980 |
-
},
|
15981 |
-
{
|
15982 |
-
"epoch": 13.646595167334281,
|
15983 |
-
"learning_rate": 1.1475431090366911e-07,
|
15984 |
-
"loss": 2.4348,
|
15985 |
-
"step": 52800
|
15986 |
-
},
|
15987 |
-
{
|
15988 |
-
"epoch": 13.651763793771805,
|
15989 |
-
"learning_rate": 1.1472189809412677e-07,
|
15990 |
-
"loss": 2.4963,
|
15991 |
-
"step": 52820
|
15992 |
-
},
|
15993 |
-
{
|
15994 |
-
"epoch": 13.65693242020933,
|
15995 |
-
"learning_rate": 1.1468948528458446e-07,
|
15996 |
-
"loss": 2.3726,
|
15997 |
-
"step": 52840
|
15998 |
-
},
|
15999 |
-
{
|
16000 |
-
"epoch": 13.662101046646853,
|
16001 |
-
"learning_rate": 1.1465707247504213e-07,
|
16002 |
-
"loss": 2.3944,
|
16003 |
-
"step": 52860
|
16004 |
-
},
|
16005 |
-
{
|
16006 |
-
"epoch": 13.667269673084379,
|
16007 |
-
"learning_rate": 1.1462465966549979e-07,
|
16008 |
-
"loss": 2.5007,
|
16009 |
-
"step": 52880
|
16010 |
-
},
|
16011 |
-
{
|
16012 |
-
"epoch": 13.672438299521902,
|
16013 |
-
"learning_rate": 1.1459224685595745e-07,
|
16014 |
-
"loss": 2.4834,
|
16015 |
-
"step": 52900
|
16016 |
-
},
|
16017 |
-
{
|
16018 |
-
"epoch": 13.677606925959426,
|
16019 |
-
"learning_rate": 1.1455983404641514e-07,
|
16020 |
-
"loss": 2.3963,
|
16021 |
-
"step": 52920
|
16022 |
-
},
|
16023 |
-
{
|
16024 |
-
"epoch": 13.68277555239695,
|
16025 |
-
"learning_rate": 1.1452742123687283e-07,
|
16026 |
-
"loss": 2.4304,
|
16027 |
-
"step": 52940
|
16028 |
-
},
|
16029 |
-
{
|
16030 |
-
"epoch": 13.687944178834474,
|
16031 |
-
"learning_rate": 1.1449500842733046e-07,
|
16032 |
-
"loss": 2.4219,
|
16033 |
-
"step": 52960
|
16034 |
-
},
|
16035 |
-
{
|
16036 |
-
"epoch": 13.693112805272,
|
16037 |
-
"learning_rate": 1.1446259561778815e-07,
|
16038 |
-
"loss": 2.5008,
|
16039 |
-
"step": 52980
|
16040 |
-
},
|
16041 |
-
{
|
16042 |
-
"epoch": 13.698281431709523,
|
16043 |
-
"learning_rate": 1.1443018280824581e-07,
|
16044 |
-
"loss": 2.4243,
|
16045 |
-
"step": 53000
|
16046 |
-
},
|
16047 |
-
{
|
16048 |
-
"epoch": 13.703450058147048,
|
16049 |
-
"learning_rate": 1.143977699987035e-07,
|
16050 |
-
"loss": 2.462,
|
16051 |
-
"step": 53020
|
16052 |
-
},
|
16053 |
-
{
|
16054 |
-
"epoch": 13.708618684584572,
|
16055 |
-
"learning_rate": 1.1436535718916113e-07,
|
16056 |
-
"loss": 2.4512,
|
16057 |
-
"step": 53040
|
16058 |
-
},
|
16059 |
-
{
|
16060 |
-
"epoch": 13.713787311022095,
|
16061 |
-
"learning_rate": 1.1433294437961882e-07,
|
16062 |
-
"loss": 2.508,
|
16063 |
-
"step": 53060
|
16064 |
-
},
|
16065 |
-
{
|
16066 |
-
"epoch": 13.71895593745962,
|
16067 |
-
"learning_rate": 1.1430053157007649e-07,
|
16068 |
-
"loss": 2.4558,
|
16069 |
-
"step": 53080
|
16070 |
-
},
|
16071 |
-
{
|
16072 |
-
"epoch": 13.724124563897144,
|
16073 |
-
"learning_rate": 1.1426811876053417e-07,
|
16074 |
-
"loss": 2.4658,
|
16075 |
-
"step": 53100
|
16076 |
-
},
|
16077 |
-
{
|
16078 |
-
"epoch": 13.729293190334669,
|
16079 |
-
"learning_rate": 1.1423570595099181e-07,
|
16080 |
-
"loss": 2.3885,
|
16081 |
-
"step": 53120
|
16082 |
-
},
|
16083 |
-
{
|
16084 |
-
"epoch": 13.734461816772193,
|
16085 |
-
"learning_rate": 1.142032931414495e-07,
|
16086 |
-
"loss": 2.44,
|
16087 |
-
"step": 53140
|
16088 |
-
},
|
16089 |
-
{
|
16090 |
-
"epoch": 13.739630443209718,
|
16091 |
-
"learning_rate": 1.1417088033190716e-07,
|
16092 |
-
"loss": 2.4517,
|
16093 |
-
"step": 53160
|
16094 |
-
},
|
16095 |
-
{
|
16096 |
-
"epoch": 13.744799069647241,
|
16097 |
-
"learning_rate": 1.1413846752236485e-07,
|
16098 |
-
"loss": 2.4766,
|
16099 |
-
"step": 53180
|
16100 |
-
},
|
16101 |
-
{
|
16102 |
-
"epoch": 13.749967696084765,
|
16103 |
-
"learning_rate": 1.1410605471282248e-07,
|
16104 |
-
"loss": 2.447,
|
16105 |
-
"step": 53200
|
16106 |
-
},
|
16107 |
-
{
|
16108 |
-
"epoch": 13.75513632252229,
|
16109 |
-
"learning_rate": 1.1407364190328017e-07,
|
16110 |
-
"loss": 2.4432,
|
16111 |
-
"step": 53220
|
16112 |
-
},
|
16113 |
-
{
|
16114 |
-
"epoch": 13.760304948959813,
|
16115 |
-
"learning_rate": 1.1404122909373786e-07,
|
16116 |
-
"loss": 2.4773,
|
16117 |
-
"step": 53240
|
16118 |
-
},
|
16119 |
-
{
|
16120 |
-
"epoch": 13.765473575397339,
|
16121 |
-
"learning_rate": 1.140088162841955e-07,
|
16122 |
-
"loss": 2.455,
|
16123 |
-
"step": 53260
|
16124 |
-
},
|
16125 |
-
{
|
16126 |
-
"epoch": 13.770642201834862,
|
16127 |
-
"learning_rate": 1.1397640347465318e-07,
|
16128 |
-
"loss": 2.4651,
|
16129 |
-
"step": 53280
|
16130 |
-
},
|
16131 |
-
{
|
16132 |
-
"epoch": 13.775810828272387,
|
16133 |
-
"learning_rate": 1.1394399066511085e-07,
|
16134 |
-
"loss": 2.4309,
|
16135 |
-
"step": 53300
|
16136 |
-
},
|
16137 |
-
{
|
16138 |
-
"epoch": 13.780979454709911,
|
16139 |
-
"learning_rate": 1.1391157785556853e-07,
|
16140 |
-
"loss": 2.4539,
|
16141 |
-
"step": 53320
|
16142 |
-
},
|
16143 |
-
{
|
16144 |
-
"epoch": 13.786148081147434,
|
16145 |
-
"learning_rate": 1.1387916504602617e-07,
|
16146 |
-
"loss": 2.4259,
|
16147 |
-
"step": 53340
|
16148 |
-
},
|
16149 |
-
{
|
16150 |
-
"epoch": 13.79131670758496,
|
16151 |
-
"learning_rate": 1.1384675223648386e-07,
|
16152 |
-
"loss": 2.357,
|
16153 |
-
"step": 53360
|
16154 |
-
},
|
16155 |
-
{
|
16156 |
-
"epoch": 13.796485334022483,
|
16157 |
-
"learning_rate": 1.1381433942694152e-07,
|
16158 |
-
"loss": 2.4122,
|
16159 |
-
"step": 53380
|
16160 |
-
},
|
16161 |
-
{
|
16162 |
-
"epoch": 13.801653960460008,
|
16163 |
-
"learning_rate": 1.1378192661739921e-07,
|
16164 |
-
"loss": 2.435,
|
16165 |
-
"step": 53400
|
16166 |
-
},
|
16167 |
-
{
|
16168 |
-
"epoch": 13.806822586897532,
|
16169 |
-
"learning_rate": 1.1374951380785684e-07,
|
16170 |
-
"loss": 2.4501,
|
16171 |
-
"step": 53420
|
16172 |
-
},
|
16173 |
-
{
|
16174 |
-
"epoch": 13.811991213335055,
|
16175 |
-
"learning_rate": 1.1371710099831453e-07,
|
16176 |
-
"loss": 2.4191,
|
16177 |
-
"step": 53440
|
16178 |
-
},
|
16179 |
-
{
|
16180 |
-
"epoch": 13.81715983977258,
|
16181 |
-
"learning_rate": 1.1368468818877219e-07,
|
16182 |
-
"loss": 2.4168,
|
16183 |
-
"step": 53460
|
16184 |
-
},
|
16185 |
-
{
|
16186 |
-
"epoch": 13.822328466210104,
|
16187 |
-
"learning_rate": 1.1365227537922988e-07,
|
16188 |
-
"loss": 2.4114,
|
16189 |
-
"step": 53480
|
16190 |
-
},
|
16191 |
-
{
|
16192 |
-
"epoch": 13.82749709264763,
|
16193 |
-
"learning_rate": 1.1361986256968752e-07,
|
16194 |
-
"loss": 2.4437,
|
16195 |
-
"step": 53500
|
16196 |
-
},
|
16197 |
-
{
|
16198 |
-
"epoch": 13.832665719085153,
|
16199 |
-
"learning_rate": 1.135874497601452e-07,
|
16200 |
-
"loss": 2.5263,
|
16201 |
-
"step": 53520
|
16202 |
-
},
|
16203 |
-
{
|
16204 |
-
"epoch": 13.837834345522678,
|
16205 |
-
"learning_rate": 1.1355503695060289e-07,
|
16206 |
-
"loss": 2.3909,
|
16207 |
-
"step": 53540
|
16208 |
-
},
|
16209 |
-
{
|
16210 |
-
"epoch": 13.843002971960201,
|
16211 |
-
"learning_rate": 1.1352262414106056e-07,
|
16212 |
-
"loss": 2.4661,
|
16213 |
-
"step": 53560
|
16214 |
-
},
|
16215 |
-
{
|
16216 |
-
"epoch": 13.848171598397727,
|
16217 |
-
"learning_rate": 1.1349021133151822e-07,
|
16218 |
-
"loss": 2.3978,
|
16219 |
-
"step": 53580
|
16220 |
-
},
|
16221 |
-
{
|
16222 |
-
"epoch": 13.85334022483525,
|
16223 |
-
"learning_rate": 1.1345779852197588e-07,
|
16224 |
-
"loss": 2.4506,
|
16225 |
-
"step": 53600
|
16226 |
-
},
|
16227 |
-
{
|
16228 |
-
"epoch": 13.858508851272774,
|
16229 |
-
"learning_rate": 1.1342538571243357e-07,
|
16230 |
-
"loss": 2.4424,
|
16231 |
-
"step": 53620
|
16232 |
-
},
|
16233 |
-
{
|
16234 |
-
"epoch": 13.863677477710299,
|
16235 |
-
"learning_rate": 1.1339297290289123e-07,
|
16236 |
-
"loss": 2.4247,
|
16237 |
-
"step": 53640
|
16238 |
-
},
|
16239 |
-
{
|
16240 |
-
"epoch": 13.868846104147822,
|
16241 |
-
"learning_rate": 1.1336056009334889e-07,
|
16242 |
-
"loss": 2.4441,
|
16243 |
-
"step": 53660
|
16244 |
-
},
|
16245 |
-
{
|
16246 |
-
"epoch": 13.874014730585348,
|
16247 |
-
"learning_rate": 1.1332814728380655e-07,
|
16248 |
-
"loss": 2.4742,
|
16249 |
-
"step": 53680
|
16250 |
-
},
|
16251 |
-
{
|
16252 |
-
"epoch": 13.879183357022871,
|
16253 |
-
"learning_rate": 1.1329573447426424e-07,
|
16254 |
-
"loss": 2.4507,
|
16255 |
-
"step": 53700
|
16256 |
-
},
|
16257 |
-
{
|
16258 |
-
"epoch": 13.884351983460395,
|
16259 |
-
"learning_rate": 1.132633216647219e-07,
|
16260 |
-
"loss": 2.4422,
|
16261 |
-
"step": 53720
|
16262 |
-
},
|
16263 |
-
{
|
16264 |
-
"epoch": 13.88952060989792,
|
16265 |
-
"learning_rate": 1.1323090885517956e-07,
|
16266 |
-
"loss": 2.5032,
|
16267 |
-
"step": 53740
|
16268 |
-
},
|
16269 |
-
{
|
16270 |
-
"epoch": 13.894689236335443,
|
16271 |
-
"learning_rate": 1.1319849604563723e-07,
|
16272 |
-
"loss": 2.4601,
|
16273 |
-
"step": 53760
|
16274 |
-
},
|
16275 |
-
{
|
16276 |
-
"epoch": 13.899857862772969,
|
16277 |
-
"learning_rate": 1.1316608323609492e-07,
|
16278 |
-
"loss": 2.3983,
|
16279 |
-
"step": 53780
|
16280 |
-
},
|
16281 |
-
{
|
16282 |
-
"epoch": 13.905026489210492,
|
16283 |
-
"learning_rate": 1.1313367042655258e-07,
|
16284 |
-
"loss": 2.3913,
|
16285 |
-
"step": 53800
|
16286 |
-
},
|
16287 |
-
{
|
16288 |
-
"epoch": 13.910195115648017,
|
16289 |
-
"learning_rate": 1.1310125761701024e-07,
|
16290 |
-
"loss": 2.4417,
|
16291 |
-
"step": 53820
|
16292 |
-
},
|
16293 |
-
{
|
16294 |
-
"epoch": 13.91536374208554,
|
16295 |
-
"learning_rate": 1.1306884480746793e-07,
|
16296 |
-
"loss": 2.422,
|
16297 |
-
"step": 53840
|
16298 |
-
},
|
16299 |
-
{
|
16300 |
-
"epoch": 13.920532368523064,
|
16301 |
-
"learning_rate": 1.1303643199792559e-07,
|
16302 |
-
"loss": 2.4629,
|
16303 |
-
"step": 53860
|
16304 |
-
},
|
16305 |
-
{
|
16306 |
-
"epoch": 13.92570099496059,
|
16307 |
-
"learning_rate": 1.1300401918838325e-07,
|
16308 |
-
"loss": 2.4489,
|
16309 |
-
"step": 53880
|
16310 |
-
},
|
16311 |
-
{
|
16312 |
-
"epoch": 13.930869621398113,
|
16313 |
-
"learning_rate": 1.1297160637884091e-07,
|
16314 |
-
"loss": 2.4243,
|
16315 |
-
"step": 53900
|
16316 |
-
},
|
16317 |
-
{
|
16318 |
-
"epoch": 13.936038247835638,
|
16319 |
-
"learning_rate": 1.129391935692986e-07,
|
16320 |
-
"loss": 2.4241,
|
16321 |
-
"step": 53920
|
16322 |
-
},
|
16323 |
-
{
|
16324 |
-
"epoch": 13.941206874273162,
|
16325 |
-
"learning_rate": 1.1290678075975626e-07,
|
16326 |
-
"loss": 2.4386,
|
16327 |
-
"step": 53940
|
16328 |
-
},
|
16329 |
-
{
|
16330 |
-
"epoch": 13.946375500710687,
|
16331 |
-
"learning_rate": 1.1287436795021392e-07,
|
16332 |
-
"loss": 2.4117,
|
16333 |
-
"step": 53960
|
16334 |
-
},
|
16335 |
-
{
|
16336 |
-
"epoch": 13.95154412714821,
|
16337 |
-
"learning_rate": 1.1284195514067159e-07,
|
16338 |
-
"loss": 2.4356,
|
16339 |
-
"step": 53980
|
16340 |
-
},
|
16341 |
-
{
|
16342 |
-
"epoch": 13.956712753585734,
|
16343 |
-
"learning_rate": 1.1280954233112928e-07,
|
16344 |
-
"loss": 2.4823,
|
16345 |
-
"step": 54000
|
16346 |
-
},
|
16347 |
-
{
|
16348 |
-
"epoch": 13.961881380023259,
|
16349 |
-
"learning_rate": 1.1277712952158694e-07,
|
16350 |
-
"loss": 2.4368,
|
16351 |
-
"step": 54020
|
16352 |
-
},
|
16353 |
-
{
|
16354 |
-
"epoch": 13.967050006460783,
|
16355 |
-
"learning_rate": 1.127447167120446e-07,
|
16356 |
-
"loss": 2.4806,
|
16357 |
-
"step": 54040
|
16358 |
-
},
|
16359 |
-
{
|
16360 |
-
"epoch": 13.972218632898308,
|
16361 |
-
"learning_rate": 1.1271230390250226e-07,
|
16362 |
-
"loss": 2.3614,
|
16363 |
-
"step": 54060
|
16364 |
-
},
|
16365 |
-
{
|
16366 |
-
"epoch": 13.977387259335831,
|
16367 |
-
"learning_rate": 1.1267989109295995e-07,
|
16368 |
-
"loss": 2.4549,
|
16369 |
-
"step": 54080
|
16370 |
-
},
|
16371 |
-
{
|
16372 |
-
"epoch": 13.982555885773357,
|
16373 |
-
"learning_rate": 1.1264747828341761e-07,
|
16374 |
-
"loss": 2.4648,
|
16375 |
-
"step": 54100
|
16376 |
-
},
|
16377 |
-
{
|
16378 |
-
"epoch": 13.98772451221088,
|
16379 |
-
"learning_rate": 1.1261506547387527e-07,
|
16380 |
-
"loss": 2.482,
|
16381 |
-
"step": 54120
|
16382 |
-
},
|
16383 |
-
{
|
16384 |
-
"epoch": 13.992893138648403,
|
16385 |
-
"learning_rate": 1.1258265266433296e-07,
|
16386 |
-
"loss": 2.4405,
|
16387 |
-
"step": 54140
|
16388 |
-
},
|
16389 |
-
{
|
16390 |
-
"epoch": 13.998061765085929,
|
16391 |
-
"learning_rate": 1.1255023985479062e-07,
|
16392 |
-
"loss": 2.4125,
|
16393 |
-
"step": 54160
|
16394 |
-
},
|
16395 |
-
{
|
16396 |
-
"epoch": 13.999870784339063,
|
16397 |
-
"eval_bleu": 13.0772,
|
16398 |
-
"eval_gen_len": 37.0436,
|
16399 |
-
"eval_loss": 2.4198501110076904,
|
16400 |
-
"eval_runtime": 802.9059,
|
16401 |
-
"eval_samples_per_second": 2.145,
|
16402 |
-
"eval_steps_per_second": 1.072,
|
16403 |
-
"step": 54167
|
16404 |
}
|
16405 |
],
|
16406 |
"logging_steps": 20,
|
@@ -16420,7 +12910,7 @@
|
|
16420 |
"attributes": {}
|
16421 |
}
|
16422 |
},
|
16423 |
-
"total_flos":
|
16424 |
"train_batch_size": 2,
|
16425 |
"trial_name": null,
|
16426 |
"trial_params": null
|
|
|
1 |
{
|
2 |
+
"best_metric": 12.1716,
|
3 |
+
"best_model_checkpoint": "/kaggle/tmp/amr-tst-indo/AMRBART-id/fine-tune/../outputs/mbart-en-id-smaller-fted/checkpoint-42562",
|
4 |
+
"epoch": 11.0,
|
5 |
"eval_steps": 500,
|
6 |
+
"global_step": 42562,
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
|
|
12891 |
"eval_samples_per_second": 2.165,
|
12892 |
"eval_steps_per_second": 1.083,
|
12893 |
"step": 42562
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
12894 |
}
|
12895 |
],
|
12896 |
"logging_steps": 20,
|
|
|
12910 |
"attributes": {}
|
12911 |
}
|
12912 |
},
|
12913 |
+
"total_flos": 2.793935715793797e+17,
|
12914 |
"train_batch_size": 2,
|
12915 |
"trial_name": null,
|
12916 |
"trial_params": null
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1575259780
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ec8f6718bbb0f1ed8877dd4712ad6aac5ba74032ff57de3fb4fc16d307e93308
|
3 |
size 1575259780
|
run-2024-10-28T15:22:08+00:00.log
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
val_outputs/val_generated_predictions_14521373.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|