|
{ |
|
"best_metric": 0.14961867034435272, |
|
"best_model_checkpoint": "/home/ubuntu/ML/source_code/fine-tuned-legalroberta/qa/checkpoint-364860", |
|
"epoch": 2.0, |
|
"global_step": 364860, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 1.9981728151437082e-05, |
|
"loss": 0.3901, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.9963456302874163e-05, |
|
"loss": 0.1441, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.9945184454311244e-05, |
|
"loss": 0.107, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.9926912605748324e-05, |
|
"loss": 0.1122, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.9908640757185405e-05, |
|
"loss": 0.1177, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.9890368908622486e-05, |
|
"loss": 0.1039, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.9872097060059567e-05, |
|
"loss": 0.0989, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.985382521149665e-05, |
|
"loss": 0.1081, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.983555336293373e-05, |
|
"loss": 0.0978, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.981728151437081e-05, |
|
"loss": 0.0979, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.9799009665807893e-05, |
|
"loss": 0.1083, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.9780737817244973e-05, |
|
"loss": 0.0935, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.9762465968682054e-05, |
|
"loss": 0.091, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.9744194120119135e-05, |
|
"loss": 0.1068, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.9725922271556216e-05, |
|
"loss": 0.1264, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.9707650422993296e-05, |
|
"loss": 0.1381, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.9689378574430377e-05, |
|
"loss": 0.0971, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.9671106725867458e-05, |
|
"loss": 0.1426, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.965283487730454e-05, |
|
"loss": 0.1167, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.963456302874162e-05, |
|
"loss": 0.1031, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.96162911801787e-05, |
|
"loss": 0.1154, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.959801933161578e-05, |
|
"loss": 0.1116, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.957974748305286e-05, |
|
"loss": 0.0929, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.9561475634489945e-05, |
|
"loss": 0.0812, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.9543203785927023e-05, |
|
"loss": 0.1107, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.9524931937364103e-05, |
|
"loss": 0.0934, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.9506660088801187e-05, |
|
"loss": 0.0963, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.9488388240238268e-05, |
|
"loss": 0.1121, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.9470116391675345e-05, |
|
"loss": 0.1094, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.945184454311243e-05, |
|
"loss": 0.0955, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.943357269454951e-05, |
|
"loss": 0.1061, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.941530084598659e-05, |
|
"loss": 0.1049, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.939702899742367e-05, |
|
"loss": 0.1018, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.9378757148860752e-05, |
|
"loss": 0.1117, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.9360485300297833e-05, |
|
"loss": 0.0995, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.9342213451734914e-05, |
|
"loss": 0.1019, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.9323941603171994e-05, |
|
"loss": 0.0892, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.9305669754609075e-05, |
|
"loss": 0.0982, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.9287397906046156e-05, |
|
"loss": 0.0943, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.9269126057483236e-05, |
|
"loss": 0.0885, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.9250854208920317e-05, |
|
"loss": 0.0956, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.9232582360357398e-05, |
|
"loss": 0.1101, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.9214310511794482e-05, |
|
"loss": 0.099, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.919603866323156e-05, |
|
"loss": 0.0919, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.917776681466864e-05, |
|
"loss": 0.0944, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.9159494966105724e-05, |
|
"loss": 0.1062, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.9141223117542805e-05, |
|
"loss": 0.0833, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.9122951268979882e-05, |
|
"loss": 0.1055, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.9104679420416966e-05, |
|
"loss": 0.0978, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.9086407571854047e-05, |
|
"loss": 0.0844, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.9068135723291127e-05, |
|
"loss": 0.092, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.9049863874728208e-05, |
|
"loss": 0.1131, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.903159202616529e-05, |
|
"loss": 0.0858, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.901332017760237e-05, |
|
"loss": 0.093, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.899504832903945e-05, |
|
"loss": 0.086, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.897677648047653e-05, |
|
"loss": 0.0785, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.895850463191361e-05, |
|
"loss": 0.0949, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.8940232783350692e-05, |
|
"loss": 0.1032, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.8921960934787773e-05, |
|
"loss": 0.0887, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.8903689086224854e-05, |
|
"loss": 0.102, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.8885417237661934e-05, |
|
"loss": 0.099, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.886714538909902e-05, |
|
"loss": 0.0886, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.8848873540536096e-05, |
|
"loss": 0.0955, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.8830601691973177e-05, |
|
"loss": 0.0786, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.881232984341026e-05, |
|
"loss": 0.109, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.879405799484734e-05, |
|
"loss": 0.0798, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.8775786146284422e-05, |
|
"loss": 0.075, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.8757514297721503e-05, |
|
"loss": 0.1025, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.8739242449158583e-05, |
|
"loss": 0.0832, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.8720970600595664e-05, |
|
"loss": 0.0942, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.8702698752032745e-05, |
|
"loss": 0.0905, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.8684426903469826e-05, |
|
"loss": 0.0967, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.8666155054906906e-05, |
|
"loss": 0.1022, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.8647883206343987e-05, |
|
"loss": 0.091, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.8629611357781068e-05, |
|
"loss": 0.1029, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.861133950921815e-05, |
|
"loss": 0.0867, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.8593067660655232e-05, |
|
"loss": 0.1072, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.857479581209231e-05, |
|
"loss": 0.1039, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.855652396352939e-05, |
|
"loss": 0.0988, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.8538252114966475e-05, |
|
"loss": 0.0769, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.8519980266403555e-05, |
|
"loss": 0.0818, |
|
"step": 40500 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.8501708417840633e-05, |
|
"loss": 0.1, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.8483436569277717e-05, |
|
"loss": 0.1093, |
|
"step": 41500 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.8465164720714797e-05, |
|
"loss": 0.0806, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.8446892872151878e-05, |
|
"loss": 0.0877, |
|
"step": 42500 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.842862102358896e-05, |
|
"loss": 0.1024, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.841034917502604e-05, |
|
"loss": 0.0899, |
|
"step": 43500 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.839207732646312e-05, |
|
"loss": 0.0768, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.83738054779002e-05, |
|
"loss": 0.0896, |
|
"step": 44500 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.835553362933728e-05, |
|
"loss": 0.0875, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.8337261780774362e-05, |
|
"loss": 0.0855, |
|
"step": 45500 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.8318989932211443e-05, |
|
"loss": 0.1109, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.8300718083648524e-05, |
|
"loss": 0.0956, |
|
"step": 46500 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.8282446235085604e-05, |
|
"loss": 0.0936, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.8264174386522685e-05, |
|
"loss": 0.0734, |
|
"step": 47500 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.824590253795977e-05, |
|
"loss": 0.0805, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.8227630689396846e-05, |
|
"loss": 0.106, |
|
"step": 48500 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.8209358840833927e-05, |
|
"loss": 0.0693, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.819108699227101e-05, |
|
"loss": 0.0893, |
|
"step": 49500 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.8172815143708092e-05, |
|
"loss": 0.096, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.815454329514517e-05, |
|
"loss": 0.0879, |
|
"step": 50500 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.8136271446582253e-05, |
|
"loss": 0.0934, |
|
"step": 51000 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.8117999598019334e-05, |
|
"loss": 0.0797, |
|
"step": 51500 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.8099727749456415e-05, |
|
"loss": 0.0825, |
|
"step": 52000 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.8081455900893495e-05, |
|
"loss": 0.0926, |
|
"step": 52500 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.8063184052330576e-05, |
|
"loss": 0.0796, |
|
"step": 53000 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.8044912203767657e-05, |
|
"loss": 0.0604, |
|
"step": 53500 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.8026640355204737e-05, |
|
"loss": 0.0844, |
|
"step": 54000 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.8008368506641818e-05, |
|
"loss": 0.0701, |
|
"step": 54500 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.79900966580789e-05, |
|
"loss": 0.0764, |
|
"step": 55000 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.797182480951598e-05, |
|
"loss": 0.0856, |
|
"step": 55500 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.795355296095306e-05, |
|
"loss": 0.0933, |
|
"step": 56000 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.793528111239014e-05, |
|
"loss": 0.0891, |
|
"step": 56500 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.791700926382722e-05, |
|
"loss": 0.0736, |
|
"step": 57000 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.7898737415264306e-05, |
|
"loss": 0.0875, |
|
"step": 57500 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.7880465566701383e-05, |
|
"loss": 0.0802, |
|
"step": 58000 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.7862193718138464e-05, |
|
"loss": 0.0759, |
|
"step": 58500 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.7843921869575548e-05, |
|
"loss": 0.0957, |
|
"step": 59000 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.782565002101263e-05, |
|
"loss": 0.0853, |
|
"step": 59500 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.7807378172449706e-05, |
|
"loss": 0.0829, |
|
"step": 60000 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.778910632388679e-05, |
|
"loss": 0.0798, |
|
"step": 60500 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.777083447532387e-05, |
|
"loss": 0.0881, |
|
"step": 61000 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.775256262676095e-05, |
|
"loss": 0.0842, |
|
"step": 61500 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.7734290778198032e-05, |
|
"loss": 0.0722, |
|
"step": 62000 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.7716018929635113e-05, |
|
"loss": 0.081, |
|
"step": 62500 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.7697747081072193e-05, |
|
"loss": 0.0696, |
|
"step": 63000 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.7679475232509274e-05, |
|
"loss": 0.098, |
|
"step": 63500 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.7661203383946355e-05, |
|
"loss": 0.067, |
|
"step": 64000 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.7642931535383436e-05, |
|
"loss": 0.0674, |
|
"step": 64500 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.7624659686820516e-05, |
|
"loss": 0.0803, |
|
"step": 65000 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.7606387838257597e-05, |
|
"loss": 0.0715, |
|
"step": 65500 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.7588115989694678e-05, |
|
"loss": 0.0857, |
|
"step": 66000 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.756984414113176e-05, |
|
"loss": 0.0849, |
|
"step": 66500 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.7551572292568842e-05, |
|
"loss": 0.08, |
|
"step": 67000 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.753330044400592e-05, |
|
"loss": 0.0651, |
|
"step": 67500 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.7515028595443004e-05, |
|
"loss": 0.0664, |
|
"step": 68000 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.7496756746880085e-05, |
|
"loss": 0.0555, |
|
"step": 68500 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.7478484898317165e-05, |
|
"loss": 0.0775, |
|
"step": 69000 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.7460213049754246e-05, |
|
"loss": 0.0731, |
|
"step": 69500 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.7441941201191327e-05, |
|
"loss": 0.0723, |
|
"step": 70000 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.7423669352628407e-05, |
|
"loss": 0.0674, |
|
"step": 70500 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.7405397504065488e-05, |
|
"loss": 0.0636, |
|
"step": 71000 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.738712565550257e-05, |
|
"loss": 0.0855, |
|
"step": 71500 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.736885380693965e-05, |
|
"loss": 0.0696, |
|
"step": 72000 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.735058195837673e-05, |
|
"loss": 0.0647, |
|
"step": 72500 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.733231010981381e-05, |
|
"loss": 0.0649, |
|
"step": 73000 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.731403826125089e-05, |
|
"loss": 0.077, |
|
"step": 73500 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.7295766412687972e-05, |
|
"loss": 0.0663, |
|
"step": 74000 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.7277494564125056e-05, |
|
"loss": 0.0675, |
|
"step": 74500 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.7259222715562134e-05, |
|
"loss": 0.0708, |
|
"step": 75000 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.7240950866999214e-05, |
|
"loss": 0.0805, |
|
"step": 75500 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.72226790184363e-05, |
|
"loss": 0.0776, |
|
"step": 76000 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.720440716987338e-05, |
|
"loss": 0.0604, |
|
"step": 76500 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.7186135321310456e-05, |
|
"loss": 0.0612, |
|
"step": 77000 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.716786347274754e-05, |
|
"loss": 0.0546, |
|
"step": 77500 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.714959162418462e-05, |
|
"loss": 0.0753, |
|
"step": 78000 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.7131319775621702e-05, |
|
"loss": 0.0688, |
|
"step": 78500 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.7113047927058783e-05, |
|
"loss": 0.0603, |
|
"step": 79000 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.7094776078495863e-05, |
|
"loss": 0.0679, |
|
"step": 79500 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.7076504229932944e-05, |
|
"loss": 0.0715, |
|
"step": 80000 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.7058232381370025e-05, |
|
"loss": 0.0655, |
|
"step": 80500 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.7039960532807105e-05, |
|
"loss": 0.0744, |
|
"step": 81000 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.7021688684244186e-05, |
|
"loss": 0.0636, |
|
"step": 81500 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.7003416835681267e-05, |
|
"loss": 0.0838, |
|
"step": 82000 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.6985144987118347e-05, |
|
"loss": 0.064, |
|
"step": 82500 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.6966873138555428e-05, |
|
"loss": 0.0632, |
|
"step": 83000 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.694860128999251e-05, |
|
"loss": 0.069, |
|
"step": 83500 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.6930329441429593e-05, |
|
"loss": 0.0617, |
|
"step": 84000 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.691205759286667e-05, |
|
"loss": 0.0716, |
|
"step": 84500 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.689378574430375e-05, |
|
"loss": 0.075, |
|
"step": 85000 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.6875513895740835e-05, |
|
"loss": 0.0729, |
|
"step": 85500 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.6857242047177916e-05, |
|
"loss": 0.0669, |
|
"step": 86000 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.6838970198614993e-05, |
|
"loss": 0.0664, |
|
"step": 86500 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.6820698350052077e-05, |
|
"loss": 0.077, |
|
"step": 87000 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.6802426501489158e-05, |
|
"loss": 0.0781, |
|
"step": 87500 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.678415465292624e-05, |
|
"loss": 0.0666, |
|
"step": 88000 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.676588280436332e-05, |
|
"loss": 0.0657, |
|
"step": 88500 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.67476109558004e-05, |
|
"loss": 0.0612, |
|
"step": 89000 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.672933910723748e-05, |
|
"loss": 0.0694, |
|
"step": 89500 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.671106725867456e-05, |
|
"loss": 0.0671, |
|
"step": 90000 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.6692795410111642e-05, |
|
"loss": 0.0729, |
|
"step": 90500 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.6674523561548723e-05, |
|
"loss": 0.0723, |
|
"step": 91000 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.6656251712985803e-05, |
|
"loss": 0.0777, |
|
"step": 91500 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.6637979864422884e-05, |
|
"loss": 0.0798, |
|
"step": 92000 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.6619708015859965e-05, |
|
"loss": 0.0548, |
|
"step": 92500 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.6601436167297046e-05, |
|
"loss": 0.0646, |
|
"step": 93000 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.658316431873413e-05, |
|
"loss": 0.0639, |
|
"step": 93500 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 1.6564892470171207e-05, |
|
"loss": 0.0529, |
|
"step": 94000 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 1.6546620621608288e-05, |
|
"loss": 0.0577, |
|
"step": 94500 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 1.6528348773045372e-05, |
|
"loss": 0.0632, |
|
"step": 95000 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 1.6510076924482452e-05, |
|
"loss": 0.0463, |
|
"step": 95500 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 1.649180507591953e-05, |
|
"loss": 0.0661, |
|
"step": 96000 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 1.6473533227356614e-05, |
|
"loss": 0.0742, |
|
"step": 96500 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 1.6455261378793695e-05, |
|
"loss": 0.0549, |
|
"step": 97000 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 1.6436989530230775e-05, |
|
"loss": 0.064, |
|
"step": 97500 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 1.6418717681667856e-05, |
|
"loss": 0.0597, |
|
"step": 98000 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 1.6400445833104937e-05, |
|
"loss": 0.0795, |
|
"step": 98500 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 1.6382173984542017e-05, |
|
"loss": 0.0658, |
|
"step": 99000 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 1.6363902135979098e-05, |
|
"loss": 0.0574, |
|
"step": 99500 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 1.634563028741618e-05, |
|
"loss": 0.0641, |
|
"step": 100000 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 1.632735843885326e-05, |
|
"loss": 0.0594, |
|
"step": 100500 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 1.630908659029034e-05, |
|
"loss": 0.0572, |
|
"step": 101000 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 1.629081474172742e-05, |
|
"loss": 0.067, |
|
"step": 101500 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 1.62725428931645e-05, |
|
"loss": 0.0617, |
|
"step": 102000 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 1.6254271044601582e-05, |
|
"loss": 0.0689, |
|
"step": 102500 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 1.6235999196038666e-05, |
|
"loss": 0.0676, |
|
"step": 103000 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 1.6217727347475744e-05, |
|
"loss": 0.067, |
|
"step": 103500 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 1.6199455498912828e-05, |
|
"loss": 0.0527, |
|
"step": 104000 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 1.618118365034991e-05, |
|
"loss": 0.0581, |
|
"step": 104500 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 1.616291180178699e-05, |
|
"loss": 0.0647, |
|
"step": 105000 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 1.614463995322407e-05, |
|
"loss": 0.0826, |
|
"step": 105500 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 1.612636810466115e-05, |
|
"loss": 0.0654, |
|
"step": 106000 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 1.610809625609823e-05, |
|
"loss": 0.066, |
|
"step": 106500 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 1.6089824407535312e-05, |
|
"loss": 0.056, |
|
"step": 107000 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 1.6071552558972393e-05, |
|
"loss": 0.0715, |
|
"step": 107500 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 1.6053280710409473e-05, |
|
"loss": 0.0664, |
|
"step": 108000 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 1.6035008861846554e-05, |
|
"loss": 0.072, |
|
"step": 108500 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 1.6016737013283635e-05, |
|
"loss": 0.065, |
|
"step": 109000 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 1.5998465164720715e-05, |
|
"loss": 0.0648, |
|
"step": 109500 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 1.5980193316157796e-05, |
|
"loss": 0.0874, |
|
"step": 110000 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 1.596192146759488e-05, |
|
"loss": 0.0721, |
|
"step": 110500 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 1.5943649619031958e-05, |
|
"loss": 0.0536, |
|
"step": 111000 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 1.5925377770469038e-05, |
|
"loss": 0.0591, |
|
"step": 111500 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 1.5907105921906122e-05, |
|
"loss": 0.0627, |
|
"step": 112000 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 1.5888834073343203e-05, |
|
"loss": 0.0583, |
|
"step": 112500 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 1.587056222478028e-05, |
|
"loss": 0.0569, |
|
"step": 113000 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 1.5852290376217364e-05, |
|
"loss": 0.0597, |
|
"step": 113500 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 1.5834018527654445e-05, |
|
"loss": 0.0672, |
|
"step": 114000 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 1.5815746679091526e-05, |
|
"loss": 0.0565, |
|
"step": 114500 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 1.5797474830528606e-05, |
|
"loss": 0.053, |
|
"step": 115000 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 1.5779202981965687e-05, |
|
"loss": 0.0535, |
|
"step": 115500 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 1.5760931133402768e-05, |
|
"loss": 0.0645, |
|
"step": 116000 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 1.574265928483985e-05, |
|
"loss": 0.059, |
|
"step": 116500 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 1.572438743627693e-05, |
|
"loss": 0.0703, |
|
"step": 117000 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 1.570611558771401e-05, |
|
"loss": 0.0471, |
|
"step": 117500 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 1.568784373915109e-05, |
|
"loss": 0.044, |
|
"step": 118000 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 1.566957189058817e-05, |
|
"loss": 0.0639, |
|
"step": 118500 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 1.5651300042025252e-05, |
|
"loss": 0.0551, |
|
"step": 119000 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 1.5633028193462333e-05, |
|
"loss": 0.0573, |
|
"step": 119500 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 1.5614756344899417e-05, |
|
"loss": 0.0453, |
|
"step": 120000 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 1.5596484496336494e-05, |
|
"loss": 0.063, |
|
"step": 120500 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 1.5578212647773575e-05, |
|
"loss": 0.0611, |
|
"step": 121000 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 1.555994079921066e-05, |
|
"loss": 0.0519, |
|
"step": 121500 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 1.554166895064774e-05, |
|
"loss": 0.0636, |
|
"step": 122000 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 1.5523397102084817e-05, |
|
"loss": 0.0803, |
|
"step": 122500 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 1.55051252535219e-05, |
|
"loss": 0.0626, |
|
"step": 123000 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 1.5486853404958982e-05, |
|
"loss": 0.0476, |
|
"step": 123500 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 1.5468581556396062e-05, |
|
"loss": 0.0617, |
|
"step": 124000 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 1.5450309707833143e-05, |
|
"loss": 0.0524, |
|
"step": 124500 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 1.5432037859270224e-05, |
|
"loss": 0.0498, |
|
"step": 125000 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 1.5413766010707305e-05, |
|
"loss": 0.0535, |
|
"step": 125500 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 1.5395494162144385e-05, |
|
"loss": 0.0497, |
|
"step": 126000 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 1.5377222313581466e-05, |
|
"loss": 0.0592, |
|
"step": 126500 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.5358950465018547e-05, |
|
"loss": 0.0443, |
|
"step": 127000 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.5340678616455627e-05, |
|
"loss": 0.0547, |
|
"step": 127500 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.5322406767892708e-05, |
|
"loss": 0.06, |
|
"step": 128000 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.530413491932979e-05, |
|
"loss": 0.0482, |
|
"step": 128500 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.528586307076687e-05, |
|
"loss": 0.0559, |
|
"step": 129000 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.5267591222203954e-05, |
|
"loss": 0.0533, |
|
"step": 129500 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.5249319373641031e-05, |
|
"loss": 0.0581, |
|
"step": 130000 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.5231047525078113e-05, |
|
"loss": 0.0615, |
|
"step": 130500 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.5212775676515194e-05, |
|
"loss": 0.0518, |
|
"step": 131000 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.5194503827952276e-05, |
|
"loss": 0.0436, |
|
"step": 131500 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.5176231979389355e-05, |
|
"loss": 0.0558, |
|
"step": 132000 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.5157960130826436e-05, |
|
"loss": 0.0472, |
|
"step": 132500 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.5139688282263518e-05, |
|
"loss": 0.0503, |
|
"step": 133000 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.5121416433700599e-05, |
|
"loss": 0.0482, |
|
"step": 133500 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.510314458513768e-05, |
|
"loss": 0.0584, |
|
"step": 134000 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.508487273657476e-05, |
|
"loss": 0.0632, |
|
"step": 134500 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.5066600888011841e-05, |
|
"loss": 0.0571, |
|
"step": 135000 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.5048329039448922e-05, |
|
"loss": 0.0524, |
|
"step": 135500 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.5030057190886003e-05, |
|
"loss": 0.0447, |
|
"step": 136000 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.5011785342323083e-05, |
|
"loss": 0.0541, |
|
"step": 136500 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.4993513493760166e-05, |
|
"loss": 0.0487, |
|
"step": 137000 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.4975241645197245e-05, |
|
"loss": 0.063, |
|
"step": 137500 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.4956969796634327e-05, |
|
"loss": 0.0576, |
|
"step": 138000 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.4938697948071408e-05, |
|
"loss": 0.0618, |
|
"step": 138500 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.4920426099508489e-05, |
|
"loss": 0.0439, |
|
"step": 139000 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.490215425094557e-05, |
|
"loss": 0.0455, |
|
"step": 139500 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.488388240238265e-05, |
|
"loss": 0.0581, |
|
"step": 140000 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.4865610553819732e-05, |
|
"loss": 0.0544, |
|
"step": 140500 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.4847338705256813e-05, |
|
"loss": 0.0521, |
|
"step": 141000 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.4829066856693892e-05, |
|
"loss": 0.0454, |
|
"step": 141500 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.4810795008130974e-05, |
|
"loss": 0.0518, |
|
"step": 142000 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.4792523159568055e-05, |
|
"loss": 0.0579, |
|
"step": 142500 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.4774251311005138e-05, |
|
"loss": 0.0357, |
|
"step": 143000 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.4755979462442217e-05, |
|
"loss": 0.0494, |
|
"step": 143500 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.4737707613879297e-05, |
|
"loss": 0.0459, |
|
"step": 144000 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.471943576531638e-05, |
|
"loss": 0.0581, |
|
"step": 144500 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.4701163916753459e-05, |
|
"loss": 0.0546, |
|
"step": 145000 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.468289206819054e-05, |
|
"loss": 0.0551, |
|
"step": 145500 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.4664620219627622e-05, |
|
"loss": 0.0476, |
|
"step": 146000 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.4646348371064702e-05, |
|
"loss": 0.0533, |
|
"step": 146500 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.4628076522501781e-05, |
|
"loss": 0.0561, |
|
"step": 147000 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.4609804673938864e-05, |
|
"loss": 0.0504, |
|
"step": 147500 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.4591532825375944e-05, |
|
"loss": 0.0398, |
|
"step": 148000 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.4573260976813027e-05, |
|
"loss": 0.054, |
|
"step": 148500 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.4554989128250106e-05, |
|
"loss": 0.047, |
|
"step": 149000 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.4536717279687187e-05, |
|
"loss": 0.047, |
|
"step": 149500 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.4518445431124269e-05, |
|
"loss": 0.0489, |
|
"step": 150000 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.450017358256135e-05, |
|
"loss": 0.0551, |
|
"step": 150500 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.4481901733998429e-05, |
|
"loss": 0.043, |
|
"step": 151000 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.4463629885435511e-05, |
|
"loss": 0.0451, |
|
"step": 151500 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.4445358036872592e-05, |
|
"loss": 0.057, |
|
"step": 152000 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.4427086188309674e-05, |
|
"loss": 0.0554, |
|
"step": 152500 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.4408814339746753e-05, |
|
"loss": 0.0469, |
|
"step": 153000 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.4390542491183834e-05, |
|
"loss": 0.0504, |
|
"step": 153500 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.4372270642620916e-05, |
|
"loss": 0.0567, |
|
"step": 154000 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.4353998794057995e-05, |
|
"loss": 0.0634, |
|
"step": 154500 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.4335726945495076e-05, |
|
"loss": 0.0471, |
|
"step": 155000 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.4317455096932158e-05, |
|
"loss": 0.0391, |
|
"step": 155500 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.4299183248369239e-05, |
|
"loss": 0.0481, |
|
"step": 156000 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.4280911399806318e-05, |
|
"loss": 0.0554, |
|
"step": 156500 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.42626395512434e-05, |
|
"loss": 0.0501, |
|
"step": 157000 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.4244367702680481e-05, |
|
"loss": 0.0456, |
|
"step": 157500 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.4226095854117564e-05, |
|
"loss": 0.0331, |
|
"step": 158000 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.4207824005554643e-05, |
|
"loss": 0.0514, |
|
"step": 158500 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.4189552156991723e-05, |
|
"loss": 0.0434, |
|
"step": 159000 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.4171280308428806e-05, |
|
"loss": 0.0519, |
|
"step": 159500 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 1.4153008459865886e-05, |
|
"loss": 0.0434, |
|
"step": 160000 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 1.4134736611302965e-05, |
|
"loss": 0.0385, |
|
"step": 160500 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 1.4116464762740048e-05, |
|
"loss": 0.058, |
|
"step": 161000 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 1.4098192914177128e-05, |
|
"loss": 0.0476, |
|
"step": 161500 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 1.4079921065614207e-05, |
|
"loss": 0.0462, |
|
"step": 162000 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 1.406164921705129e-05, |
|
"loss": 0.0447, |
|
"step": 162500 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 1.404337736848837e-05, |
|
"loss": 0.048, |
|
"step": 163000 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 1.4025105519925453e-05, |
|
"loss": 0.0442, |
|
"step": 163500 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 1.4006833671362532e-05, |
|
"loss": 0.0555, |
|
"step": 164000 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 1.3988561822799613e-05, |
|
"loss": 0.0434, |
|
"step": 164500 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 1.3970289974236695e-05, |
|
"loss": 0.0577, |
|
"step": 165000 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.3952018125673776e-05, |
|
"loss": 0.0522, |
|
"step": 165500 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.3933746277110855e-05, |
|
"loss": 0.0412, |
|
"step": 166000 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.3915474428547937e-05, |
|
"loss": 0.0601, |
|
"step": 166500 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.3897202579985018e-05, |
|
"loss": 0.0553, |
|
"step": 167000 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.38789307314221e-05, |
|
"loss": 0.0522, |
|
"step": 167500 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.386065888285918e-05, |
|
"loss": 0.0505, |
|
"step": 168000 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.384238703429626e-05, |
|
"loss": 0.0497, |
|
"step": 168500 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.3824115185733342e-05, |
|
"loss": 0.0378, |
|
"step": 169000 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.3805843337170423e-05, |
|
"loss": 0.0399, |
|
"step": 169500 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.3787571488607504e-05, |
|
"loss": 0.0485, |
|
"step": 170000 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.3769299640044584e-05, |
|
"loss": 0.0416, |
|
"step": 170500 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.3751027791481665e-05, |
|
"loss": 0.0486, |
|
"step": 171000 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.3732755942918746e-05, |
|
"loss": 0.0498, |
|
"step": 171500 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.3714484094355827e-05, |
|
"loss": 0.0515, |
|
"step": 172000 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.3696212245792909e-05, |
|
"loss": 0.0403, |
|
"step": 172500 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.367794039722999e-05, |
|
"loss": 0.0461, |
|
"step": 173000 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.3659668548667069e-05, |
|
"loss": 0.0557, |
|
"step": 173500 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.3641396700104151e-05, |
|
"loss": 0.0602, |
|
"step": 174000 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.3623124851541232e-05, |
|
"loss": 0.0416, |
|
"step": 174500 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.3604853002978312e-05, |
|
"loss": 0.0472, |
|
"step": 175000 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.3586581154415393e-05, |
|
"loss": 0.0475, |
|
"step": 175500 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.3568309305852474e-05, |
|
"loss": 0.0478, |
|
"step": 176000 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 1.3550037457289556e-05, |
|
"loss": 0.0448, |
|
"step": 176500 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 1.3531765608726637e-05, |
|
"loss": 0.0444, |
|
"step": 177000 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 1.3513493760163716e-05, |
|
"loss": 0.0459, |
|
"step": 177500 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.3495221911600798e-05, |
|
"loss": 0.0381, |
|
"step": 178000 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.3476950063037879e-05, |
|
"loss": 0.0402, |
|
"step": 178500 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.3458678214474961e-05, |
|
"loss": 0.0338, |
|
"step": 179000 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.344040636591204e-05, |
|
"loss": 0.0527, |
|
"step": 179500 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.3422134517349121e-05, |
|
"loss": 0.0381, |
|
"step": 180000 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.3403862668786203e-05, |
|
"loss": 0.0473, |
|
"step": 180500 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.3385590820223282e-05, |
|
"loss": 0.0475, |
|
"step": 181000 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.3367318971660363e-05, |
|
"loss": 0.0482, |
|
"step": 181500 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 1.3349047123097446e-05, |
|
"loss": 0.0432, |
|
"step": 182000 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 0.15653175115585327, |
|
"eval_runtime": 1676.2688, |
|
"eval_samples_per_second": 138.216, |
|
"eval_steps_per_second": 13.822, |
|
"step": 182430 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 1.3330775274534526e-05, |
|
"loss": 0.0411, |
|
"step": 182500 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 1.3312503425971605e-05, |
|
"loss": 0.0428, |
|
"step": 183000 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 1.3294231577408688e-05, |
|
"loss": 0.0454, |
|
"step": 183500 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 1.3275959728845768e-05, |
|
"loss": 0.0453, |
|
"step": 184000 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 1.325768788028285e-05, |
|
"loss": 0.0451, |
|
"step": 184500 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 1.323941603171993e-05, |
|
"loss": 0.0426, |
|
"step": 185000 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 1.322114418315701e-05, |
|
"loss": 0.0386, |
|
"step": 185500 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 1.3202872334594093e-05, |
|
"loss": 0.0422, |
|
"step": 186000 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 1.3184600486031174e-05, |
|
"loss": 0.0491, |
|
"step": 186500 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 1.3166328637468253e-05, |
|
"loss": 0.0493, |
|
"step": 187000 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 1.3148056788905335e-05, |
|
"loss": 0.0388, |
|
"step": 187500 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 1.3129784940342416e-05, |
|
"loss": 0.0429, |
|
"step": 188000 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 1.3111513091779498e-05, |
|
"loss": 0.0501, |
|
"step": 188500 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 1.3093241243216577e-05, |
|
"loss": 0.0423, |
|
"step": 189000 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 1.3074969394653658e-05, |
|
"loss": 0.0458, |
|
"step": 189500 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 1.305669754609074e-05, |
|
"loss": 0.0411, |
|
"step": 190000 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 1.303842569752782e-05, |
|
"loss": 0.0411, |
|
"step": 190500 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 1.30201538489649e-05, |
|
"loss": 0.037, |
|
"step": 191000 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 1.3001882000401982e-05, |
|
"loss": 0.039, |
|
"step": 191500 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 1.2983610151839063e-05, |
|
"loss": 0.0468, |
|
"step": 192000 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 1.2965338303276142e-05, |
|
"loss": 0.0404, |
|
"step": 192500 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 1.2947066454713224e-05, |
|
"loss": 0.0394, |
|
"step": 193000 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 1.2928794606150305e-05, |
|
"loss": 0.0454, |
|
"step": 193500 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 1.2910522757587387e-05, |
|
"loss": 0.0402, |
|
"step": 194000 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.2892250909024466e-05, |
|
"loss": 0.0507, |
|
"step": 194500 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.2873979060461547e-05, |
|
"loss": 0.0429, |
|
"step": 195000 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.285570721189863e-05, |
|
"loss": 0.0452, |
|
"step": 195500 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.283743536333571e-05, |
|
"loss": 0.042, |
|
"step": 196000 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.281916351477279e-05, |
|
"loss": 0.0482, |
|
"step": 196500 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.2800891666209872e-05, |
|
"loss": 0.0457, |
|
"step": 197000 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.2782619817646952e-05, |
|
"loss": 0.0383, |
|
"step": 197500 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.2764347969084031e-05, |
|
"loss": 0.0384, |
|
"step": 198000 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.2746076120521114e-05, |
|
"loss": 0.0558, |
|
"step": 198500 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.2727804271958194e-05, |
|
"loss": 0.0362, |
|
"step": 199000 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.2709532423395277e-05, |
|
"loss": 0.044, |
|
"step": 199500 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.2691260574832356e-05, |
|
"loss": 0.0475, |
|
"step": 200000 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.2672988726269437e-05, |
|
"loss": 0.0404, |
|
"step": 200500 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.2654716877706519e-05, |
|
"loss": 0.0384, |
|
"step": 201000 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.26364450291436e-05, |
|
"loss": 0.0528, |
|
"step": 201500 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.261817318058068e-05, |
|
"loss": 0.0572, |
|
"step": 202000 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.2599901332017761e-05, |
|
"loss": 0.0365, |
|
"step": 202500 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.2581629483454842e-05, |
|
"loss": 0.0354, |
|
"step": 203000 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.2563357634891924e-05, |
|
"loss": 0.0391, |
|
"step": 203500 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.2545085786329003e-05, |
|
"loss": 0.0394, |
|
"step": 204000 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.2526813937766086e-05, |
|
"loss": 0.044, |
|
"step": 204500 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.2508542089203166e-05, |
|
"loss": 0.0436, |
|
"step": 205000 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.2490270240640247e-05, |
|
"loss": 0.0294, |
|
"step": 205500 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.2471998392077328e-05, |
|
"loss": 0.0422, |
|
"step": 206000 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.2453726543514408e-05, |
|
"loss": 0.0428, |
|
"step": 206500 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.2435454694951489e-05, |
|
"loss": 0.0414, |
|
"step": 207000 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.241718284638857e-05, |
|
"loss": 0.0411, |
|
"step": 207500 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.239891099782565e-05, |
|
"loss": 0.0261, |
|
"step": 208000 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.2380639149262733e-05, |
|
"loss": 0.0327, |
|
"step": 208500 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.2362367300699814e-05, |
|
"loss": 0.0373, |
|
"step": 209000 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.2344095452136893e-05, |
|
"loss": 0.0467, |
|
"step": 209500 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.2325823603573975e-05, |
|
"loss": 0.0375, |
|
"step": 210000 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.2307551755011056e-05, |
|
"loss": 0.042, |
|
"step": 210500 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 1.2289279906448138e-05, |
|
"loss": 0.0365, |
|
"step": 211000 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 1.2271008057885217e-05, |
|
"loss": 0.0464, |
|
"step": 211500 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 1.2252736209322298e-05, |
|
"loss": 0.0398, |
|
"step": 212000 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 1.223446436075938e-05, |
|
"loss": 0.0391, |
|
"step": 212500 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 1.221619251219646e-05, |
|
"loss": 0.0413, |
|
"step": 213000 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 1.219792066363354e-05, |
|
"loss": 0.0358, |
|
"step": 213500 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 1.2179648815070622e-05, |
|
"loss": 0.043, |
|
"step": 214000 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 1.2161376966507703e-05, |
|
"loss": 0.0413, |
|
"step": 214500 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 1.2143105117944785e-05, |
|
"loss": 0.0513, |
|
"step": 215000 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 1.2124833269381864e-05, |
|
"loss": 0.0421, |
|
"step": 215500 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 1.2106561420818945e-05, |
|
"loss": 0.0474, |
|
"step": 216000 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 1.2088289572256027e-05, |
|
"loss": 0.0445, |
|
"step": 216500 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 1.2070017723693106e-05, |
|
"loss": 0.0396, |
|
"step": 217000 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 1.2051745875130187e-05, |
|
"loss": 0.038, |
|
"step": 217500 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 1.203347402656727e-05, |
|
"loss": 0.0269, |
|
"step": 218000 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 1.201520217800435e-05, |
|
"loss": 0.0422, |
|
"step": 218500 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 1.199693032944143e-05, |
|
"loss": 0.0522, |
|
"step": 219000 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 1.1978658480878512e-05, |
|
"loss": 0.035, |
|
"step": 219500 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 1.1960386632315592e-05, |
|
"loss": 0.0388, |
|
"step": 220000 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 1.1942114783752675e-05, |
|
"loss": 0.033, |
|
"step": 220500 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 1.1923842935189754e-05, |
|
"loss": 0.0459, |
|
"step": 221000 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 1.1905571086626834e-05, |
|
"loss": 0.0368, |
|
"step": 221500 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 1.1887299238063917e-05, |
|
"loss": 0.0415, |
|
"step": 222000 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 1.1869027389500997e-05, |
|
"loss": 0.0373, |
|
"step": 222500 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 1.1850755540938076e-05, |
|
"loss": 0.0308, |
|
"step": 223000 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 1.1832483692375159e-05, |
|
"loss": 0.0378, |
|
"step": 223500 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 1.181421184381224e-05, |
|
"loss": 0.0424, |
|
"step": 224000 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 1.1795939995249322e-05, |
|
"loss": 0.0516, |
|
"step": 224500 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 1.1777668146686401e-05, |
|
"loss": 0.0349, |
|
"step": 225000 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 1.1759396298123482e-05, |
|
"loss": 0.0281, |
|
"step": 225500 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 1.1741124449560564e-05, |
|
"loss": 0.0424, |
|
"step": 226000 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 1.1722852600997643e-05, |
|
"loss": 0.033, |
|
"step": 226500 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 1.1704580752434724e-05, |
|
"loss": 0.0482, |
|
"step": 227000 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 1.1686308903871806e-05, |
|
"loss": 0.0361, |
|
"step": 227500 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 1.1668037055308887e-05, |
|
"loss": 0.0392, |
|
"step": 228000 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 1.1649765206745966e-05, |
|
"loss": 0.0357, |
|
"step": 228500 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 1.1631493358183048e-05, |
|
"loss": 0.0428, |
|
"step": 229000 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 1.1613221509620129e-05, |
|
"loss": 0.0337, |
|
"step": 229500 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 1.1594949661057211e-05, |
|
"loss": 0.0395, |
|
"step": 230000 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 1.157667781249429e-05, |
|
"loss": 0.0376, |
|
"step": 230500 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 1.1558405963931371e-05, |
|
"loss": 0.0342, |
|
"step": 231000 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 1.1540134115368453e-05, |
|
"loss": 0.0442, |
|
"step": 231500 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 1.1521862266805534e-05, |
|
"loss": 0.0445, |
|
"step": 232000 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 1.1503590418242613e-05, |
|
"loss": 0.0396, |
|
"step": 232500 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 1.1485318569679696e-05, |
|
"loss": 0.039, |
|
"step": 233000 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 1.1467046721116776e-05, |
|
"loss": 0.0394, |
|
"step": 233500 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 1.1448774872553857e-05, |
|
"loss": 0.0299, |
|
"step": 234000 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 1.1430503023990938e-05, |
|
"loss": 0.0251, |
|
"step": 234500 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 1.1412231175428018e-05, |
|
"loss": 0.0359, |
|
"step": 235000 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 1.13939593268651e-05, |
|
"loss": 0.0364, |
|
"step": 235500 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 1.137568747830218e-05, |
|
"loss": 0.0334, |
|
"step": 236000 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 1.135741562973926e-05, |
|
"loss": 0.0369, |
|
"step": 236500 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 1.1339143781176343e-05, |
|
"loss": 0.0326, |
|
"step": 237000 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 1.1320871932613424e-05, |
|
"loss": 0.0429, |
|
"step": 237500 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 1.1302600084050504e-05, |
|
"loss": 0.0433, |
|
"step": 238000 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 1.1284328235487585e-05, |
|
"loss": 0.0383, |
|
"step": 238500 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 1.1266056386924666e-05, |
|
"loss": 0.0335, |
|
"step": 239000 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 1.1247784538361748e-05, |
|
"loss": 0.0388, |
|
"step": 239500 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 1.1229512689798827e-05, |
|
"loss": 0.0349, |
|
"step": 240000 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 1.121124084123591e-05, |
|
"loss": 0.0311, |
|
"step": 240500 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 1.119296899267299e-05, |
|
"loss": 0.0453, |
|
"step": 241000 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 1.117469714411007e-05, |
|
"loss": 0.0392, |
|
"step": 241500 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 1.1156425295547152e-05, |
|
"loss": 0.0429, |
|
"step": 242000 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 1.1138153446984232e-05, |
|
"loss": 0.0441, |
|
"step": 242500 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 1.1119881598421315e-05, |
|
"loss": 0.0365, |
|
"step": 243000 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 1.1101609749858394e-05, |
|
"loss": 0.0374, |
|
"step": 243500 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 1.1083337901295474e-05, |
|
"loss": 0.0434, |
|
"step": 244000 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 1.1065066052732557e-05, |
|
"loss": 0.0411, |
|
"step": 244500 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 1.1046794204169637e-05, |
|
"loss": 0.0409, |
|
"step": 245000 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 1.1028522355606716e-05, |
|
"loss": 0.0373, |
|
"step": 245500 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 1.1010250507043799e-05, |
|
"loss": 0.0427, |
|
"step": 246000 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 1.099197865848088e-05, |
|
"loss": 0.0351, |
|
"step": 246500 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 1.0973706809917962e-05, |
|
"loss": 0.0285, |
|
"step": 247000 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 1.0955434961355041e-05, |
|
"loss": 0.0364, |
|
"step": 247500 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 1.0937163112792122e-05, |
|
"loss": 0.0448, |
|
"step": 248000 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 1.0918891264229204e-05, |
|
"loss": 0.0333, |
|
"step": 248500 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 1.0900619415666285e-05, |
|
"loss": 0.0427, |
|
"step": 249000 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 1.0882347567103364e-05, |
|
"loss": 0.0393, |
|
"step": 249500 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 1.0864075718540446e-05, |
|
"loss": 0.0413, |
|
"step": 250000 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 1.0845803869977527e-05, |
|
"loss": 0.0322, |
|
"step": 250500 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 1.082753202141461e-05, |
|
"loss": 0.0341, |
|
"step": 251000 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 1.0809260172851688e-05, |
|
"loss": 0.0363, |
|
"step": 251500 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 1.0790988324288769e-05, |
|
"loss": 0.0316, |
|
"step": 252000 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 1.0772716475725851e-05, |
|
"loss": 0.0393, |
|
"step": 252500 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 1.075444462716293e-05, |
|
"loss": 0.0328, |
|
"step": 253000 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 1.0736172778600011e-05, |
|
"loss": 0.0354, |
|
"step": 253500 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 1.0717900930037093e-05, |
|
"loss": 0.03, |
|
"step": 254000 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 1.0699629081474174e-05, |
|
"loss": 0.0423, |
|
"step": 254500 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 1.0681357232911253e-05, |
|
"loss": 0.04, |
|
"step": 255000 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 1.0663085384348335e-05, |
|
"loss": 0.029, |
|
"step": 255500 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 1.0644813535785416e-05, |
|
"loss": 0.039, |
|
"step": 256000 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 1.0626541687222499e-05, |
|
"loss": 0.0335, |
|
"step": 256500 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 1.0608269838659578e-05, |
|
"loss": 0.0317, |
|
"step": 257000 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 1.0589997990096658e-05, |
|
"loss": 0.0515, |
|
"step": 257500 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 1.057172614153374e-05, |
|
"loss": 0.0376, |
|
"step": 258000 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 1.0553454292970821e-05, |
|
"loss": 0.0413, |
|
"step": 258500 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 1.05351824444079e-05, |
|
"loss": 0.0363, |
|
"step": 259000 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 1.0516910595844983e-05, |
|
"loss": 0.0336, |
|
"step": 259500 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 1.0498638747282063e-05, |
|
"loss": 0.0405, |
|
"step": 260000 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 1.0480366898719146e-05, |
|
"loss": 0.042, |
|
"step": 260500 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 1.0462095050156225e-05, |
|
"loss": 0.0366, |
|
"step": 261000 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 1.0443823201593306e-05, |
|
"loss": 0.041, |
|
"step": 261500 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 1.0425551353030388e-05, |
|
"loss": 0.0277, |
|
"step": 262000 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 1.0407279504467467e-05, |
|
"loss": 0.0321, |
|
"step": 262500 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 1.0389007655904548e-05, |
|
"loss": 0.0412, |
|
"step": 263000 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 1.037073580734163e-05, |
|
"loss": 0.0411, |
|
"step": 263500 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 1.035246395877871e-05, |
|
"loss": 0.0294, |
|
"step": 264000 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 1.033419211021579e-05, |
|
"loss": 0.0361, |
|
"step": 264500 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 1.0315920261652872e-05, |
|
"loss": 0.0506, |
|
"step": 265000 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 1.0297648413089953e-05, |
|
"loss": 0.038, |
|
"step": 265500 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 1.0279376564527035e-05, |
|
"loss": 0.0323, |
|
"step": 266000 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 1.0261104715964114e-05, |
|
"loss": 0.0346, |
|
"step": 266500 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 1.0242832867401195e-05, |
|
"loss": 0.0336, |
|
"step": 267000 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 1.0224561018838277e-05, |
|
"loss": 0.0387, |
|
"step": 267500 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 1.0206289170275358e-05, |
|
"loss": 0.0362, |
|
"step": 268000 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 1.0188017321712437e-05, |
|
"loss": 0.0304, |
|
"step": 268500 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 1.016974547314952e-05, |
|
"loss": 0.038, |
|
"step": 269000 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 1.01514736245866e-05, |
|
"loss": 0.036, |
|
"step": 269500 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 1.013320177602368e-05, |
|
"loss": 0.0385, |
|
"step": 270000 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 1.0114929927460762e-05, |
|
"loss": 0.033, |
|
"step": 270500 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 1.0096658078897842e-05, |
|
"loss": 0.0283, |
|
"step": 271000 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 1.0078386230334925e-05, |
|
"loss": 0.0372, |
|
"step": 271500 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 1.0060114381772004e-05, |
|
"loss": 0.0387, |
|
"step": 272000 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 1.0041842533209086e-05, |
|
"loss": 0.0378, |
|
"step": 272500 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 1.0023570684646167e-05, |
|
"loss": 0.0308, |
|
"step": 273000 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 1.0005298836083247e-05, |
|
"loss": 0.0363, |
|
"step": 273500 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 9.987026987520328e-06, |
|
"loss": 0.0289, |
|
"step": 274000 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 9.968755138957409e-06, |
|
"loss": 0.0329, |
|
"step": 274500 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 9.95048329039449e-06, |
|
"loss": 0.0385, |
|
"step": 275000 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 9.93221144183157e-06, |
|
"loss": 0.0415, |
|
"step": 275500 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 9.913939593268653e-06, |
|
"loss": 0.0353, |
|
"step": 276000 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 9.895667744705733e-06, |
|
"loss": 0.0308, |
|
"step": 276500 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 9.877395896142814e-06, |
|
"loss": 0.0356, |
|
"step": 277000 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 9.859124047579895e-06, |
|
"loss": 0.0348, |
|
"step": 277500 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 9.840852199016975e-06, |
|
"loss": 0.0439, |
|
"step": 278000 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 9.822580350454056e-06, |
|
"loss": 0.0329, |
|
"step": 278500 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 9.804308501891137e-06, |
|
"loss": 0.0386, |
|
"step": 279000 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 9.786036653328217e-06, |
|
"loss": 0.0445, |
|
"step": 279500 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 9.767764804765298e-06, |
|
"loss": 0.0385, |
|
"step": 280000 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 9.74949295620238e-06, |
|
"loss": 0.0308, |
|
"step": 280500 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 9.73122110763946e-06, |
|
"loss": 0.0322, |
|
"step": 281000 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 9.712949259076542e-06, |
|
"loss": 0.0424, |
|
"step": 281500 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 9.694677410513623e-06, |
|
"loss": 0.0274, |
|
"step": 282000 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 9.676405561950703e-06, |
|
"loss": 0.0471, |
|
"step": 282500 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 9.658133713387784e-06, |
|
"loss": 0.0425, |
|
"step": 283000 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 9.639861864824865e-06, |
|
"loss": 0.0353, |
|
"step": 283500 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 9.621590016261945e-06, |
|
"loss": 0.0364, |
|
"step": 284000 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 9.603318167699028e-06, |
|
"loss": 0.0258, |
|
"step": 284500 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 9.585046319136107e-06, |
|
"loss": 0.0458, |
|
"step": 285000 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 9.56677447057319e-06, |
|
"loss": 0.0319, |
|
"step": 285500 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 9.54850262201027e-06, |
|
"loss": 0.045, |
|
"step": 286000 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 9.53023077344735e-06, |
|
"loss": 0.0317, |
|
"step": 286500 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 9.511958924884431e-06, |
|
"loss": 0.0307, |
|
"step": 287000 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 9.493687076321512e-06, |
|
"loss": 0.0373, |
|
"step": 287500 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 9.475415227758593e-06, |
|
"loss": 0.0294, |
|
"step": 288000 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 9.457143379195673e-06, |
|
"loss": 0.038, |
|
"step": 288500 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 9.438871530632754e-06, |
|
"loss": 0.0328, |
|
"step": 289000 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 9.420599682069835e-06, |
|
"loss": 0.0416, |
|
"step": 289500 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 9.402327833506917e-06, |
|
"loss": 0.033, |
|
"step": 290000 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 9.384055984943998e-06, |
|
"loss": 0.0253, |
|
"step": 290500 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 9.365784136381079e-06, |
|
"loss": 0.0316, |
|
"step": 291000 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 9.34751228781816e-06, |
|
"loss": 0.0327, |
|
"step": 291500 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 9.32924043925524e-06, |
|
"loss": 0.0287, |
|
"step": 292000 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 9.31096859069232e-06, |
|
"loss": 0.0327, |
|
"step": 292500 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 9.292696742129403e-06, |
|
"loss": 0.0341, |
|
"step": 293000 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 9.274424893566482e-06, |
|
"loss": 0.0399, |
|
"step": 293500 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 9.256153045003565e-06, |
|
"loss": 0.021, |
|
"step": 294000 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 9.237881196440645e-06, |
|
"loss": 0.0511, |
|
"step": 294500 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 9.219609347877726e-06, |
|
"loss": 0.0241, |
|
"step": 295000 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 9.201337499314807e-06, |
|
"loss": 0.0296, |
|
"step": 295500 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 9.183065650751887e-06, |
|
"loss": 0.037, |
|
"step": 296000 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 9.164793802188968e-06, |
|
"loss": 0.0362, |
|
"step": 296500 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 9.146521953626049e-06, |
|
"loss": 0.0349, |
|
"step": 297000 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 9.12825010506313e-06, |
|
"loss": 0.0367, |
|
"step": 297500 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 9.10997825650021e-06, |
|
"loss": 0.0343, |
|
"step": 298000 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 9.091706407937293e-06, |
|
"loss": 0.0296, |
|
"step": 298500 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 9.073434559374372e-06, |
|
"loss": 0.0305, |
|
"step": 299000 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 9.055162710811454e-06, |
|
"loss": 0.0389, |
|
"step": 299500 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 9.036890862248535e-06, |
|
"loss": 0.0324, |
|
"step": 300000 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 9.018619013685615e-06, |
|
"loss": 0.0243, |
|
"step": 300500 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 9.000347165122696e-06, |
|
"loss": 0.0287, |
|
"step": 301000 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 8.982075316559777e-06, |
|
"loss": 0.0306, |
|
"step": 301500 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 8.963803467996857e-06, |
|
"loss": 0.0323, |
|
"step": 302000 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 8.94553161943394e-06, |
|
"loss": 0.0309, |
|
"step": 302500 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 8.927259770871019e-06, |
|
"loss": 0.031, |
|
"step": 303000 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 8.908987922308101e-06, |
|
"loss": 0.0367, |
|
"step": 303500 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 8.890716073745182e-06, |
|
"loss": 0.0331, |
|
"step": 304000 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 8.872444225182263e-06, |
|
"loss": 0.0323, |
|
"step": 304500 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 8.854172376619343e-06, |
|
"loss": 0.0362, |
|
"step": 305000 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 8.835900528056424e-06, |
|
"loss": 0.0269, |
|
"step": 305500 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 8.817628679493505e-06, |
|
"loss": 0.0395, |
|
"step": 306000 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 8.799356830930585e-06, |
|
"loss": 0.0424, |
|
"step": 306500 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 8.781084982367666e-06, |
|
"loss": 0.0296, |
|
"step": 307000 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 8.762813133804747e-06, |
|
"loss": 0.0375, |
|
"step": 307500 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 8.74454128524183e-06, |
|
"loss": 0.0321, |
|
"step": 308000 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 8.72626943667891e-06, |
|
"loss": 0.0321, |
|
"step": 308500 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 8.70799758811599e-06, |
|
"loss": 0.0331, |
|
"step": 309000 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 8.689725739553071e-06, |
|
"loss": 0.0259, |
|
"step": 309500 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 8.671453890990152e-06, |
|
"loss": 0.0258, |
|
"step": 310000 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 8.653182042427233e-06, |
|
"loss": 0.035, |
|
"step": 310500 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 8.634910193864315e-06, |
|
"loss": 0.027, |
|
"step": 311000 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 8.616638345301394e-06, |
|
"loss": 0.0329, |
|
"step": 311500 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 8.598366496738476e-06, |
|
"loss": 0.021, |
|
"step": 312000 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 8.580094648175557e-06, |
|
"loss": 0.0329, |
|
"step": 312500 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 8.561822799612638e-06, |
|
"loss": 0.0496, |
|
"step": 313000 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 8.543550951049719e-06, |
|
"loss": 0.0434, |
|
"step": 313500 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 8.5252791024868e-06, |
|
"loss": 0.0299, |
|
"step": 314000 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 8.50700725392388e-06, |
|
"loss": 0.0358, |
|
"step": 314500 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 8.48873540536096e-06, |
|
"loss": 0.0337, |
|
"step": 315000 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 8.470463556798041e-06, |
|
"loss": 0.0348, |
|
"step": 315500 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 8.452191708235122e-06, |
|
"loss": 0.0372, |
|
"step": 316000 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 8.433919859672204e-06, |
|
"loss": 0.0284, |
|
"step": 316500 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 8.415648011109283e-06, |
|
"loss": 0.0297, |
|
"step": 317000 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 8.397376162546366e-06, |
|
"loss": 0.0294, |
|
"step": 317500 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 8.379104313983447e-06, |
|
"loss": 0.0302, |
|
"step": 318000 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 8.360832465420527e-06, |
|
"loss": 0.0305, |
|
"step": 318500 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 8.342560616857608e-06, |
|
"loss": 0.0222, |
|
"step": 319000 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 8.324288768294689e-06, |
|
"loss": 0.0289, |
|
"step": 319500 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 8.30601691973177e-06, |
|
"loss": 0.0253, |
|
"step": 320000 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 8.287745071168852e-06, |
|
"loss": 0.0365, |
|
"step": 320500 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 8.26947322260593e-06, |
|
"loss": 0.0357, |
|
"step": 321000 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 8.251201374043013e-06, |
|
"loss": 0.0365, |
|
"step": 321500 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 8.232929525480094e-06, |
|
"loss": 0.0324, |
|
"step": 322000 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 8.214657676917175e-06, |
|
"loss": 0.0319, |
|
"step": 322500 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 8.196385828354255e-06, |
|
"loss": 0.0339, |
|
"step": 323000 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 8.178113979791336e-06, |
|
"loss": 0.0271, |
|
"step": 323500 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 8.159842131228417e-06, |
|
"loss": 0.031, |
|
"step": 324000 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 8.141570282665497e-06, |
|
"loss": 0.0304, |
|
"step": 324500 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 8.123298434102578e-06, |
|
"loss": 0.0335, |
|
"step": 325000 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 8.105026585539659e-06, |
|
"loss": 0.0321, |
|
"step": 325500 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 8.086754736976741e-06, |
|
"loss": 0.0231, |
|
"step": 326000 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 8.068482888413822e-06, |
|
"loss": 0.0433, |
|
"step": 326500 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 8.050211039850903e-06, |
|
"loss": 0.0387, |
|
"step": 327000 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 8.031939191287983e-06, |
|
"loss": 0.038, |
|
"step": 327500 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 8.013667342725064e-06, |
|
"loss": 0.0336, |
|
"step": 328000 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 7.995395494162145e-06, |
|
"loss": 0.0404, |
|
"step": 328500 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 7.977123645599227e-06, |
|
"loss": 0.025, |
|
"step": 329000 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 7.958851797036306e-06, |
|
"loss": 0.0259, |
|
"step": 329500 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 7.940579948473388e-06, |
|
"loss": 0.0327, |
|
"step": 330000 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 7.922308099910469e-06, |
|
"loss": 0.0227, |
|
"step": 330500 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 7.90403625134755e-06, |
|
"loss": 0.0393, |
|
"step": 331000 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 7.88576440278463e-06, |
|
"loss": 0.0336, |
|
"step": 331500 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 7.867492554221711e-06, |
|
"loss": 0.026, |
|
"step": 332000 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 7.849220705658792e-06, |
|
"loss": 0.0341, |
|
"step": 332500 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 7.830948857095873e-06, |
|
"loss": 0.0322, |
|
"step": 333000 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 7.812677008532953e-06, |
|
"loss": 0.0307, |
|
"step": 333500 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 7.794405159970034e-06, |
|
"loss": 0.0321, |
|
"step": 334000 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 7.776133311407116e-06, |
|
"loss": 0.0319, |
|
"step": 334500 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 7.757861462844195e-06, |
|
"loss": 0.0265, |
|
"step": 335000 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 7.739589614281278e-06, |
|
"loss": 0.0357, |
|
"step": 335500 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 7.721317765718359e-06, |
|
"loss": 0.0284, |
|
"step": 336000 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 7.70304591715544e-06, |
|
"loss": 0.0337, |
|
"step": 336500 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 7.68477406859252e-06, |
|
"loss": 0.0263, |
|
"step": 337000 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 7.6665022200296e-06, |
|
"loss": 0.0292, |
|
"step": 337500 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 7.648230371466681e-06, |
|
"loss": 0.0239, |
|
"step": 338000 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 7.629958522903764e-06, |
|
"loss": 0.0426, |
|
"step": 338500 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 7.6116866743408436e-06, |
|
"loss": 0.0303, |
|
"step": 339000 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 7.593414825777925e-06, |
|
"loss": 0.0347, |
|
"step": 339500 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 7.575142977215005e-06, |
|
"loss": 0.037, |
|
"step": 340000 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 7.5568711286520865e-06, |
|
"loss": 0.0295, |
|
"step": 340500 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 7.538599280089167e-06, |
|
"loss": 0.0261, |
|
"step": 341000 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 7.520327431526249e-06, |
|
"loss": 0.0329, |
|
"step": 341500 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 7.502055582963329e-06, |
|
"loss": 0.0392, |
|
"step": 342000 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 7.483783734400409e-06, |
|
"loss": 0.0299, |
|
"step": 342500 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 7.465511885837491e-06, |
|
"loss": 0.0313, |
|
"step": 343000 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 7.4472400372745716e-06, |
|
"loss": 0.0312, |
|
"step": 343500 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 7.428968188711653e-06, |
|
"loss": 0.0329, |
|
"step": 344000 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 7.410696340148733e-06, |
|
"loss": 0.0227, |
|
"step": 344500 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 7.3924244915858145e-06, |
|
"loss": 0.0264, |
|
"step": 345000 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 7.374152643022895e-06, |
|
"loss": 0.0277, |
|
"step": 345500 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 7.355880794459977e-06, |
|
"loss": 0.0301, |
|
"step": 346000 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 7.337608945897057e-06, |
|
"loss": 0.035, |
|
"step": 346500 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 7.319337097334138e-06, |
|
"loss": 0.0374, |
|
"step": 347000 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 7.301065248771219e-06, |
|
"loss": 0.0193, |
|
"step": 347500 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 7.2827934002083e-06, |
|
"loss": 0.0264, |
|
"step": 348000 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 7.26452155164538e-06, |
|
"loss": 0.034, |
|
"step": 348500 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 7.246249703082462e-06, |
|
"loss": 0.0288, |
|
"step": 349000 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 7.2279778545195425e-06, |
|
"loss": 0.0323, |
|
"step": 349500 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 7.209706005956624e-06, |
|
"loss": 0.0279, |
|
"step": 350000 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 7.191434157393704e-06, |
|
"loss": 0.0285, |
|
"step": 350500 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 7.173162308830785e-06, |
|
"loss": 0.0383, |
|
"step": 351000 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 7.154890460267866e-06, |
|
"loss": 0.033, |
|
"step": 351500 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 7.136618611704946e-06, |
|
"loss": 0.0347, |
|
"step": 352000 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 7.1183467631420275e-06, |
|
"loss": 0.0314, |
|
"step": 352500 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 7.100074914579108e-06, |
|
"loss": 0.0306, |
|
"step": 353000 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 7.08180306601619e-06, |
|
"loss": 0.0464, |
|
"step": 353500 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 7.06353121745327e-06, |
|
"loss": 0.0314, |
|
"step": 354000 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 7.045259368890351e-06, |
|
"loss": 0.0206, |
|
"step": 354500 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 7.026987520327432e-06, |
|
"loss": 0.0206, |
|
"step": 355000 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 7.008715671764513e-06, |
|
"loss": 0.0222, |
|
"step": 355500 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 6.990443823201593e-06, |
|
"loss": 0.0324, |
|
"step": 356000 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 6.972171974638675e-06, |
|
"loss": 0.0431, |
|
"step": 356500 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 6.9539001260757555e-06, |
|
"loss": 0.0363, |
|
"step": 357000 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 6.935628277512837e-06, |
|
"loss": 0.0307, |
|
"step": 357500 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 6.917356428949917e-06, |
|
"loss": 0.0223, |
|
"step": 358000 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 6.8990845803869985e-06, |
|
"loss": 0.0276, |
|
"step": 358500 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 6.880812731824079e-06, |
|
"loss": 0.0283, |
|
"step": 359000 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 6.862540883261161e-06, |
|
"loss": 0.0323, |
|
"step": 359500 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 6.844269034698241e-06, |
|
"loss": 0.0323, |
|
"step": 360000 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 6.825997186135321e-06, |
|
"loss": 0.0334, |
|
"step": 360500 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 6.807725337572403e-06, |
|
"loss": 0.0325, |
|
"step": 361000 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 6.7894534890094835e-06, |
|
"loss": 0.0333, |
|
"step": 361500 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 6.771181640446565e-06, |
|
"loss": 0.0287, |
|
"step": 362000 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 6.752909791883645e-06, |
|
"loss": 0.0192, |
|
"step": 362500 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 6.7346379433207264e-06, |
|
"loss": 0.03, |
|
"step": 363000 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 6.716366094757807e-06, |
|
"loss": 0.0238, |
|
"step": 363500 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 6.698094246194889e-06, |
|
"loss": 0.0262, |
|
"step": 364000 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 6.6798223976319685e-06, |
|
"loss": 0.0262, |
|
"step": 364500 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 0.14961867034435272, |
|
"eval_runtime": 1674.7145, |
|
"eval_samples_per_second": 138.344, |
|
"eval_steps_per_second": 13.835, |
|
"step": 364860 |
|
} |
|
], |
|
"max_steps": 547290, |
|
"num_train_epochs": 3, |
|
"total_flos": 7.150255099702272e+17, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|