{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.6484149855907781, "eval_steps": 250, "global_step": 2250, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00028818443804034583, "grad_norm": 342.1254510360843, "learning_rate": 0.0, "loss": 5.5528, "step": 1 }, { "epoch": 0.00028818443804034583, "eval_websight_new_IoU": 0.0, "eval_websight_new_MAE_x": 683.8400268554688, "eval_websight_new_MAE_y": 817.5599975585938, "eval_websight_new_coord_probability": 0.0006622532964684069, "eval_websight_new_correct_token_probability": 3.301670119526534e-07, "eval_websight_new_inside_bbox": 0.0, "eval_websight_new_loss": 5.271888256072998, "eval_websight_new_range_100_probability": 6.324639252852648e-05, "eval_websight_new_range_20_probability": 1.297366907238029e-05, "eval_websight_new_range_5_probability": 3.246676669732551e-06, "eval_websight_new_runtime": 34.3296, "eval_websight_new_samples_per_second": 1.456, "eval_websight_new_steps_per_second": 0.204, "step": 1 }, { "epoch": 0.00028818443804034583, "eval_seeclick_IoU": 0.029257936403155327, "eval_seeclick_MAE_x": 728.1400146484375, "eval_seeclick_MAE_y": 768.47998046875, "eval_seeclick_coord_probability": 0.0003137658059131354, "eval_seeclick_correct_token_probability": 1.5348230419931497e-07, "eval_seeclick_inside_bbox": 0.10000000149011612, "eval_seeclick_loss": 6.226981163024902, "eval_seeclick_range_100_probability": 3.0610663088737056e-05, "eval_seeclick_range_20_probability": 6.1479131545638666e-06, "eval_seeclick_range_5_probability": 1.5359454437202658e-06, "eval_seeclick_runtime": 43.0329, "eval_seeclick_samples_per_second": 1.162, "eval_seeclick_steps_per_second": 0.163, "step": 1 }, { "epoch": 0.00028818443804034583, "eval_icons_IoU": 0.004958232864737511, "eval_icons_MAE_x": 745.6400146484375, "eval_icons_MAE_y": 752.8400268554688, "eval_icons_coord_probability": 0.0001312530366703868, "eval_icons_correct_token_probability": 6.412179232029303e-08, "eval_icons_inside_bbox": 0.0, "eval_icons_loss": 5.205017566680908, "eval_icons_range_100_probability": 1.1899669516424183e-05, "eval_icons_range_20_probability": 2.5593178634153446e-06, "eval_icons_range_5_probability": 6.412645348063961e-07, "eval_icons_runtime": 43.1355, "eval_icons_samples_per_second": 1.159, "eval_icons_steps_per_second": 0.162, "step": 1 }, { "epoch": 0.0005763688760806917, "grad_norm": 374.14019160055807, "learning_rate": 1.2553691692674561e-06, "loss": 6.1136, "step": 2 }, { "epoch": 0.0008645533141210375, "grad_norm": 347.4988078431115, "learning_rate": 1.989713057850388e-06, "loss": 5.0944, "step": 3 }, { "epoch": 0.0011527377521613833, "grad_norm": 179.43459303960572, "learning_rate": 2.5107383385349122e-06, "loss": 5.0628, "step": 4 }, { "epoch": 0.001440922190201729, "grad_norm": 264.3045842109189, "learning_rate": 2.914876943577515e-06, "loss": 5.98, "step": 5 }, { "epoch": 0.001729106628242075, "grad_norm": 159.84065752881398, "learning_rate": 3.245082227117844e-06, "loss": 4.9271, "step": 6 }, { "epoch": 0.0020172910662824206, "grad_norm": 93.20770627312804, "learning_rate": 3.524266816342358e-06, "loss": 4.8324, "step": 7 }, { "epoch": 0.0023054755043227667, "grad_norm": 134.3055088466808, "learning_rate": 3.7661075078023677e-06, "loss": 5.6501, "step": 8 }, { "epoch": 0.0025936599423631124, "grad_norm": 70.87239560354378, "learning_rate": 3.979426115700776e-06, "loss": 4.7434, "step": 9 }, { "epoch": 0.002881844380403458, "grad_norm": 50.71713625909726, "learning_rate": 4.170246112844972e-06, "loss": 4.662, "step": 10 }, { "epoch": 0.003170028818443804, "grad_norm": 57.917505846640466, "learning_rate": 4.342863797226275e-06, "loss": 5.3537, "step": 11 }, { "epoch": 0.00345821325648415, "grad_norm": 42.88832778930361, "learning_rate": 4.5004513963853e-06, "loss": 4.4709, "step": 12 }, { "epoch": 0.0037463976945244955, "grad_norm": 42.62371538537093, "learning_rate": 4.645417934887083e-06, "loss": 4.3543, "step": 13 }, { "epoch": 0.004034582132564841, "grad_norm": 42.273228097799816, "learning_rate": 4.779635985609814e-06, "loss": 5.129, "step": 14 }, { "epoch": 0.004322766570605188, "grad_norm": 41.22565825127231, "learning_rate": 4.904590001427903e-06, "loss": 4.2895, "step": 15 }, { "epoch": 0.004610951008645533, "grad_norm": 36.4504773077764, "learning_rate": 5.0214766770698244e-06, "loss": 4.1565, "step": 16 }, { "epoch": 0.004899135446685879, "grad_norm": 34.29091312887054, "learning_rate": 5.131274831432035e-06, "loss": 4.8424, "step": 17 }, { "epoch": 0.005187319884726225, "grad_norm": 48.649272813076784, "learning_rate": 5.234795284968231e-06, "loss": 4.0544, "step": 18 }, { "epoch": 0.00547550432276657, "grad_norm": 31.959874950729677, "learning_rate": 5.3327172336600444e-06, "loss": 3.8417, "step": 19 }, { "epoch": 0.005763688760806916, "grad_norm": 28.61342603792043, "learning_rate": 5.425615282112428e-06, "loss": 4.437, "step": 20 }, { "epoch": 0.006051873198847263, "grad_norm": 47.66734385668608, "learning_rate": 5.513979874192746e-06, "loss": 3.8685, "step": 21 }, { "epoch": 0.006340057636887608, "grad_norm": 28.43431869602734, "learning_rate": 5.598232966493732e-06, "loss": 3.6345, "step": 22 }, { "epoch": 0.006628242074927954, "grad_norm": 27.765318083952838, "learning_rate": 5.6787402149051605e-06, "loss": 4.1856, "step": 23 }, { "epoch": 0.0069164265129683, "grad_norm": 51.311972962556034, "learning_rate": 5.755820565652757e-06, "loss": 3.6776, "step": 24 }, { "epoch": 0.007204610951008645, "grad_norm": 30.939045377510407, "learning_rate": 5.82975388715503e-06, "loss": 3.4193, "step": 25 }, { "epoch": 0.007492795389048991, "grad_norm": 25.204320422929406, "learning_rate": 5.900787104154539e-06, "loss": 3.9317, "step": 26 }, { "epoch": 0.0077809798270893375, "grad_norm": 47.505367505045086, "learning_rate": 5.9691391735511625e-06, "loss": 3.427, "step": 27 }, { "epoch": 0.008069164265129682, "grad_norm": 43.00244755896901, "learning_rate": 6.03500515487727e-06, "loss": 3.1939, "step": 28 }, { "epoch": 0.008357348703170028, "grad_norm": 24.738277068462082, "learning_rate": 6.0985595661703904e-06, "loss": 3.6996, "step": 29 }, { "epoch": 0.008645533141210375, "grad_norm": 27.13279303141791, "learning_rate": 6.159959170695358e-06, "loss": 3.2539, "step": 30 }, { "epoch": 0.008933717579250721, "grad_norm": 30.518635329934547, "learning_rate": 6.219345306558267e-06, "loss": 3.0071, "step": 31 }, { "epoch": 0.009221902017291067, "grad_norm": 25.330093895707904, "learning_rate": 6.27684584633728e-06, "loss": 3.4751, "step": 32 }, { "epoch": 0.009510086455331412, "grad_norm": 20.378547867700338, "learning_rate": 6.332576855076663e-06, "loss": 3.0759, "step": 33 }, { "epoch": 0.009798270893371758, "grad_norm": 22.71599913260458, "learning_rate": 6.386644000699491e-06, "loss": 2.7783, "step": 34 }, { "epoch": 0.010086455331412104, "grad_norm": 27.406387142794618, "learning_rate": 6.4391437599198735e-06, "loss": 3.2274, "step": 35 }, { "epoch": 0.01037463976945245, "grad_norm": 19.64330264999394, "learning_rate": 6.490164454235688e-06, "loss": 2.8711, "step": 36 }, { "epoch": 0.010662824207492795, "grad_norm": 21.322369767837493, "learning_rate": 6.5397871439471684e-06, "loss": 2.5704, "step": 37 }, { "epoch": 0.01095100864553314, "grad_norm": 32.31738147342337, "learning_rate": 6.5880864029275e-06, "loss": 3.0204, "step": 38 }, { "epoch": 0.011239193083573486, "grad_norm": 20.961007392609183, "learning_rate": 6.635130992737469e-06, "loss": 2.6381, "step": 39 }, { "epoch": 0.011527377521613832, "grad_norm": 17.38274192908384, "learning_rate": 6.680984451379884e-06, "loss": 2.3938, "step": 40 }, { "epoch": 0.01181556195965418, "grad_norm": 19.507273735865926, "learning_rate": 6.725705609344599e-06, "loss": 2.7986, "step": 41 }, { "epoch": 0.012103746397694525, "grad_norm": 15.160391921438011, "learning_rate": 6.769349043460203e-06, "loss": 2.515, "step": 42 }, { "epoch": 0.01239193083573487, "grad_norm": 15.153798413143186, "learning_rate": 6.811965477335649e-06, "loss": 2.2172, "step": 43 }, { "epoch": 0.012680115273775217, "grad_norm": 15.60502737286468, "learning_rate": 6.853602135761187e-06, "loss": 2.5867, "step": 44 }, { "epoch": 0.012968299711815562, "grad_norm": 18.906378023267003, "learning_rate": 6.89430305927829e-06, "loss": 2.3077, "step": 45 }, { "epoch": 0.013256484149855908, "grad_norm": 22.483387730826415, "learning_rate": 6.934109384172617e-06, "loss": 2.0573, "step": 46 }, { "epoch": 0.013544668587896254, "grad_norm": 16.918339744078224, "learning_rate": 6.973059592352829e-06, "loss": 2.4301, "step": 47 }, { "epoch": 0.0138328530259366, "grad_norm": 14.359138587395373, "learning_rate": 7.011189734920213e-06, "loss": 2.1018, "step": 48 }, { "epoch": 0.014121037463976945, "grad_norm": 12.785390213380659, "learning_rate": 7.048533632684716e-06, "loss": 1.9024, "step": 49 }, { "epoch": 0.01440922190201729, "grad_norm": 13.822682100623124, "learning_rate": 7.085123056422486e-06, "loss": 2.2607, "step": 50 }, { "epoch": 0.014697406340057636, "grad_norm": 11.248470676887981, "learning_rate": 7.1209878892824225e-06, "loss": 1.9559, "step": 51 }, { "epoch": 0.014697406340057636, "eval_websight_new_IoU": 0.010778559371829033, "eval_websight_new_MAE_x": 363.0199890136719, "eval_websight_new_MAE_y": 421.5400085449219, "eval_websight_new_coord_probability": 0.0072825150564312935, "eval_websight_new_correct_token_probability": 5.725064966100035e-06, "eval_websight_new_inside_bbox": 0.019999999552965164, "eval_websight_new_loss": 1.8606456518173218, "eval_websight_new_range_100_probability": 0.0008474837522953749, "eval_websight_new_range_20_probability": 0.0001835049915825948, "eval_websight_new_range_5_probability": 4.652893039747141e-05, "eval_websight_new_runtime": 33.4595, "eval_websight_new_samples_per_second": 1.494, "eval_websight_new_steps_per_second": 0.209, "step": 51 }, { "epoch": 0.014697406340057636, "eval_seeclick_IoU": 0.009962150827050209, "eval_seeclick_MAE_x": 473.05999755859375, "eval_seeclick_MAE_y": 601.760009765625, "eval_seeclick_coord_probability": 0.007472059689462185, "eval_seeclick_correct_token_probability": 5.521962975763017e-06, "eval_seeclick_inside_bbox": 0.05999999865889549, "eval_seeclick_loss": 2.1386313438415527, "eval_seeclick_range_100_probability": 0.0008697480079717934, "eval_seeclick_range_20_probability": 0.00017912087787408382, "eval_seeclick_range_5_probability": 4.512786472332664e-05, "eval_seeclick_runtime": 42.4315, "eval_seeclick_samples_per_second": 1.178, "eval_seeclick_steps_per_second": 0.165, "step": 51 }, { "epoch": 0.014697406340057636, "eval_icons_IoU": 0.0015337333315983415, "eval_icons_MAE_x": 481.1199951171875, "eval_icons_MAE_y": 428.3800048828125, "eval_icons_coord_probability": 0.007426741532981396, "eval_icons_correct_token_probability": 5.428708846011432e-06, "eval_icons_inside_bbox": 0.0, "eval_icons_loss": 1.865750789642334, "eval_icons_range_100_probability": 0.0008250317187048495, "eval_icons_range_20_probability": 0.0002058930549537763, "eval_icons_range_5_probability": 5.214904376771301e-05, "eval_icons_runtime": 42.6726, "eval_icons_samples_per_second": 1.172, "eval_icons_steps_per_second": 0.164, "step": 51 }, { "epoch": 0.014985590778097982, "grad_norm": 9.772026011470567, "learning_rate": 7.156156273421995e-06, "loss": 1.7996, "step": 52 }, { "epoch": 0.01527377521613833, "grad_norm": 12.657681600064546, "learning_rate": 7.1906547426750725e-06, "loss": 2.0842, "step": 53 }, { "epoch": 0.015561959654178675, "grad_norm": 9.613342066404323, "learning_rate": 7.224508342818619e-06, "loss": 1.8363, "step": 54 }, { "epoch": 0.01585014409221902, "grad_norm": 11.090633439192665, "learning_rate": 7.25774074080379e-06, "loss": 1.7267, "step": 55 }, { "epoch": 0.016138328530259365, "grad_norm": 13.126846403161675, "learning_rate": 7.290374324144728e-06, "loss": 1.9524, "step": 56 }, { "epoch": 0.016426512968299712, "grad_norm": 7.940529281455624, "learning_rate": 7.3224302915104325e-06, "loss": 1.7391, "step": 57 }, { "epoch": 0.016714697406340056, "grad_norm": 10.327915839031606, "learning_rate": 7.3539287354378455e-06, "loss": 1.6814, "step": 58 }, { "epoch": 0.017002881844380403, "grad_norm": 12.465134904038113, "learning_rate": 7.38488871797435e-06, "loss": 1.8826, "step": 59 }, { "epoch": 0.01729106628242075, "grad_norm": 9.518374438515043, "learning_rate": 7.415328339962814e-06, "loss": 1.7086, "step": 60 }, { "epoch": 0.017579250720461095, "grad_norm": 10.860650734562737, "learning_rate": 7.4452648045998055e-06, "loss": 1.6188, "step": 61 }, { "epoch": 0.017867435158501442, "grad_norm": 7.824123766994198, "learning_rate": 7.474714475825724e-06, "loss": 1.7929, "step": 62 }, { "epoch": 0.018155619596541786, "grad_norm": 8.104791112765197, "learning_rate": 7.503692932043134e-06, "loss": 1.6061, "step": 63 }, { "epoch": 0.018443804034582133, "grad_norm": 8.993205128398746, "learning_rate": 7.532215015604735e-06, "loss": 1.5625, "step": 64 }, { "epoch": 0.018731988472622477, "grad_norm": 8.176567292713766, "learning_rate": 7.560294878464597e-06, "loss": 1.7126, "step": 65 }, { "epoch": 0.019020172910662825, "grad_norm": 7.132222732428861, "learning_rate": 7.587946024344118e-06, "loss": 1.5697, "step": 66 }, { "epoch": 0.01930835734870317, "grad_norm": 8.127988816552792, "learning_rate": 7.615181347727268e-06, "loss": 1.5083, "step": 67 }, { "epoch": 0.019596541786743516, "grad_norm": 10.280846062275057, "learning_rate": 7.642013169966947e-06, "loss": 1.6577, "step": 68 }, { "epoch": 0.01988472622478386, "grad_norm": 6.578790507069004, "learning_rate": 7.668453272755549e-06, "loss": 1.5532, "step": 69 }, { "epoch": 0.020172910662824207, "grad_norm": 10.222158144657401, "learning_rate": 7.69451292918733e-06, "loss": 1.4516, "step": 70 }, { "epoch": 0.020461095100864555, "grad_norm": 8.033521635824966, "learning_rate": 7.720202932617524e-06, "loss": 1.5968, "step": 71 }, { "epoch": 0.0207492795389049, "grad_norm": 6.283023849212295, "learning_rate": 7.745533623503144e-06, "loss": 1.4865, "step": 72 }, { "epoch": 0.021037463976945246, "grad_norm": 11.541429466909177, "learning_rate": 7.770514914392505e-06, "loss": 1.3943, "step": 73 }, { "epoch": 0.02132564841498559, "grad_norm": 6.810441213975766, "learning_rate": 7.795156313214624e-06, "loss": 1.5454, "step": 74 }, { "epoch": 0.021613832853025938, "grad_norm": 7.50975549320261, "learning_rate": 7.819466945005417e-06, "loss": 1.4353, "step": 75 }, { "epoch": 0.02190201729106628, "grad_norm": 7.460716357179955, "learning_rate": 7.843455572194956e-06, "loss": 1.3462, "step": 76 }, { "epoch": 0.02219020172910663, "grad_norm": 7.5211049677596, "learning_rate": 7.867130613568634e-06, "loss": 1.4812, "step": 77 }, { "epoch": 0.022478386167146973, "grad_norm": 6.112353977933034, "learning_rate": 7.890500162004926e-06, "loss": 1.3572, "step": 78 }, { "epoch": 0.02276657060518732, "grad_norm": 10.0445955512976, "learning_rate": 7.913572001083272e-06, "loss": 1.296, "step": 79 }, { "epoch": 0.023054755043227664, "grad_norm": 6.124066253475085, "learning_rate": 7.93635362064734e-06, "loss": 1.4276, "step": 80 }, { "epoch": 0.02334293948126801, "grad_norm": 6.316055475181084, "learning_rate": 7.958852231401552e-06, "loss": 1.3069, "step": 81 }, { "epoch": 0.02363112391930836, "grad_norm": 6.578049211563219, "learning_rate": 7.981074778612054e-06, "loss": 1.2461, "step": 82 }, { "epoch": 0.023919308357348703, "grad_norm": 7.444772664188824, "learning_rate": 8.003027954977264e-06, "loss": 1.3804, "step": 83 }, { "epoch": 0.02420749279538905, "grad_norm": 6.5602592933063395, "learning_rate": 8.024718212727658e-06, "loss": 1.27, "step": 84 }, { "epoch": 0.024495677233429394, "grad_norm": 6.942091189640846, "learning_rate": 8.04615177500955e-06, "loss": 1.1982, "step": 85 }, { "epoch": 0.02478386167146974, "grad_norm": 6.438612199361165, "learning_rate": 8.067334646603105e-06, "loss": 1.3237, "step": 86 }, { "epoch": 0.025072046109510086, "grad_norm": 5.547834949064384, "learning_rate": 8.088272624020777e-06, "loss": 1.2506, "step": 87 }, { "epoch": 0.025360230547550433, "grad_norm": 7.768644495811314, "learning_rate": 8.108971305028645e-06, "loss": 1.1535, "step": 88 }, { "epoch": 0.025648414985590777, "grad_norm": 6.0374882659496505, "learning_rate": 8.129436097629779e-06, "loss": 1.2679, "step": 89 }, { "epoch": 0.025936599423631124, "grad_norm": 5.491709410176797, "learning_rate": 8.149672228545746e-06, "loss": 1.1669, "step": 90 }, { "epoch": 0.02622478386167147, "grad_norm": 6.2627480902334725, "learning_rate": 8.16968475122944e-06, "loss": 1.104, "step": 91 }, { "epoch": 0.026512968299711816, "grad_norm": 6.052249863888647, "learning_rate": 8.189478553440074e-06, "loss": 1.225, "step": 92 }, { "epoch": 0.02680115273775216, "grad_norm": 5.386417663387223, "learning_rate": 8.209058364408657e-06, "loss": 1.1572, "step": 93 }, { "epoch": 0.027089337175792507, "grad_norm": 6.281249602777029, "learning_rate": 8.228428761620285e-06, "loss": 1.0714, "step": 94 }, { "epoch": 0.027377521613832854, "grad_norm": 5.836760674441374, "learning_rate": 8.247594177237559e-06, "loss": 1.1769, "step": 95 }, { "epoch": 0.0276657060518732, "grad_norm": 5.266384098127444, "learning_rate": 8.266558904187668e-06, "loss": 1.0825, "step": 96 }, { "epoch": 0.027953890489913546, "grad_norm": 5.929747177357997, "learning_rate": 8.28532710193407e-06, "loss": 1.0329, "step": 97 }, { "epoch": 0.02824207492795389, "grad_norm": 5.622834489781087, "learning_rate": 8.303902801952174e-06, "loss": 1.1473, "step": 98 }, { "epoch": 0.028530259365994237, "grad_norm": 5.150387365532939, "learning_rate": 8.32228991292705e-06, "loss": 1.0387, "step": 99 }, { "epoch": 0.02881844380403458, "grad_norm": 5.629354297769571, "learning_rate": 8.340492225689944e-06, "loss": 0.9917, "step": 100 }, { "epoch": 0.02910662824207493, "grad_norm": 5.70491743358872, "learning_rate": 8.358513417909158e-06, "loss": 1.1078, "step": 101 }, { "epoch": 0.029394812680115272, "grad_norm": 5.053891183636838, "learning_rate": 8.376357058549878e-06, "loss": 1.0231, "step": 102 }, { "epoch": 0.02968299711815562, "grad_norm": 5.675823683975542, "learning_rate": 8.394026612116405e-06, "loss": 0.9555, "step": 103 }, { "epoch": 0.029971181556195964, "grad_norm": 5.463972580079715, "learning_rate": 8.41152544268945e-06, "loss": 1.0771, "step": 104 }, { "epoch": 0.03025936599423631, "grad_norm": 4.999123462776384, "learning_rate": 8.42885681777026e-06, "loss": 0.9859, "step": 105 }, { "epoch": 0.03054755043227666, "grad_norm": 5.497431583114866, "learning_rate": 8.446023911942528e-06, "loss": 0.9313, "step": 106 }, { "epoch": 0.030835734870317003, "grad_norm": 5.391678295140468, "learning_rate": 8.463029810362388e-06, "loss": 1.0353, "step": 107 }, { "epoch": 0.03112391930835735, "grad_norm": 4.891677757189106, "learning_rate": 8.479877512086076e-06, "loss": 0.9647, "step": 108 }, { "epoch": 0.0314121037463977, "grad_norm": 5.610726694133473, "learning_rate": 8.496569933244228e-06, "loss": 0.8908, "step": 109 }, { "epoch": 0.03170028818443804, "grad_norm": 5.817735866127734, "learning_rate": 8.513109910071246e-06, "loss": 1.0143, "step": 110 }, { "epoch": 0.031988472622478385, "grad_norm": 4.871866053471662, "learning_rate": 8.529500201797555e-06, "loss": 0.9479, "step": 111 }, { "epoch": 0.03227665706051873, "grad_norm": 5.31764218824648, "learning_rate": 8.545743493412182e-06, "loss": 0.8705, "step": 112 }, { "epoch": 0.03256484149855908, "grad_norm": 5.081131222064519, "learning_rate": 8.561842398302535e-06, "loss": 0.9911, "step": 113 }, { "epoch": 0.032853025936599424, "grad_norm": 4.610845570193001, "learning_rate": 8.577799460777888e-06, "loss": 0.9058, "step": 114 }, { "epoch": 0.03314121037463977, "grad_norm": 5.478425064148529, "learning_rate": 8.593617158482677e-06, "loss": 0.8407, "step": 115 }, { "epoch": 0.03342939481268011, "grad_norm": 4.9308395705928385, "learning_rate": 8.609297904705302e-06, "loss": 0.9558, "step": 116 }, { "epoch": 0.03371757925072046, "grad_norm": 4.4313658226774, "learning_rate": 8.624844050587858e-06, "loss": 0.9037, "step": 117 }, { "epoch": 0.03400576368876081, "grad_norm": 5.1957219052400525, "learning_rate": 8.640257887241806e-06, "loss": 0.8201, "step": 118 }, { "epoch": 0.03429394812680115, "grad_norm": 4.637596099695778, "learning_rate": 8.655541647774393e-06, "loss": 0.9466, "step": 119 }, { "epoch": 0.0345821325648415, "grad_norm": 4.210947996449233, "learning_rate": 8.67069750923027e-06, "loss": 0.8812, "step": 120 }, { "epoch": 0.034870317002881845, "grad_norm": 4.656876963041862, "learning_rate": 8.68572759445255e-06, "loss": 0.7964, "step": 121 }, { "epoch": 0.03515850144092219, "grad_norm": 4.598584830368883, "learning_rate": 8.700633973867262e-06, "loss": 0.9246, "step": 122 }, { "epoch": 0.03544668587896253, "grad_norm": 4.2321321744684415, "learning_rate": 8.715418667194985e-06, "loss": 0.8862, "step": 123 }, { "epoch": 0.035734870317002884, "grad_norm": 4.600363613453042, "learning_rate": 8.73008364509318e-06, "loss": 0.7905, "step": 124 }, { "epoch": 0.03602305475504323, "grad_norm": 4.198077372961204, "learning_rate": 8.744630830732545e-06, "loss": 0.9113, "step": 125 }, { "epoch": 0.03631123919308357, "grad_norm": 3.7501430166268435, "learning_rate": 8.75906210131059e-06, "loss": 0.8415, "step": 126 }, { "epoch": 0.036599423631123916, "grad_norm": 4.419910726022508, "learning_rate": 8.773379289505366e-06, "loss": 0.752, "step": 127 }, { "epoch": 0.03688760806916427, "grad_norm": 3.997242466871886, "learning_rate": 8.787584184872193e-06, "loss": 0.8776, "step": 128 }, { "epoch": 0.03717579250720461, "grad_norm": 3.534947859211222, "learning_rate": 8.801678535186037e-06, "loss": 0.8254, "step": 129 }, { "epoch": 0.037463976945244955, "grad_norm": 4.428264660634081, "learning_rate": 8.815664047732054e-06, "loss": 0.7262, "step": 130 }, { "epoch": 0.037752161383285306, "grad_norm": 3.8981897659675044, "learning_rate": 8.829542390546686e-06, "loss": 0.8572, "step": 131 }, { "epoch": 0.03804034582132565, "grad_norm": 3.4427721693779323, "learning_rate": 8.843315193611574e-06, "loss": 0.8358, "step": 132 }, { "epoch": 0.038328530259365994, "grad_norm": 3.9905563522535226, "learning_rate": 8.856984050002403e-06, "loss": 0.7198, "step": 133 }, { "epoch": 0.03861671469740634, "grad_norm": 3.706844015046595, "learning_rate": 8.870550516994724e-06, "loss": 0.8521, "step": 134 }, { "epoch": 0.03890489913544669, "grad_norm": 3.186539722022302, "learning_rate": 8.884016117128679e-06, "loss": 0.8066, "step": 135 }, { "epoch": 0.03919308357348703, "grad_norm": 3.8254743339541295, "learning_rate": 8.897382339234405e-06, "loss": 0.712, "step": 136 }, { "epoch": 0.039481268011527376, "grad_norm": 3.500633177755673, "learning_rate": 8.910650639419908e-06, "loss": 0.8213, "step": 137 }, { "epoch": 0.03976945244956772, "grad_norm": 3.079033701156694, "learning_rate": 8.923822442023006e-06, "loss": 0.8025, "step": 138 }, { "epoch": 0.04005763688760807, "grad_norm": 3.548379275418023, "learning_rate": 8.936899140528881e-06, "loss": 0.6977, "step": 139 }, { "epoch": 0.040345821325648415, "grad_norm": 3.0662833147413306, "learning_rate": 8.949882098454784e-06, "loss": 0.7955, "step": 140 }, { "epoch": 0.04063400576368876, "grad_norm": 2.768024362479292, "learning_rate": 8.962772650203216e-06, "loss": 0.7805, "step": 141 }, { "epoch": 0.04092219020172911, "grad_norm": 3.2237959260312223, "learning_rate": 8.975572101884981e-06, "loss": 0.6739, "step": 142 }, { "epoch": 0.041210374639769454, "grad_norm": 2.7734152839051283, "learning_rate": 8.988281732113356e-06, "loss": 0.8002, "step": 143 }, { "epoch": 0.0414985590778098, "grad_norm": 2.250679674647897, "learning_rate": 9.0009027927706e-06, "loss": 0.7839, "step": 144 }, { "epoch": 0.04178674351585014, "grad_norm": 3.1539427375278724, "learning_rate": 9.013436509747905e-06, "loss": 0.6708, "step": 145 }, { "epoch": 0.04207492795389049, "grad_norm": 2.4309547666916878, "learning_rate": 9.025884083659961e-06, "loss": 0.7986, "step": 146 }, { "epoch": 0.042363112391930836, "grad_norm": 2.01309967729301, "learning_rate": 9.038246690535104e-06, "loss": 0.766, "step": 147 }, { "epoch": 0.04265129682997118, "grad_norm": 3.1015443582771693, "learning_rate": 9.05052548248208e-06, "loss": 0.6383, "step": 148 }, { "epoch": 0.042939481268011524, "grad_norm": 2.2747882705969196, "learning_rate": 9.062721588334355e-06, "loss": 0.7744, "step": 149 }, { "epoch": 0.043227665706051875, "grad_norm": 1.874678323320859, "learning_rate": 9.074836114272873e-06, "loss": 0.7578, "step": 150 }, { "epoch": 0.04351585014409222, "grad_norm": 2.73548781933127, "learning_rate": 9.08687014442814e-06, "loss": 0.6651, "step": 151 }, { "epoch": 0.04380403458213256, "grad_norm": 2.07172770752639, "learning_rate": 9.098824741462414e-06, "loss": 0.7656, "step": 152 }, { "epoch": 0.044092219020172914, "grad_norm": 2.426006794767885, "learning_rate": 9.11070094713281e-06, "loss": 0.7627, "step": 153 }, { "epoch": 0.04438040345821326, "grad_norm": 2.6608504533829533, "learning_rate": 9.12249978283609e-06, "loss": 0.6537, "step": 154 }, { "epoch": 0.0446685878962536, "grad_norm": 2.1136226301554877, "learning_rate": 9.134222250135783e-06, "loss": 0.7838, "step": 155 }, { "epoch": 0.044956772334293946, "grad_norm": 1.7542973386472682, "learning_rate": 9.145869331272382e-06, "loss": 0.7444, "step": 156 }, { "epoch": 0.0452449567723343, "grad_norm": 2.649915281124027, "learning_rate": 9.15744198965723e-06, "loss": 0.6349, "step": 157 }, { "epoch": 0.04553314121037464, "grad_norm": 1.9522249812385406, "learning_rate": 9.16894117035073e-06, "loss": 0.7693, "step": 158 }, { "epoch": 0.045821325648414984, "grad_norm": 1.8613859415577985, "learning_rate": 9.18036780052546e-06, "loss": 0.7551, "step": 159 }, { "epoch": 0.04610951008645533, "grad_norm": 2.627616979894939, "learning_rate": 9.191722789914796e-06, "loss": 0.6259, "step": 160 }, { "epoch": 0.04639769452449568, "grad_norm": 2.3971206406714805, "learning_rate": 9.203007031247519e-06, "loss": 0.7471, "step": 161 }, { "epoch": 0.04668587896253602, "grad_norm": 1.8047186312374244, "learning_rate": 9.214221400669006e-06, "loss": 0.7432, "step": 162 }, { "epoch": 0.04697406340057637, "grad_norm": 2.631275686948647, "learning_rate": 9.225366758149435e-06, "loss": 0.6261, "step": 163 }, { "epoch": 0.04726224783861672, "grad_norm": 2.345316372051671, "learning_rate": 9.23644394787951e-06, "loss": 0.737, "step": 164 }, { "epoch": 0.04755043227665706, "grad_norm": 1.8856785473095028, "learning_rate": 9.247453798654177e-06, "loss": 0.7196, "step": 165 }, { "epoch": 0.047838616714697406, "grad_norm": 2.6540847088644464, "learning_rate": 9.258397124244722e-06, "loss": 0.6193, "step": 166 }, { "epoch": 0.04812680115273775, "grad_norm": 1.9912082892186653, "learning_rate": 9.2692747237597e-06, "loss": 0.7346, "step": 167 }, { "epoch": 0.0484149855907781, "grad_norm": 1.9022891213962014, "learning_rate": 9.280087381995114e-06, "loss": 0.7211, "step": 168 }, { "epoch": 0.048703170028818445, "grad_norm": 2.855680994233883, "learning_rate": 9.290835869774166e-06, "loss": 0.6233, "step": 169 }, { "epoch": 0.04899135446685879, "grad_norm": 1.8992692898334764, "learning_rate": 9.301520944277006e-06, "loss": 0.7406, "step": 170 }, { "epoch": 0.04927953890489913, "grad_norm": 1.8061418922496177, "learning_rate": 9.31214334936082e-06, "loss": 0.7206, "step": 171 }, { "epoch": 0.04956772334293948, "grad_norm": 3.037852726523747, "learning_rate": 9.32270381587056e-06, "loss": 0.605, "step": 172 }, { "epoch": 0.04985590778097983, "grad_norm": 2.1650228003560645, "learning_rate": 9.333203061940696e-06, "loss": 0.7204, "step": 173 }, { "epoch": 0.05014409221902017, "grad_norm": 2.1427179809283405, "learning_rate": 9.343641793288234e-06, "loss": 0.7136, "step": 174 }, { "epoch": 0.05043227665706052, "grad_norm": 2.7496668086096823, "learning_rate": 9.354020703497389e-06, "loss": 0.6172, "step": 175 }, { "epoch": 0.050720461095100866, "grad_norm": 2.469362015895731, "learning_rate": 9.3643404742961e-06, "loss": 0.7237, "step": 176 }, { "epoch": 0.05100864553314121, "grad_norm": 2.1037607295004004, "learning_rate": 9.374601775824737e-06, "loss": 0.7092, "step": 177 }, { "epoch": 0.051296829971181554, "grad_norm": 2.6105303567931823, "learning_rate": 9.384805266897236e-06, "loss": 0.6011, "step": 178 }, { "epoch": 0.051585014409221905, "grad_norm": 2.4143791803196546, "learning_rate": 9.394951595254911e-06, "loss": 0.7142, "step": 179 }, { "epoch": 0.05187319884726225, "grad_norm": 1.7090641211386361, "learning_rate": 9.405041397813202e-06, "loss": 0.6932, "step": 180 }, { "epoch": 0.05216138328530259, "grad_norm": 2.999007860084135, "learning_rate": 9.41507530090159e-06, "loss": 0.6001, "step": 181 }, { "epoch": 0.05244956772334294, "grad_norm": 2.5026419005785936, "learning_rate": 9.425053920496896e-06, "loss": 0.6877, "step": 182 }, { "epoch": 0.05273775216138329, "grad_norm": 1.5888472945782244, "learning_rate": 9.434977862450193e-06, "loss": 0.6942, "step": 183 }, { "epoch": 0.05302593659942363, "grad_norm": 2.621111877977237, "learning_rate": 9.44484772270753e-06, "loss": 0.5951, "step": 184 }, { "epoch": 0.053314121037463975, "grad_norm": 1.9813131092916778, "learning_rate": 9.454664087524683e-06, "loss": 0.6937, "step": 185 }, { "epoch": 0.05360230547550432, "grad_norm": 1.870575239040046, "learning_rate": 9.464427533676113e-06, "loss": 0.7063, "step": 186 }, { "epoch": 0.05389048991354467, "grad_norm": 2.353331569351921, "learning_rate": 9.47413862865831e-06, "loss": 0.5961, "step": 187 }, { "epoch": 0.054178674351585014, "grad_norm": 1.794698720408445, "learning_rate": 9.483797930887741e-06, "loss": 0.6878, "step": 188 }, { "epoch": 0.05446685878962536, "grad_norm": 2.0116398279578016, "learning_rate": 9.493405989893522e-06, "loss": 0.6783, "step": 189 }, { "epoch": 0.05475504322766571, "grad_norm": 2.741855766208623, "learning_rate": 9.502963346505015e-06, "loss": 0.5801, "step": 190 }, { "epoch": 0.05504322766570605, "grad_norm": 2.38993948854855, "learning_rate": 9.512470533034512e-06, "loss": 0.711, "step": 191 }, { "epoch": 0.0553314121037464, "grad_norm": 1.9766927221810036, "learning_rate": 9.521928073455125e-06, "loss": 0.6893, "step": 192 }, { "epoch": 0.05561959654178674, "grad_norm": 2.6954209105434246, "learning_rate": 9.531336483574083e-06, "loss": 0.5935, "step": 193 }, { "epoch": 0.05590778097982709, "grad_norm": 2.2338669962164928, "learning_rate": 9.540696271201526e-06, "loss": 0.6966, "step": 194 }, { "epoch": 0.056195965417867436, "grad_norm": 1.5236329673106341, "learning_rate": 9.550007936314986e-06, "loss": 0.672, "step": 195 }, { "epoch": 0.05648414985590778, "grad_norm": 3.0857241897790577, "learning_rate": 9.559271971219628e-06, "loss": 0.5804, "step": 196 }, { "epoch": 0.056772334293948123, "grad_norm": 3.08494922150488, "learning_rate": 9.568488860704454e-06, "loss": 0.7248, "step": 197 }, { "epoch": 0.057060518731988474, "grad_norm": 1.7510197355931456, "learning_rate": 9.577659082194506e-06, "loss": 0.6751, "step": 198 }, { "epoch": 0.05734870317002882, "grad_norm": 3.8852483936339173, "learning_rate": 9.586783105899282e-06, "loss": 0.5727, "step": 199 }, { "epoch": 0.05763688760806916, "grad_norm": 3.929409013124462, "learning_rate": 9.595861394957398e-06, "loss": 0.6807, "step": 200 }, { "epoch": 0.05792507204610951, "grad_norm": 1.5785031461666155, "learning_rate": 9.604894405577658e-06, "loss": 0.6806, "step": 201 }, { "epoch": 0.05821325648414986, "grad_norm": 3.1878732741345246, "learning_rate": 9.613882587176614e-06, "loss": 0.5734, "step": 202 }, { "epoch": 0.0585014409221902, "grad_norm": 3.1854201233525523, "learning_rate": 9.622826382512747e-06, "loss": 0.6922, "step": 203 }, { "epoch": 0.058789625360230545, "grad_norm": 1.8526625111973063, "learning_rate": 9.631726227817333e-06, "loss": 0.6509, "step": 204 }, { "epoch": 0.059077809798270896, "grad_norm": 2.7294299798853348, "learning_rate": 9.640582552922113e-06, "loss": 0.5793, "step": 205 }, { "epoch": 0.05936599423631124, "grad_norm": 3.3941744620197984, "learning_rate": 9.64939578138386e-06, "loss": 0.6888, "step": 206 }, { "epoch": 0.059654178674351584, "grad_norm": 1.9883789922192263, "learning_rate": 9.658166330605937e-06, "loss": 0.6699, "step": 207 }, { "epoch": 0.05994236311239193, "grad_norm": 2.758267978856053, "learning_rate": 9.666894611956906e-06, "loss": 0.5638, "step": 208 }, { "epoch": 0.06023054755043228, "grad_norm": 1.8171103966983788, "learning_rate": 9.67558103088632e-06, "loss": 0.6893, "step": 209 }, { "epoch": 0.06051873198847262, "grad_norm": 2.1888003833329712, "learning_rate": 9.684225987037717e-06, "loss": 0.6822, "step": 210 }, { "epoch": 0.060806916426512966, "grad_norm": 2.6024308217775913, "learning_rate": 9.69282987435897e-06, "loss": 0.5748, "step": 211 }, { "epoch": 0.06109510086455332, "grad_norm": 1.8525049491547192, "learning_rate": 9.701393081209986e-06, "loss": 0.666, "step": 212 }, { "epoch": 0.06138328530259366, "grad_norm": 2.0029479305883666, "learning_rate": 9.709915990467912e-06, "loss": 0.6596, "step": 213 }, { "epoch": 0.061671469740634005, "grad_norm": 2.76483624262101, "learning_rate": 9.718398979629844e-06, "loss": 0.575, "step": 214 }, { "epoch": 0.06195965417867435, "grad_norm": 1.6541639511464132, "learning_rate": 9.726842420913163e-06, "loss": 0.6784, "step": 215 }, { "epoch": 0.0622478386167147, "grad_norm": 2.20119899046267, "learning_rate": 9.735246681353532e-06, "loss": 0.6708, "step": 216 }, { "epoch": 0.06253602305475504, "grad_norm": 3.0557683367774144, "learning_rate": 9.743612122900627e-06, "loss": 0.5411, "step": 217 }, { "epoch": 0.0628242074927954, "grad_norm": 1.8749695111686993, "learning_rate": 9.751939102511684e-06, "loss": 0.6743, "step": 218 }, { "epoch": 0.06311239193083573, "grad_norm": 1.482297281360979, "learning_rate": 9.760227972242893e-06, "loss": 0.664, "step": 219 }, { "epoch": 0.06340057636887608, "grad_norm": 3.422962977642731, "learning_rate": 9.768479079338704e-06, "loss": 0.5599, "step": 220 }, { "epoch": 0.06368876080691642, "grad_norm": 1.9338581326114699, "learning_rate": 9.776692766319116e-06, "loss": 0.6797, "step": 221 }, { "epoch": 0.06397694524495677, "grad_norm": 1.2609633785017869, "learning_rate": 9.78486937106501e-06, "loss": 0.669, "step": 222 }, { "epoch": 0.06426512968299712, "grad_norm": 3.5571118701177826, "learning_rate": 9.793009226901535e-06, "loss": 0.555, "step": 223 }, { "epoch": 0.06455331412103746, "grad_norm": 2.9462476487007074, "learning_rate": 9.801112662679638e-06, "loss": 0.6844, "step": 224 }, { "epoch": 0.06484149855907781, "grad_norm": 1.5889839937538799, "learning_rate": 9.809180002855807e-06, "loss": 0.6593, "step": 225 }, { "epoch": 0.06512968299711816, "grad_norm": 3.7602466428938217, "learning_rate": 9.817211567569991e-06, "loss": 0.5668, "step": 226 }, { "epoch": 0.0654178674351585, "grad_norm": 2.030001995574258, "learning_rate": 9.825207672721862e-06, "loss": 0.6841, "step": 227 }, { "epoch": 0.06570605187319885, "grad_norm": 2.233169312531071, "learning_rate": 9.833168630045344e-06, "loss": 0.6487, "step": 228 }, { "epoch": 0.0659942363112392, "grad_norm": 2.5736297275905424, "learning_rate": 9.841094747181556e-06, "loss": 0.562, "step": 229 }, { "epoch": 0.06628242074927954, "grad_norm": 2.291853135098293, "learning_rate": 9.848986327750132e-06, "loss": 0.6562, "step": 230 }, { "epoch": 0.06657060518731989, "grad_norm": 1.8728261595091804, "learning_rate": 9.85684367141902e-06, "loss": 0.6473, "step": 231 }, { "epoch": 0.06685878962536022, "grad_norm": 2.583515503245737, "learning_rate": 9.864667073972758e-06, "loss": 0.5391, "step": 232 }, { "epoch": 0.06714697406340057, "grad_norm": 1.747241423930901, "learning_rate": 9.872456827379283e-06, "loss": 0.6686, "step": 233 }, { "epoch": 0.06743515850144093, "grad_norm": 1.3701584391406454, "learning_rate": 9.880213219855314e-06, "loss": 0.6519, "step": 234 }, { "epoch": 0.06772334293948126, "grad_norm": 3.4287773141998237, "learning_rate": 9.887936535930344e-06, "loss": 0.5332, "step": 235 }, { "epoch": 0.06801152737752161, "grad_norm": 1.7901734286428725, "learning_rate": 9.895627056509262e-06, "loss": 0.6371, "step": 236 }, { "epoch": 0.06829971181556196, "grad_norm": 1.6397325447306252, "learning_rate": 9.90328505893366e-06, "loss": 0.6518, "step": 237 }, { "epoch": 0.0685878962536023, "grad_norm": 2.745535047975084, "learning_rate": 9.91091081704185e-06, "loss": 0.5285, "step": 238 }, { "epoch": 0.06887608069164265, "grad_norm": 1.3871287337041953, "learning_rate": 9.91850460122761e-06, "loss": 0.6584, "step": 239 }, { "epoch": 0.069164265129683, "grad_norm": 3.6251919092078766, "learning_rate": 9.926066678497726e-06, "loss": 0.6559, "step": 240 }, { "epoch": 0.06945244956772334, "grad_norm": 1.982539491543836, "learning_rate": 9.93359731252832e-06, "loss": 0.5447, "step": 241 }, { "epoch": 0.06974063400576369, "grad_norm": 3.5261780386103623, "learning_rate": 9.941096763720006e-06, "loss": 0.6706, "step": 242 }, { "epoch": 0.07002881844380403, "grad_norm": 1.911397605573411, "learning_rate": 9.948565289251939e-06, "loss": 0.6375, "step": 243 }, { "epoch": 0.07031700288184438, "grad_norm": 2.285007903188301, "learning_rate": 9.956003143134718e-06, "loss": 0.5516, "step": 244 }, { "epoch": 0.07060518731988473, "grad_norm": 1.7565615568409982, "learning_rate": 9.963410576262233e-06, "loss": 0.6452, "step": 245 }, { "epoch": 0.07089337175792507, "grad_norm": 1.9698534688810974, "learning_rate": 9.97078783646244e-06, "loss": 0.6603, "step": 246 }, { "epoch": 0.07118155619596542, "grad_norm": 2.5102089169613193, "learning_rate": 9.978135168547126e-06, "loss": 0.5465, "step": 247 }, { "epoch": 0.07146974063400577, "grad_norm": 1.7303136494447997, "learning_rate": 9.985452814360637e-06, "loss": 0.6296, "step": 248 }, { "epoch": 0.0717579250720461, "grad_norm": 1.11740583936279, "learning_rate": 9.992741012827652e-06, "loss": 0.6626, "step": 249 }, { "epoch": 0.07204610951008646, "grad_norm": 4.068639871871646, "learning_rate": 1e-05, "loss": 0.524, "step": 250 }, { "epoch": 0.07204610951008646, "eval_websight_new_IoU": 0.005624114070087671, "eval_websight_new_MAE_x": 305.29998779296875, "eval_websight_new_MAE_y": 421.6199951171875, "eval_websight_new_coord_probability": 0.717555820941925, "eval_websight_new_correct_token_probability": 0.005016825627535582, "eval_websight_new_inside_bbox": 0.019999999552965164, "eval_websight_new_loss": 0.5373218059539795, "eval_websight_new_range_100_probability": 0.11471053212881088, "eval_websight_new_range_20_probability": 0.027784831821918488, "eval_websight_new_range_5_probability": 0.009580916725099087, "eval_websight_new_runtime": 34.5936, "eval_websight_new_samples_per_second": 1.445, "eval_websight_new_steps_per_second": 0.202, "step": 250 }, { "epoch": 0.07204610951008646, "eval_seeclick_IoU": 0.0044020069763064384, "eval_seeclick_MAE_x": 491.29998779296875, "eval_seeclick_MAE_y": 601.5999755859375, "eval_seeclick_coord_probability": 0.7230032086372375, "eval_seeclick_correct_token_probability": 0.004017556551843882, "eval_seeclick_inside_bbox": 0.05999999865889549, "eval_seeclick_loss": 0.5848657488822937, "eval_seeclick_range_100_probability": 0.12672168016433716, "eval_seeclick_range_20_probability": 0.03310141712427139, "eval_seeclick_range_5_probability": 0.009075821377336979, "eval_seeclick_runtime": 42.2516, "eval_seeclick_samples_per_second": 1.183, "eval_seeclick_steps_per_second": 0.166, "step": 250 }, { "epoch": 0.07204610951008646, "eval_icons_IoU": 0.0, "eval_icons_MAE_x": 620.3400268554688, "eval_icons_MAE_y": 428.2200012207031, "eval_icons_coord_probability": 0.7095824480056763, "eval_icons_correct_token_probability": 0.002493268111720681, "eval_icons_inside_bbox": 0.0, "eval_icons_loss": 0.6193642616271973, "eval_icons_range_100_probability": 0.22341783344745636, "eval_icons_range_20_probability": 0.06995800137519836, "eval_icons_range_5_probability": 0.018928950652480125, "eval_icons_runtime": 43.3165, "eval_icons_samples_per_second": 1.154, "eval_icons_steps_per_second": 0.162, "step": 250 }, { "epoch": 0.07233429394812681, "grad_norm": 1.4228089719555008, "learning_rate": 1e-05, "loss": 0.6483, "step": 251 }, { "epoch": 0.07262247838616714, "grad_norm": 1.4726753507806032, "learning_rate": 1e-05, "loss": 0.6351, "step": 252 }, { "epoch": 0.0729106628242075, "grad_norm": 2.5730722722107267, "learning_rate": 1e-05, "loss": 0.5223, "step": 253 }, { "epoch": 0.07319884726224783, "grad_norm": 1.6668812999341671, "learning_rate": 1e-05, "loss": 0.6679, "step": 254 }, { "epoch": 0.07348703170028818, "grad_norm": 1.2513607195413772, "learning_rate": 1e-05, "loss": 0.6325, "step": 255 }, { "epoch": 0.07377521613832853, "grad_norm": 4.071539500131123, "learning_rate": 1e-05, "loss": 0.5157, "step": 256 }, { "epoch": 0.07406340057636887, "grad_norm": 2.7672062492774874, "learning_rate": 1e-05, "loss": 0.6494, "step": 257 }, { "epoch": 0.07435158501440922, "grad_norm": 1.5093154739351127, "learning_rate": 1e-05, "loss": 0.6371, "step": 258 }, { "epoch": 0.07463976945244957, "grad_norm": 2.6997085498081073, "learning_rate": 1e-05, "loss": 0.5089, "step": 259 }, { "epoch": 0.07492795389048991, "grad_norm": 2.533984042380557, "learning_rate": 1e-05, "loss": 0.6263, "step": 260 }, { "epoch": 0.07521613832853026, "grad_norm": 1.3204406580989467, "learning_rate": 1e-05, "loss": 0.6447, "step": 261 }, { "epoch": 0.07550432276657061, "grad_norm": 2.0059628945560593, "learning_rate": 1e-05, "loss": 0.5287, "step": 262 }, { "epoch": 0.07579250720461095, "grad_norm": 2.055654119489871, "learning_rate": 1e-05, "loss": 0.6279, "step": 263 }, { "epoch": 0.0760806916426513, "grad_norm": 1.6448620459151915, "learning_rate": 1e-05, "loss": 0.6352, "step": 264 }, { "epoch": 0.07636887608069164, "grad_norm": 2.325043774953538, "learning_rate": 1e-05, "loss": 0.5142, "step": 265 }, { "epoch": 0.07665706051873199, "grad_norm": 1.5225370910236076, "learning_rate": 1e-05, "loss": 0.6399, "step": 266 }, { "epoch": 0.07694524495677234, "grad_norm": 1.8189992608714503, "learning_rate": 1e-05, "loss": 0.6533, "step": 267 }, { "epoch": 0.07723342939481267, "grad_norm": 2.1680286049966604, "learning_rate": 1e-05, "loss": 0.5007, "step": 268 }, { "epoch": 0.07752161383285303, "grad_norm": 3.9033303562018555, "learning_rate": 1e-05, "loss": 0.6512, "step": 269 }, { "epoch": 0.07780979827089338, "grad_norm": 1.0489029684926237, "learning_rate": 1e-05, "loss": 0.6411, "step": 270 }, { "epoch": 0.07809798270893371, "grad_norm": 3.264163568820453, "learning_rate": 1e-05, "loss": 0.5221, "step": 271 }, { "epoch": 0.07838616714697406, "grad_norm": 1.8691632613548, "learning_rate": 1e-05, "loss": 0.6288, "step": 272 }, { "epoch": 0.07867435158501442, "grad_norm": 1.2243714190429924, "learning_rate": 1e-05, "loss": 0.6447, "step": 273 }, { "epoch": 0.07896253602305475, "grad_norm": 2.578411914899318, "learning_rate": 1e-05, "loss": 0.5224, "step": 274 }, { "epoch": 0.0792507204610951, "grad_norm": 1.4950396387903995, "learning_rate": 1e-05, "loss": 0.648, "step": 275 }, { "epoch": 0.07953890489913544, "grad_norm": 1.5020358400986247, "learning_rate": 1e-05, "loss": 0.6502, "step": 276 }, { "epoch": 0.07982708933717579, "grad_norm": 1.9186735356285112, "learning_rate": 1e-05, "loss": 0.5023, "step": 277 }, { "epoch": 0.08011527377521614, "grad_norm": 1.5322143468026472, "learning_rate": 1e-05, "loss": 0.6347, "step": 278 }, { "epoch": 0.08040345821325648, "grad_norm": 0.9374030254644127, "learning_rate": 1e-05, "loss": 0.6476, "step": 279 }, { "epoch": 0.08069164265129683, "grad_norm": 2.453546930013716, "learning_rate": 1e-05, "loss": 0.5115, "step": 280 }, { "epoch": 0.08097982708933718, "grad_norm": 1.525975667252706, "learning_rate": 1e-05, "loss": 0.6151, "step": 281 }, { "epoch": 0.08126801152737752, "grad_norm": 1.0710752976867846, "learning_rate": 1e-05, "loss": 0.6447, "step": 282 }, { "epoch": 0.08155619596541787, "grad_norm": 2.322551664301356, "learning_rate": 1e-05, "loss": 0.495, "step": 283 }, { "epoch": 0.08184438040345822, "grad_norm": 1.7580633577114375, "learning_rate": 1e-05, "loss": 0.6064, "step": 284 }, { "epoch": 0.08213256484149856, "grad_norm": 0.98115460737944, "learning_rate": 1e-05, "loss": 0.6263, "step": 285 }, { "epoch": 0.08242074927953891, "grad_norm": 1.9173191103169804, "learning_rate": 1e-05, "loss": 0.4884, "step": 286 }, { "epoch": 0.08270893371757924, "grad_norm": 2.2318016275660106, "learning_rate": 1e-05, "loss": 0.6324, "step": 287 }, { "epoch": 0.0829971181556196, "grad_norm": 1.1124465074110315, "learning_rate": 1e-05, "loss": 0.6226, "step": 288 }, { "epoch": 0.08328530259365995, "grad_norm": 2.4864735694052196, "learning_rate": 1e-05, "loss": 0.5043, "step": 289 }, { "epoch": 0.08357348703170028, "grad_norm": 1.6794470102067223, "learning_rate": 1e-05, "loss": 0.6325, "step": 290 }, { "epoch": 0.08386167146974063, "grad_norm": 1.4315168581948596, "learning_rate": 1e-05, "loss": 0.6229, "step": 291 }, { "epoch": 0.08414985590778098, "grad_norm": 3.4610927042017146, "learning_rate": 1e-05, "loss": 0.5054, "step": 292 }, { "epoch": 0.08443804034582132, "grad_norm": 2.7295557683706115, "learning_rate": 1e-05, "loss": 0.6311, "step": 293 }, { "epoch": 0.08472622478386167, "grad_norm": 1.0191280989324512, "learning_rate": 1e-05, "loss": 0.6217, "step": 294 }, { "epoch": 0.08501440922190202, "grad_norm": 1.9113003308196765, "learning_rate": 1e-05, "loss": 0.497, "step": 295 }, { "epoch": 0.08530259365994236, "grad_norm": 1.5293635470357692, "learning_rate": 1e-05, "loss": 0.5931, "step": 296 }, { "epoch": 0.08559077809798271, "grad_norm": 1.0296246850281991, "learning_rate": 1e-05, "loss": 0.6092, "step": 297 }, { "epoch": 0.08587896253602305, "grad_norm": 2.4398890133828863, "learning_rate": 1e-05, "loss": 0.5003, "step": 298 }, { "epoch": 0.0861671469740634, "grad_norm": 1.36298863155866, "learning_rate": 1e-05, "loss": 0.6619, "step": 299 }, { "epoch": 0.08645533141210375, "grad_norm": 1.0616564773227006, "learning_rate": 1e-05, "loss": 0.6138, "step": 300 }, { "epoch": 0.08674351585014409, "grad_norm": 1.8635512737499922, "learning_rate": 1e-05, "loss": 0.5046, "step": 301 }, { "epoch": 0.08703170028818444, "grad_norm": 2.196386789113054, "learning_rate": 1e-05, "loss": 0.666, "step": 302 }, { "epoch": 0.08731988472622479, "grad_norm": 1.1206758236967824, "learning_rate": 1e-05, "loss": 0.6356, "step": 303 }, { "epoch": 0.08760806916426513, "grad_norm": 2.2610129339817098, "learning_rate": 1e-05, "loss": 0.4975, "step": 304 }, { "epoch": 0.08789625360230548, "grad_norm": 2.9824324925991728, "learning_rate": 1e-05, "loss": 0.6007, "step": 305 }, { "epoch": 0.08818443804034583, "grad_norm": 1.2488543260255234, "learning_rate": 1e-05, "loss": 0.6167, "step": 306 }, { "epoch": 0.08847262247838616, "grad_norm": 2.425461246914256, "learning_rate": 1e-05, "loss": 0.5042, "step": 307 }, { "epoch": 0.08876080691642652, "grad_norm": 2.538485398849825, "learning_rate": 1e-05, "loss": 0.5771, "step": 308 }, { "epoch": 0.08904899135446685, "grad_norm": 1.4754723688947828, "learning_rate": 1e-05, "loss": 0.6217, "step": 309 }, { "epoch": 0.0893371757925072, "grad_norm": 1.4215133658462726, "learning_rate": 1e-05, "loss": 0.5053, "step": 310 }, { "epoch": 0.08962536023054755, "grad_norm": 4.144241341252387, "learning_rate": 1e-05, "loss": 0.602, "step": 311 }, { "epoch": 0.08991354466858789, "grad_norm": 1.2942442719728213, "learning_rate": 1e-05, "loss": 0.6281, "step": 312 }, { "epoch": 0.09020172910662824, "grad_norm": 3.009460634432624, "learning_rate": 1e-05, "loss": 0.5039, "step": 313 }, { "epoch": 0.0904899135446686, "grad_norm": 1.96286308394569, "learning_rate": 1e-05, "loss": 0.6256, "step": 314 }, { "epoch": 0.09077809798270893, "grad_norm": 1.2472287157708806, "learning_rate": 1e-05, "loss": 0.6124, "step": 315 }, { "epoch": 0.09106628242074928, "grad_norm": 1.9483588310709774, "learning_rate": 1e-05, "loss": 0.4877, "step": 316 }, { "epoch": 0.09135446685878963, "grad_norm": 2.539116120554538, "learning_rate": 1e-05, "loss": 0.6175, "step": 317 }, { "epoch": 0.09164265129682997, "grad_norm": 0.9162670920016374, "learning_rate": 1e-05, "loss": 0.604, "step": 318 }, { "epoch": 0.09193083573487032, "grad_norm": 3.9023447397861357, "learning_rate": 1e-05, "loss": 0.4915, "step": 319 }, { "epoch": 0.09221902017291066, "grad_norm": 3.676239398664723, "learning_rate": 1e-05, "loss": 0.6053, "step": 320 }, { "epoch": 0.09250720461095101, "grad_norm": 1.2454556899715417, "learning_rate": 1e-05, "loss": 0.6014, "step": 321 }, { "epoch": 0.09279538904899136, "grad_norm": 1.2458713787172075, "learning_rate": 1e-05, "loss": 0.4911, "step": 322 }, { "epoch": 0.0930835734870317, "grad_norm": 1.866125511189336, "learning_rate": 1e-05, "loss": 0.5944, "step": 323 }, { "epoch": 0.09337175792507205, "grad_norm": 1.190775702837205, "learning_rate": 1e-05, "loss": 0.6043, "step": 324 }, { "epoch": 0.0936599423631124, "grad_norm": 2.197604572030403, "learning_rate": 1e-05, "loss": 0.4878, "step": 325 }, { "epoch": 0.09394812680115273, "grad_norm": 1.6586113238955515, "learning_rate": 1e-05, "loss": 0.6132, "step": 326 }, { "epoch": 0.09423631123919309, "grad_norm": 1.1165382336552463, "learning_rate": 1e-05, "loss": 0.6293, "step": 327 }, { "epoch": 0.09452449567723344, "grad_norm": 2.970713124920137, "learning_rate": 1e-05, "loss": 0.4946, "step": 328 }, { "epoch": 0.09481268011527377, "grad_norm": 4.388834623628037, "learning_rate": 1e-05, "loss": 0.6313, "step": 329 }, { "epoch": 0.09510086455331412, "grad_norm": 1.1696448667484844, "learning_rate": 1e-05, "loss": 0.6282, "step": 330 }, { "epoch": 0.09538904899135446, "grad_norm": 1.7819396121412316, "learning_rate": 1e-05, "loss": 0.4802, "step": 331 }, { "epoch": 0.09567723342939481, "grad_norm": 1.819061336829697, "learning_rate": 1e-05, "loss": 0.5983, "step": 332 }, { "epoch": 0.09596541786743516, "grad_norm": 0.9982024428129266, "learning_rate": 1e-05, "loss": 0.611, "step": 333 }, { "epoch": 0.0962536023054755, "grad_norm": 2.0853295943930963, "learning_rate": 1e-05, "loss": 0.4803, "step": 334 }, { "epoch": 0.09654178674351585, "grad_norm": 2.2050193425368882, "learning_rate": 1e-05, "loss": 0.596, "step": 335 }, { "epoch": 0.0968299711815562, "grad_norm": 1.106955298911173, "learning_rate": 1e-05, "loss": 0.62, "step": 336 }, { "epoch": 0.09711815561959654, "grad_norm": 1.5018103554984554, "learning_rate": 1e-05, "loss": 0.493, "step": 337 }, { "epoch": 0.09740634005763689, "grad_norm": 3.096782380894111, "learning_rate": 1e-05, "loss": 0.6066, "step": 338 }, { "epoch": 0.09769452449567724, "grad_norm": 1.109033538155625, "learning_rate": 1e-05, "loss": 0.6185, "step": 339 }, { "epoch": 0.09798270893371758, "grad_norm": 2.342402894228767, "learning_rate": 1e-05, "loss": 0.4814, "step": 340 }, { "epoch": 0.09827089337175793, "grad_norm": 2.2503418644206037, "learning_rate": 1e-05, "loss": 0.6239, "step": 341 }, { "epoch": 0.09855907780979827, "grad_norm": 1.46191303710194, "learning_rate": 1e-05, "loss": 0.6265, "step": 342 }, { "epoch": 0.09884726224783862, "grad_norm": 3.1383944100395422, "learning_rate": 1e-05, "loss": 0.4819, "step": 343 }, { "epoch": 0.09913544668587897, "grad_norm": 2.7192549497220133, "learning_rate": 1e-05, "loss": 0.6198, "step": 344 }, { "epoch": 0.0994236311239193, "grad_norm": 1.5945490264235185, "learning_rate": 1e-05, "loss": 0.6076, "step": 345 }, { "epoch": 0.09971181556195965, "grad_norm": 1.4625706633703939, "learning_rate": 1e-05, "loss": 0.4706, "step": 346 }, { "epoch": 0.1, "grad_norm": 2.4528442192556827, "learning_rate": 1e-05, "loss": 0.616, "step": 347 }, { "epoch": 0.10028818443804034, "grad_norm": 1.0813691773744674, "learning_rate": 1e-05, "loss": 0.5994, "step": 348 }, { "epoch": 0.1005763688760807, "grad_norm": 3.8456540718069, "learning_rate": 1e-05, "loss": 0.4725, "step": 349 }, { "epoch": 0.10086455331412104, "grad_norm": 2.4854019549048316, "learning_rate": 1e-05, "loss": 0.5904, "step": 350 }, { "epoch": 0.10115273775216138, "grad_norm": 1.0716853325250935, "learning_rate": 1e-05, "loss": 0.6085, "step": 351 }, { "epoch": 0.10144092219020173, "grad_norm": 1.922098976464248, "learning_rate": 1e-05, "loss": 0.4914, "step": 352 }, { "epoch": 0.10172910662824207, "grad_norm": 3.2765105360887437, "learning_rate": 1e-05, "loss": 0.5768, "step": 353 }, { "epoch": 0.10201729106628242, "grad_norm": 1.9159363681527248, "learning_rate": 1e-05, "loss": 0.603, "step": 354 }, { "epoch": 0.10230547550432277, "grad_norm": 1.2291699169535677, "learning_rate": 1e-05, "loss": 0.4772, "step": 355 }, { "epoch": 0.10259365994236311, "grad_norm": 1.7153489314835757, "learning_rate": 1e-05, "loss": 0.5885, "step": 356 }, { "epoch": 0.10288184438040346, "grad_norm": 1.038416686137203, "learning_rate": 1e-05, "loss": 0.6023, "step": 357 }, { "epoch": 0.10317002881844381, "grad_norm": 2.3964872434191378, "learning_rate": 1e-05, "loss": 0.4559, "step": 358 }, { "epoch": 0.10345821325648415, "grad_norm": 2.2157046001432477, "learning_rate": 1e-05, "loss": 0.6163, "step": 359 }, { "epoch": 0.1037463976945245, "grad_norm": 1.6549497014322723, "learning_rate": 1e-05, "loss": 0.6095, "step": 360 }, { "epoch": 0.10403458213256485, "grad_norm": 2.109594268757546, "learning_rate": 1e-05, "loss": 0.4789, "step": 361 }, { "epoch": 0.10432276657060519, "grad_norm": 2.33695262239474, "learning_rate": 1e-05, "loss": 0.5936, "step": 362 }, { "epoch": 0.10461095100864554, "grad_norm": 1.0480109993898619, "learning_rate": 1e-05, "loss": 0.6026, "step": 363 }, { "epoch": 0.10489913544668587, "grad_norm": 1.7960215966295159, "learning_rate": 1e-05, "loss": 0.4674, "step": 364 }, { "epoch": 0.10518731988472622, "grad_norm": 1.568070753765201, "learning_rate": 1e-05, "loss": 0.5882, "step": 365 }, { "epoch": 0.10547550432276658, "grad_norm": 0.9702989899863376, "learning_rate": 1e-05, "loss": 0.5989, "step": 366 }, { "epoch": 0.10576368876080691, "grad_norm": 2.0281939963414506, "learning_rate": 1e-05, "loss": 0.4724, "step": 367 }, { "epoch": 0.10605187319884726, "grad_norm": 2.838118600581876, "learning_rate": 1e-05, "loss": 0.6056, "step": 368 }, { "epoch": 0.10634005763688761, "grad_norm": 1.2565327880090889, "learning_rate": 1e-05, "loss": 0.6003, "step": 369 }, { "epoch": 0.10662824207492795, "grad_norm": 7.053305951324435, "learning_rate": 1e-05, "loss": 0.4619, "step": 370 }, { "epoch": 0.1069164265129683, "grad_norm": 2.979895739358605, "learning_rate": 1e-05, "loss": 0.6202, "step": 371 }, { "epoch": 0.10720461095100864, "grad_norm": 1.0289485571866739, "learning_rate": 1e-05, "loss": 0.6135, "step": 372 }, { "epoch": 0.10749279538904899, "grad_norm": 2.219135041670798, "learning_rate": 1e-05, "loss": 0.4634, "step": 373 }, { "epoch": 0.10778097982708934, "grad_norm": 3.53549590626335, "learning_rate": 1e-05, "loss": 0.611, "step": 374 }, { "epoch": 0.10806916426512968, "grad_norm": 1.3712409115660724, "learning_rate": 1e-05, "loss": 0.6088, "step": 375 }, { "epoch": 0.10835734870317003, "grad_norm": 3.2608083851225844, "learning_rate": 1e-05, "loss": 0.4678, "step": 376 }, { "epoch": 0.10864553314121038, "grad_norm": 2.533467816629062, "learning_rate": 1e-05, "loss": 0.5746, "step": 377 }, { "epoch": 0.10893371757925072, "grad_norm": 1.089968662571353, "learning_rate": 1e-05, "loss": 0.6114, "step": 378 }, { "epoch": 0.10922190201729107, "grad_norm": 2.334185467609639, "learning_rate": 1e-05, "loss": 0.4675, "step": 379 }, { "epoch": 0.10951008645533142, "grad_norm": 4.320804894689986, "learning_rate": 1e-05, "loss": 0.5976, "step": 380 }, { "epoch": 0.10979827089337175, "grad_norm": 1.5020827021639047, "learning_rate": 1e-05, "loss": 0.5988, "step": 381 }, { "epoch": 0.1100864553314121, "grad_norm": 2.9204916922627935, "learning_rate": 1e-05, "loss": 0.4747, "step": 382 }, { "epoch": 0.11037463976945244, "grad_norm": 1.937969689473588, "learning_rate": 1e-05, "loss": 0.5801, "step": 383 }, { "epoch": 0.1106628242074928, "grad_norm": 1.146181403848022, "learning_rate": 1e-05, "loss": 0.61, "step": 384 }, { "epoch": 0.11095100864553314, "grad_norm": 2.5044582375638473, "learning_rate": 1e-05, "loss": 0.4579, "step": 385 }, { "epoch": 0.11123919308357348, "grad_norm": 2.5779323013366646, "learning_rate": 1e-05, "loss": 0.58, "step": 386 }, { "epoch": 0.11152737752161383, "grad_norm": 1.5585121429585433, "learning_rate": 1e-05, "loss": 0.5955, "step": 387 }, { "epoch": 0.11181556195965418, "grad_norm": 3.0977720667830804, "learning_rate": 1e-05, "loss": 0.4567, "step": 388 }, { "epoch": 0.11210374639769452, "grad_norm": 2.155885125341923, "learning_rate": 1e-05, "loss": 0.5899, "step": 389 }, { "epoch": 0.11239193083573487, "grad_norm": 1.1642592175270015, "learning_rate": 1e-05, "loss": 0.6091, "step": 390 }, { "epoch": 0.11268011527377522, "grad_norm": 1.4538116292580978, "learning_rate": 1e-05, "loss": 0.4796, "step": 391 }, { "epoch": 0.11296829971181556, "grad_norm": 1.4638013421860472, "learning_rate": 1e-05, "loss": 0.6164, "step": 392 }, { "epoch": 0.11325648414985591, "grad_norm": 1.0857945976367118, "learning_rate": 1e-05, "loss": 0.5894, "step": 393 }, { "epoch": 0.11354466858789625, "grad_norm": 3.472121988951218, "learning_rate": 1e-05, "loss": 0.4398, "step": 394 }, { "epoch": 0.1138328530259366, "grad_norm": 2.389754760861759, "learning_rate": 1e-05, "loss": 0.5925, "step": 395 }, { "epoch": 0.11412103746397695, "grad_norm": 1.5297289713983515, "learning_rate": 1e-05, "loss": 0.6061, "step": 396 }, { "epoch": 0.11440922190201729, "grad_norm": 2.1975425756434346, "learning_rate": 1e-05, "loss": 0.4559, "step": 397 }, { "epoch": 0.11469740634005764, "grad_norm": 1.8118169367788128, "learning_rate": 1e-05, "loss": 0.6079, "step": 398 }, { "epoch": 0.11498559077809799, "grad_norm": 1.2899700312487847, "learning_rate": 1e-05, "loss": 0.6056, "step": 399 }, { "epoch": 0.11527377521613832, "grad_norm": 2.878458234124749, "learning_rate": 1e-05, "loss": 0.4528, "step": 400 }, { "epoch": 0.11556195965417868, "grad_norm": 3.2408021931829905, "learning_rate": 1e-05, "loss": 0.5949, "step": 401 }, { "epoch": 0.11585014409221903, "grad_norm": 1.1833224811328793, "learning_rate": 1e-05, "loss": 0.5915, "step": 402 }, { "epoch": 0.11613832853025936, "grad_norm": 3.6325309431722776, "learning_rate": 1e-05, "loss": 0.4572, "step": 403 }, { "epoch": 0.11642651296829971, "grad_norm": 3.063499009766476, "learning_rate": 1e-05, "loss": 0.619, "step": 404 }, { "epoch": 0.11671469740634005, "grad_norm": 1.2112833482431893, "learning_rate": 1e-05, "loss": 0.5913, "step": 405 }, { "epoch": 0.1170028818443804, "grad_norm": 1.896771682236592, "learning_rate": 1e-05, "loss": 0.4453, "step": 406 }, { "epoch": 0.11729106628242075, "grad_norm": 2.4464259596256586, "learning_rate": 1e-05, "loss": 0.6023, "step": 407 }, { "epoch": 0.11757925072046109, "grad_norm": 1.0719402348113563, "learning_rate": 1e-05, "loss": 0.5975, "step": 408 }, { "epoch": 0.11786743515850144, "grad_norm": 1.6364604305431, "learning_rate": 1e-05, "loss": 0.4546, "step": 409 }, { "epoch": 0.11815561959654179, "grad_norm": 1.9226298920457043, "learning_rate": 1e-05, "loss": 0.5961, "step": 410 }, { "epoch": 0.11844380403458213, "grad_norm": 1.8746477254601188, "learning_rate": 1e-05, "loss": 0.6082, "step": 411 }, { "epoch": 0.11873198847262248, "grad_norm": 2.9894133572547164, "learning_rate": 1e-05, "loss": 0.4684, "step": 412 }, { "epoch": 0.11902017291066283, "grad_norm": 2.424736405906688, "learning_rate": 1e-05, "loss": 0.5914, "step": 413 }, { "epoch": 0.11930835734870317, "grad_norm": 1.0130518532973238, "learning_rate": 1e-05, "loss": 0.597, "step": 414 }, { "epoch": 0.11959654178674352, "grad_norm": 2.5468271019669584, "learning_rate": 1e-05, "loss": 0.4359, "step": 415 }, { "epoch": 0.11988472622478386, "grad_norm": 3.017222974339237, "learning_rate": 1e-05, "loss": 0.5772, "step": 416 }, { "epoch": 0.1201729106628242, "grad_norm": 1.1193530816512283, "learning_rate": 1e-05, "loss": 0.6038, "step": 417 }, { "epoch": 0.12046109510086456, "grad_norm": 2.762433222336846, "learning_rate": 1e-05, "loss": 0.4539, "step": 418 }, { "epoch": 0.1207492795389049, "grad_norm": 2.6483256028210294, "learning_rate": 1e-05, "loss": 0.5616, "step": 419 }, { "epoch": 0.12103746397694524, "grad_norm": 1.2060796510223488, "learning_rate": 1e-05, "loss": 0.5961, "step": 420 }, { "epoch": 0.1213256484149856, "grad_norm": 2.6985162932794675, "learning_rate": 1e-05, "loss": 0.4467, "step": 421 }, { "epoch": 0.12161383285302593, "grad_norm": 2.379383849739707, "learning_rate": 1e-05, "loss": 0.5801, "step": 422 }, { "epoch": 0.12190201729106628, "grad_norm": 1.652316386479186, "learning_rate": 1e-05, "loss": 0.5923, "step": 423 }, { "epoch": 0.12219020172910663, "grad_norm": 2.679015156162535, "learning_rate": 1e-05, "loss": 0.4607, "step": 424 }, { "epoch": 0.12247838616714697, "grad_norm": 2.786695020448989, "learning_rate": 1e-05, "loss": 0.5573, "step": 425 }, { "epoch": 0.12276657060518732, "grad_norm": 1.2997492162885074, "learning_rate": 1e-05, "loss": 0.5965, "step": 426 }, { "epoch": 0.12305475504322766, "grad_norm": 2.6794316898764845, "learning_rate": 1e-05, "loss": 0.4543, "step": 427 }, { "epoch": 0.12334293948126801, "grad_norm": 2.243329329192076, "learning_rate": 1e-05, "loss": 0.5881, "step": 428 }, { "epoch": 0.12363112391930836, "grad_norm": 1.3668366873798132, "learning_rate": 1e-05, "loss": 0.5906, "step": 429 }, { "epoch": 0.1239193083573487, "grad_norm": 2.8724616352195658, "learning_rate": 1e-05, "loss": 0.4461, "step": 430 }, { "epoch": 0.12420749279538905, "grad_norm": 3.3669598621394816, "learning_rate": 1e-05, "loss": 0.5512, "step": 431 }, { "epoch": 0.1244956772334294, "grad_norm": 1.7398174447091903, "learning_rate": 1e-05, "loss": 0.5987, "step": 432 }, { "epoch": 0.12478386167146974, "grad_norm": 2.125126459710019, "learning_rate": 1e-05, "loss": 0.4399, "step": 433 }, { "epoch": 0.1250720461095101, "grad_norm": 1.9668441023227823, "learning_rate": 1e-05, "loss": 0.5572, "step": 434 }, { "epoch": 0.12536023054755044, "grad_norm": 1.4135472721759272, "learning_rate": 1e-05, "loss": 0.5917, "step": 435 }, { "epoch": 0.1256484149855908, "grad_norm": 2.6557482752537354, "learning_rate": 1e-05, "loss": 0.4216, "step": 436 }, { "epoch": 0.1259365994236311, "grad_norm": 1.9071029220151672, "learning_rate": 1e-05, "loss": 0.5837, "step": 437 }, { "epoch": 0.12622478386167146, "grad_norm": 1.5418422732583796, "learning_rate": 1e-05, "loss": 0.6133, "step": 438 }, { "epoch": 0.12651296829971181, "grad_norm": 3.490914074598071, "learning_rate": 1e-05, "loss": 0.43, "step": 439 }, { "epoch": 0.12680115273775217, "grad_norm": 2.6635869698716643, "learning_rate": 1e-05, "loss": 0.5967, "step": 440 }, { "epoch": 0.12708933717579252, "grad_norm": 1.0604435551098819, "learning_rate": 1e-05, "loss": 0.5985, "step": 441 }, { "epoch": 0.12737752161383284, "grad_norm": 3.469986231723641, "learning_rate": 1e-05, "loss": 0.451, "step": 442 }, { "epoch": 0.1276657060518732, "grad_norm": 2.833020290341045, "learning_rate": 1e-05, "loss": 0.5725, "step": 443 }, { "epoch": 0.12795389048991354, "grad_norm": 1.0051541036464473, "learning_rate": 1e-05, "loss": 0.5809, "step": 444 }, { "epoch": 0.1282420749279539, "grad_norm": 1.8142770422576717, "learning_rate": 1e-05, "loss": 0.4529, "step": 445 }, { "epoch": 0.12853025936599424, "grad_norm": 3.9811298650754323, "learning_rate": 1e-05, "loss": 0.5488, "step": 446 }, { "epoch": 0.1288184438040346, "grad_norm": 1.218327361702739, "learning_rate": 1e-05, "loss": 0.5773, "step": 447 }, { "epoch": 0.12910662824207492, "grad_norm": 2.201286370240042, "learning_rate": 1e-05, "loss": 0.4534, "step": 448 }, { "epoch": 0.12939481268011527, "grad_norm": 1.7478083257483388, "learning_rate": 1e-05, "loss": 0.6036, "step": 449 }, { "epoch": 0.12968299711815562, "grad_norm": 1.427855547613876, "learning_rate": 1e-05, "loss": 0.5867, "step": 450 }, { "epoch": 0.12997118155619597, "grad_norm": 1.4621987146776787, "learning_rate": 1e-05, "loss": 0.4413, "step": 451 }, { "epoch": 0.13025936599423632, "grad_norm": 1.4666772098953285, "learning_rate": 1e-05, "loss": 0.5748, "step": 452 }, { "epoch": 0.13054755043227664, "grad_norm": 1.213595026215865, "learning_rate": 1e-05, "loss": 0.5982, "step": 453 }, { "epoch": 0.130835734870317, "grad_norm": 2.355536720115263, "learning_rate": 1e-05, "loss": 0.445, "step": 454 }, { "epoch": 0.13112391930835735, "grad_norm": 2.920619249973288, "learning_rate": 1e-05, "loss": 0.5817, "step": 455 }, { "epoch": 0.1314121037463977, "grad_norm": 1.3428697439989274, "learning_rate": 1e-05, "loss": 0.5694, "step": 456 }, { "epoch": 0.13170028818443805, "grad_norm": 1.743289386727595, "learning_rate": 1e-05, "loss": 0.4312, "step": 457 }, { "epoch": 0.1319884726224784, "grad_norm": 2.255421712493541, "learning_rate": 1e-05, "loss": 0.5958, "step": 458 }, { "epoch": 0.13227665706051872, "grad_norm": 1.1463419678293008, "learning_rate": 1e-05, "loss": 0.5809, "step": 459 }, { "epoch": 0.13256484149855907, "grad_norm": 2.746233739943775, "learning_rate": 1e-05, "loss": 0.4354, "step": 460 }, { "epoch": 0.13285302593659942, "grad_norm": 2.4487357894621224, "learning_rate": 1e-05, "loss": 0.5515, "step": 461 }, { "epoch": 0.13314121037463977, "grad_norm": 1.735262829126143, "learning_rate": 1e-05, "loss": 0.5891, "step": 462 }, { "epoch": 0.13342939481268012, "grad_norm": 3.6653451701990347, "learning_rate": 1e-05, "loss": 0.4155, "step": 463 }, { "epoch": 0.13371757925072045, "grad_norm": 3.039932302637225, "learning_rate": 1e-05, "loss": 0.5491, "step": 464 }, { "epoch": 0.1340057636887608, "grad_norm": 1.1770276022521715, "learning_rate": 1e-05, "loss": 0.581, "step": 465 }, { "epoch": 0.13429394812680115, "grad_norm": 4.167641032839015, "learning_rate": 1e-05, "loss": 0.442, "step": 466 }, { "epoch": 0.1345821325648415, "grad_norm": 1.6053065509968596, "learning_rate": 1e-05, "loss": 0.5822, "step": 467 }, { "epoch": 0.13487031700288185, "grad_norm": 1.5032997619338024, "learning_rate": 1e-05, "loss": 0.5772, "step": 468 }, { "epoch": 0.1351585014409222, "grad_norm": 3.270618757600374, "learning_rate": 1e-05, "loss": 0.4348, "step": 469 }, { "epoch": 0.13544668587896252, "grad_norm": 2.791741501570059, "learning_rate": 1e-05, "loss": 0.5844, "step": 470 }, { "epoch": 0.13573487031700288, "grad_norm": 1.1770255237309266, "learning_rate": 1e-05, "loss": 0.574, "step": 471 }, { "epoch": 0.13602305475504323, "grad_norm": 4.736775435466847, "learning_rate": 1e-05, "loss": 0.4244, "step": 472 }, { "epoch": 0.13631123919308358, "grad_norm": 1.7837329171849312, "learning_rate": 1e-05, "loss": 0.593, "step": 473 }, { "epoch": 0.13659942363112393, "grad_norm": 1.149430666649388, "learning_rate": 1e-05, "loss": 0.5877, "step": 474 }, { "epoch": 0.13688760806916425, "grad_norm": 2.318623131551547, "learning_rate": 1e-05, "loss": 0.4416, "step": 475 }, { "epoch": 0.1371757925072046, "grad_norm": 2.8828159601083247, "learning_rate": 1e-05, "loss": 0.5403, "step": 476 }, { "epoch": 0.13746397694524495, "grad_norm": 1.723279748983139, "learning_rate": 1e-05, "loss": 0.5806, "step": 477 }, { "epoch": 0.1377521613832853, "grad_norm": 1.7523594088222687, "learning_rate": 1e-05, "loss": 0.4269, "step": 478 }, { "epoch": 0.13804034582132566, "grad_norm": 4.094525500917702, "learning_rate": 1e-05, "loss": 0.5912, "step": 479 }, { "epoch": 0.138328530259366, "grad_norm": 1.219868256258515, "learning_rate": 1e-05, "loss": 0.5927, "step": 480 }, { "epoch": 0.13861671469740633, "grad_norm": 3.8503024267859858, "learning_rate": 1e-05, "loss": 0.4374, "step": 481 }, { "epoch": 0.13890489913544668, "grad_norm": 3.7215917656004915, "learning_rate": 1e-05, "loss": 0.57, "step": 482 }, { "epoch": 0.13919308357348703, "grad_norm": 1.326839803262488, "learning_rate": 1e-05, "loss": 0.5719, "step": 483 }, { "epoch": 0.13948126801152738, "grad_norm": 3.1101448521293342, "learning_rate": 1e-05, "loss": 0.4192, "step": 484 }, { "epoch": 0.13976945244956773, "grad_norm": 1.5713028450262407, "learning_rate": 1e-05, "loss": 0.5521, "step": 485 }, { "epoch": 0.14005763688760806, "grad_norm": 1.4812387301104615, "learning_rate": 1e-05, "loss": 0.5961, "step": 486 }, { "epoch": 0.1403458213256484, "grad_norm": 3.112719692709699, "learning_rate": 1e-05, "loss": 0.425, "step": 487 }, { "epoch": 0.14063400576368876, "grad_norm": 3.117300102115347, "learning_rate": 1e-05, "loss": 0.5913, "step": 488 }, { "epoch": 0.1409221902017291, "grad_norm": 1.281642593906318, "learning_rate": 1e-05, "loss": 0.568, "step": 489 }, { "epoch": 0.14121037463976946, "grad_norm": 3.130491230859037, "learning_rate": 1e-05, "loss": 0.4269, "step": 490 }, { "epoch": 0.1414985590778098, "grad_norm": 2.4765855622656496, "learning_rate": 1e-05, "loss": 0.5728, "step": 491 }, { "epoch": 0.14178674351585013, "grad_norm": 1.3995884984100582, "learning_rate": 1e-05, "loss": 0.5658, "step": 492 }, { "epoch": 0.14207492795389048, "grad_norm": 1.739429378644102, "learning_rate": 1e-05, "loss": 0.4273, "step": 493 }, { "epoch": 0.14236311239193083, "grad_norm": 2.0295421203874318, "learning_rate": 1e-05, "loss": 0.5422, "step": 494 }, { "epoch": 0.14265129682997119, "grad_norm": 1.3170660825967557, "learning_rate": 1e-05, "loss": 0.5949, "step": 495 }, { "epoch": 0.14293948126801154, "grad_norm": 2.6091612579691197, "learning_rate": 1e-05, "loss": 0.4109, "step": 496 }, { "epoch": 0.14322766570605186, "grad_norm": 3.6534911957875242, "learning_rate": 1e-05, "loss": 0.5765, "step": 497 }, { "epoch": 0.1435158501440922, "grad_norm": 1.2762671692669125, "learning_rate": 1e-05, "loss": 0.5757, "step": 498 }, { "epoch": 0.14380403458213256, "grad_norm": 2.0657668399019644, "learning_rate": 1e-05, "loss": 0.4346, "step": 499 }, { "epoch": 0.1440922190201729, "grad_norm": 1.914852314151166, "learning_rate": 1e-05, "loss": 0.57, "step": 500 }, { "epoch": 0.1440922190201729, "eval_websight_new_IoU": 0.01463321316987276, "eval_websight_new_MAE_x": 286.7799987792969, "eval_websight_new_MAE_y": 421.5400085449219, "eval_websight_new_coord_probability": 0.73215651512146, "eval_websight_new_correct_token_probability": 0.0021086945198476315, "eval_websight_new_inside_bbox": 0.019999999552965164, "eval_websight_new_loss": 0.4361584484577179, "eval_websight_new_range_100_probability": 0.0867907926440239, "eval_websight_new_range_20_probability": 0.017009830102324486, "eval_websight_new_range_5_probability": 0.0048213982954621315, "eval_websight_new_runtime": 34.1634, "eval_websight_new_samples_per_second": 1.464, "eval_websight_new_steps_per_second": 0.205, "step": 500 }, { "epoch": 0.1440922190201729, "eval_seeclick_IoU": 0.03221755847334862, "eval_seeclick_MAE_x": 473.1000061035156, "eval_seeclick_MAE_y": 601.5999755859375, "eval_seeclick_coord_probability": 0.7388027310371399, "eval_seeclick_correct_token_probability": 0.002116676652804017, "eval_seeclick_inside_bbox": 0.05999999865889549, "eval_seeclick_loss": 0.4738258123397827, "eval_seeclick_range_100_probability": 0.1094726100564003, "eval_seeclick_range_20_probability": 0.029427895322442055, "eval_seeclick_range_5_probability": 0.005799091421067715, "eval_seeclick_runtime": 42.7801, "eval_seeclick_samples_per_second": 1.169, "eval_seeclick_steps_per_second": 0.164, "step": 500 }, { "epoch": 0.1440922190201729, "eval_icons_IoU": 0.0019313165685161948, "eval_icons_MAE_x": 473.5400085449219, "eval_icons_MAE_y": 448.8800048828125, "eval_icons_coord_probability": 0.7443975806236267, "eval_icons_correct_token_probability": 0.0036313915625214577, "eval_icons_inside_bbox": 0.0, "eval_icons_loss": 0.5505882501602173, "eval_icons_range_100_probability": 0.2653256356716156, "eval_icons_range_20_probability": 0.09607495367527008, "eval_icons_range_5_probability": 0.02670016698539257, "eval_icons_runtime": 43.0247, "eval_icons_samples_per_second": 1.162, "eval_icons_steps_per_second": 0.163, "step": 500 }, { "epoch": 0.14438040345821326, "grad_norm": 1.8448071513284736, "learning_rate": 1e-05, "loss": 0.579, "step": 501 }, { "epoch": 0.14466858789625361, "grad_norm": 3.1326321130794916, "learning_rate": 1e-05, "loss": 0.4292, "step": 502 }, { "epoch": 0.14495677233429394, "grad_norm": 1.9410665755091052, "learning_rate": 1e-05, "loss": 0.5894, "step": 503 }, { "epoch": 0.1452449567723343, "grad_norm": 1.8431514766601858, "learning_rate": 1e-05, "loss": 0.5626, "step": 504 }, { "epoch": 0.14553314121037464, "grad_norm": 3.548647616359119, "learning_rate": 1e-05, "loss": 0.4507, "step": 505 }, { "epoch": 0.145821325648415, "grad_norm": 2.5418153973369972, "learning_rate": 1e-05, "loss": 0.562, "step": 506 }, { "epoch": 0.14610951008645534, "grad_norm": 1.3964779815637525, "learning_rate": 1e-05, "loss": 0.5818, "step": 507 }, { "epoch": 0.14639769452449566, "grad_norm": 2.5166763904911633, "learning_rate": 1e-05, "loss": 0.4376, "step": 508 }, { "epoch": 0.14668587896253601, "grad_norm": 1.6927625167741949, "learning_rate": 1e-05, "loss": 0.5474, "step": 509 }, { "epoch": 0.14697406340057637, "grad_norm": 1.7542553767621742, "learning_rate": 1e-05, "loss": 0.5643, "step": 510 }, { "epoch": 0.14726224783861672, "grad_norm": 2.9349303403137244, "learning_rate": 1e-05, "loss": 0.4413, "step": 511 }, { "epoch": 0.14755043227665707, "grad_norm": 1.9760710131732777, "learning_rate": 1e-05, "loss": 0.5451, "step": 512 }, { "epoch": 0.14783861671469742, "grad_norm": 1.449825928543996, "learning_rate": 1e-05, "loss": 0.5691, "step": 513 }, { "epoch": 0.14812680115273774, "grad_norm": 4.073458689095103, "learning_rate": 1e-05, "loss": 0.4131, "step": 514 }, { "epoch": 0.1484149855907781, "grad_norm": 2.801815070531123, "learning_rate": 1e-05, "loss": 0.571, "step": 515 }, { "epoch": 0.14870317002881844, "grad_norm": 2.126190537412442, "learning_rate": 1e-05, "loss": 0.5916, "step": 516 }, { "epoch": 0.1489913544668588, "grad_norm": 1.4518893561832873, "learning_rate": 1e-05, "loss": 0.4218, "step": 517 }, { "epoch": 0.14927953890489915, "grad_norm": 3.180599704926899, "learning_rate": 1e-05, "loss": 0.5596, "step": 518 }, { "epoch": 0.14956772334293947, "grad_norm": 1.4565202457275603, "learning_rate": 1e-05, "loss": 0.5746, "step": 519 }, { "epoch": 0.14985590778097982, "grad_norm": 2.1482842856757154, "learning_rate": 1e-05, "loss": 0.4091, "step": 520 }, { "epoch": 0.15014409221902017, "grad_norm": 2.2267048860256082, "learning_rate": 1e-05, "loss": 0.5719, "step": 521 }, { "epoch": 0.15043227665706052, "grad_norm": 1.1936041027002116, "learning_rate": 1e-05, "loss": 0.5787, "step": 522 }, { "epoch": 0.15072046109510087, "grad_norm": 4.109477616663216, "learning_rate": 1e-05, "loss": 0.4387, "step": 523 }, { "epoch": 0.15100864553314122, "grad_norm": 3.145535845039723, "learning_rate": 1e-05, "loss": 0.5522, "step": 524 }, { "epoch": 0.15129682997118155, "grad_norm": 1.174791602931962, "learning_rate": 1e-05, "loss": 0.5596, "step": 525 }, { "epoch": 0.1515850144092219, "grad_norm": 3.2884466993425816, "learning_rate": 1e-05, "loss": 0.4216, "step": 526 }, { "epoch": 0.15187319884726225, "grad_norm": 2.1818494760236895, "learning_rate": 1e-05, "loss": 0.5388, "step": 527 }, { "epoch": 0.1521613832853026, "grad_norm": 1.2537362607320528, "learning_rate": 1e-05, "loss": 0.583, "step": 528 }, { "epoch": 0.15244956772334295, "grad_norm": 1.9383363422951394, "learning_rate": 1e-05, "loss": 0.4226, "step": 529 }, { "epoch": 0.15273775216138327, "grad_norm": 3.4770943698334813, "learning_rate": 1e-05, "loss": 0.5313, "step": 530 }, { "epoch": 0.15302593659942362, "grad_norm": 1.6673674360106558, "learning_rate": 1e-05, "loss": 0.5822, "step": 531 }, { "epoch": 0.15331412103746397, "grad_norm": 1.9256390602904025, "learning_rate": 1e-05, "loss": 0.41, "step": 532 }, { "epoch": 0.15360230547550432, "grad_norm": 3.4372138253608133, "learning_rate": 1e-05, "loss": 0.5564, "step": 533 }, { "epoch": 0.15389048991354468, "grad_norm": 1.139878259611146, "learning_rate": 1e-05, "loss": 0.5731, "step": 534 }, { "epoch": 0.15417867435158503, "grad_norm": 2.5258128668121658, "learning_rate": 1e-05, "loss": 0.4287, "step": 535 }, { "epoch": 0.15446685878962535, "grad_norm": 2.264631946485219, "learning_rate": 1e-05, "loss": 0.5801, "step": 536 }, { "epoch": 0.1547550432276657, "grad_norm": 1.15717597829775, "learning_rate": 1e-05, "loss": 0.5764, "step": 537 }, { "epoch": 0.15504322766570605, "grad_norm": 2.19598195468663, "learning_rate": 1e-05, "loss": 0.4152, "step": 538 }, { "epoch": 0.1553314121037464, "grad_norm": 1.5974769077267827, "learning_rate": 1e-05, "loss": 0.5567, "step": 539 }, { "epoch": 0.15561959654178675, "grad_norm": 1.6439935505821885, "learning_rate": 1e-05, "loss": 0.5763, "step": 540 }, { "epoch": 0.15590778097982708, "grad_norm": 1.8035286132055963, "learning_rate": 1e-05, "loss": 0.4244, "step": 541 }, { "epoch": 0.15619596541786743, "grad_norm": 3.3289981506089608, "learning_rate": 1e-05, "loss": 0.5306, "step": 542 }, { "epoch": 0.15648414985590778, "grad_norm": 1.3575419479537025, "learning_rate": 1e-05, "loss": 0.5522, "step": 543 }, { "epoch": 0.15677233429394813, "grad_norm": 4.060042360386747, "learning_rate": 1e-05, "loss": 0.4354, "step": 544 }, { "epoch": 0.15706051873198848, "grad_norm": 2.719748279737449, "learning_rate": 1e-05, "loss": 0.5546, "step": 545 }, { "epoch": 0.15734870317002883, "grad_norm": 1.1607951338878562, "learning_rate": 1e-05, "loss": 0.5659, "step": 546 }, { "epoch": 0.15763688760806915, "grad_norm": 2.2262089217725882, "learning_rate": 1e-05, "loss": 0.4216, "step": 547 }, { "epoch": 0.1579250720461095, "grad_norm": 2.152306989000815, "learning_rate": 1e-05, "loss": 0.5796, "step": 548 }, { "epoch": 0.15821325648414986, "grad_norm": 1.9929809674573478, "learning_rate": 1e-05, "loss": 0.5694, "step": 549 }, { "epoch": 0.1585014409221902, "grad_norm": 1.6817494253176306, "learning_rate": 1e-05, "loss": 0.4244, "step": 550 }, { "epoch": 0.15878962536023056, "grad_norm": 3.3303932325186083, "learning_rate": 1e-05, "loss": 0.5609, "step": 551 }, { "epoch": 0.15907780979827088, "grad_norm": 1.4879625617138847, "learning_rate": 1e-05, "loss": 0.551, "step": 552 }, { "epoch": 0.15936599423631123, "grad_norm": 1.7347403864057702, "learning_rate": 1e-05, "loss": 0.4, "step": 553 }, { "epoch": 0.15965417867435158, "grad_norm": 1.7652909618145627, "learning_rate": 1e-05, "loss": 0.5797, "step": 554 }, { "epoch": 0.15994236311239193, "grad_norm": 1.7303096867055798, "learning_rate": 1e-05, "loss": 0.5722, "step": 555 }, { "epoch": 0.16023054755043228, "grad_norm": 1.7126273955726241, "learning_rate": 1e-05, "loss": 0.3985, "step": 556 }, { "epoch": 0.16051873198847263, "grad_norm": 1.4902869117595645, "learning_rate": 1e-05, "loss": 0.5386, "step": 557 }, { "epoch": 0.16080691642651296, "grad_norm": 1.2679155140149418, "learning_rate": 1e-05, "loss": 0.5573, "step": 558 }, { "epoch": 0.1610951008645533, "grad_norm": 4.209937191606016, "learning_rate": 1e-05, "loss": 0.386, "step": 559 }, { "epoch": 0.16138328530259366, "grad_norm": 2.1327055619520903, "learning_rate": 1e-05, "loss": 0.5512, "step": 560 }, { "epoch": 0.161671469740634, "grad_norm": 1.8745967111214592, "learning_rate": 1e-05, "loss": 0.5706, "step": 561 }, { "epoch": 0.16195965417867436, "grad_norm": 3.004705979570041, "learning_rate": 1e-05, "loss": 0.3967, "step": 562 }, { "epoch": 0.16224783861671468, "grad_norm": 3.705567418644865, "learning_rate": 1e-05, "loss": 0.5565, "step": 563 }, { "epoch": 0.16253602305475504, "grad_norm": 1.879018023645976, "learning_rate": 1e-05, "loss": 0.5613, "step": 564 }, { "epoch": 0.1628242074927954, "grad_norm": 2.2060710644133685, "learning_rate": 1e-05, "loss": 0.3915, "step": 565 }, { "epoch": 0.16311239193083574, "grad_norm": 2.07947260611661, "learning_rate": 1e-05, "loss": 0.5639, "step": 566 }, { "epoch": 0.1634005763688761, "grad_norm": 1.6378908418091689, "learning_rate": 1e-05, "loss": 0.5699, "step": 567 }, { "epoch": 0.16368876080691644, "grad_norm": 3.000890981624951, "learning_rate": 1e-05, "loss": 0.4058, "step": 568 }, { "epoch": 0.16397694524495676, "grad_norm": 2.126292063396071, "learning_rate": 1e-05, "loss": 0.5236, "step": 569 }, { "epoch": 0.1642651296829971, "grad_norm": 1.5309333888574757, "learning_rate": 1e-05, "loss": 0.5497, "step": 570 }, { "epoch": 0.16455331412103746, "grad_norm": 4.411481950034936, "learning_rate": 1e-05, "loss": 0.4087, "step": 571 }, { "epoch": 0.16484149855907781, "grad_norm": 2.537810253449623, "learning_rate": 1e-05, "loss": 0.502, "step": 572 }, { "epoch": 0.16512968299711817, "grad_norm": 1.3858443640190015, "learning_rate": 1e-05, "loss": 0.5688, "step": 573 }, { "epoch": 0.1654178674351585, "grad_norm": 3.1550677968939924, "learning_rate": 1e-05, "loss": 0.4229, "step": 574 }, { "epoch": 0.16570605187319884, "grad_norm": 1.833600330517365, "learning_rate": 1e-05, "loss": 0.539, "step": 575 }, { "epoch": 0.1659942363112392, "grad_norm": 1.8400406919145424, "learning_rate": 1e-05, "loss": 0.5519, "step": 576 }, { "epoch": 0.16628242074927954, "grad_norm": 2.606139497529331, "learning_rate": 1e-05, "loss": 0.4, "step": 577 }, { "epoch": 0.1665706051873199, "grad_norm": 2.0181636162880308, "learning_rate": 1e-05, "loss": 0.5852, "step": 578 }, { "epoch": 0.16685878962536024, "grad_norm": 1.6839491696790014, "learning_rate": 1e-05, "loss": 0.5687, "step": 579 }, { "epoch": 0.16714697406340057, "grad_norm": 3.854172359890152, "learning_rate": 1e-05, "loss": 0.4043, "step": 580 }, { "epoch": 0.16743515850144092, "grad_norm": 3.0960591563701434, "learning_rate": 1e-05, "loss": 0.5288, "step": 581 }, { "epoch": 0.16772334293948127, "grad_norm": 2.1395720918645265, "learning_rate": 1e-05, "loss": 0.5714, "step": 582 }, { "epoch": 0.16801152737752162, "grad_norm": 3.4674051279756855, "learning_rate": 1e-05, "loss": 0.4016, "step": 583 }, { "epoch": 0.16829971181556197, "grad_norm": 3.0515951310737193, "learning_rate": 1e-05, "loss": 0.5489, "step": 584 }, { "epoch": 0.1685878962536023, "grad_norm": 1.3654445978000875, "learning_rate": 1e-05, "loss": 0.5505, "step": 585 }, { "epoch": 0.16887608069164264, "grad_norm": 2.571044000683727, "learning_rate": 1e-05, "loss": 0.4286, "step": 586 }, { "epoch": 0.169164265129683, "grad_norm": 2.262059543392501, "learning_rate": 1e-05, "loss": 0.5691, "step": 587 }, { "epoch": 0.16945244956772335, "grad_norm": 2.3723813138832313, "learning_rate": 1e-05, "loss": 0.5505, "step": 588 }, { "epoch": 0.1697406340057637, "grad_norm": 2.2783042616446365, "learning_rate": 1e-05, "loss": 0.4146, "step": 589 }, { "epoch": 0.17002881844380405, "grad_norm": 3.3019889343278206, "learning_rate": 1e-05, "loss": 0.5751, "step": 590 }, { "epoch": 0.17031700288184437, "grad_norm": 1.4577549924848103, "learning_rate": 1e-05, "loss": 0.5515, "step": 591 }, { "epoch": 0.17060518731988472, "grad_norm": 4.535404100747485, "learning_rate": 1e-05, "loss": 0.3968, "step": 592 }, { "epoch": 0.17089337175792507, "grad_norm": 2.007603274675207, "learning_rate": 1e-05, "loss": 0.5427, "step": 593 }, { "epoch": 0.17118155619596542, "grad_norm": 1.38832578741342, "learning_rate": 1e-05, "loss": 0.558, "step": 594 }, { "epoch": 0.17146974063400577, "grad_norm": 2.5680251846041497, "learning_rate": 1e-05, "loss": 0.3968, "step": 595 }, { "epoch": 0.1717579250720461, "grad_norm": 3.6595150087684094, "learning_rate": 1e-05, "loss": 0.5459, "step": 596 }, { "epoch": 0.17204610951008645, "grad_norm": 2.2315423996352934, "learning_rate": 1e-05, "loss": 0.5505, "step": 597 }, { "epoch": 0.1723342939481268, "grad_norm": 2.401220339770531, "learning_rate": 1e-05, "loss": 0.3946, "step": 598 }, { "epoch": 0.17262247838616715, "grad_norm": 2.1875816223005824, "learning_rate": 1e-05, "loss": 0.5647, "step": 599 }, { "epoch": 0.1729106628242075, "grad_norm": 1.973156691171414, "learning_rate": 1e-05, "loss": 0.5456, "step": 600 }, { "epoch": 0.17319884726224785, "grad_norm": 3.448549394643256, "learning_rate": 1e-05, "loss": 0.396, "step": 601 }, { "epoch": 0.17348703170028817, "grad_norm": 2.295210141260302, "learning_rate": 1e-05, "loss": 0.5485, "step": 602 }, { "epoch": 0.17377521613832853, "grad_norm": 1.9451387743663426, "learning_rate": 1e-05, "loss": 0.5646, "step": 603 }, { "epoch": 0.17406340057636888, "grad_norm": 3.865943774754975, "learning_rate": 1e-05, "loss": 0.3931, "step": 604 }, { "epoch": 0.17435158501440923, "grad_norm": 2.878396544005281, "learning_rate": 1e-05, "loss": 0.5514, "step": 605 }, { "epoch": 0.17463976945244958, "grad_norm": 1.4862887491338082, "learning_rate": 1e-05, "loss": 0.5606, "step": 606 }, { "epoch": 0.1749279538904899, "grad_norm": 1.886792204755533, "learning_rate": 1e-05, "loss": 0.3835, "step": 607 }, { "epoch": 0.17521613832853025, "grad_norm": 2.93011855036884, "learning_rate": 1e-05, "loss": 0.548, "step": 608 }, { "epoch": 0.1755043227665706, "grad_norm": 2.327293328559529, "learning_rate": 1e-05, "loss": 0.5625, "step": 609 }, { "epoch": 0.17579250720461095, "grad_norm": 3.8373218363052906, "learning_rate": 1e-05, "loss": 0.3928, "step": 610 }, { "epoch": 0.1760806916426513, "grad_norm": 2.4655793860953956, "learning_rate": 1e-05, "loss": 0.5778, "step": 611 }, { "epoch": 0.17636887608069166, "grad_norm": 1.849299129303253, "learning_rate": 1e-05, "loss": 0.5402, "step": 612 }, { "epoch": 0.17665706051873198, "grad_norm": 5.654202874511459, "learning_rate": 1e-05, "loss": 0.3951, "step": 613 }, { "epoch": 0.17694524495677233, "grad_norm": 4.320339997345097, "learning_rate": 1e-05, "loss": 0.5332, "step": 614 }, { "epoch": 0.17723342939481268, "grad_norm": 1.9081807363214465, "learning_rate": 1e-05, "loss": 0.5533, "step": 615 }, { "epoch": 0.17752161383285303, "grad_norm": 1.6998027358164398, "learning_rate": 1e-05, "loss": 0.4038, "step": 616 }, { "epoch": 0.17780979827089338, "grad_norm": 1.8637243057421535, "learning_rate": 1e-05, "loss": 0.5201, "step": 617 }, { "epoch": 0.1780979827089337, "grad_norm": 2.3678771540872177, "learning_rate": 1e-05, "loss": 0.5529, "step": 618 }, { "epoch": 0.17838616714697406, "grad_norm": 5.57593000831114, "learning_rate": 1e-05, "loss": 0.4073, "step": 619 }, { "epoch": 0.1786743515850144, "grad_norm": 4.218960810252173, "learning_rate": 1e-05, "loss": 0.5473, "step": 620 }, { "epoch": 0.17896253602305476, "grad_norm": 1.6011666265746873, "learning_rate": 1e-05, "loss": 0.5605, "step": 621 }, { "epoch": 0.1792507204610951, "grad_norm": 2.7083163727318076, "learning_rate": 1e-05, "loss": 0.3935, "step": 622 }, { "epoch": 0.17953890489913546, "grad_norm": 1.6812217120008799, "learning_rate": 1e-05, "loss": 0.5633, "step": 623 }, { "epoch": 0.17982708933717578, "grad_norm": 1.6809096482590729, "learning_rate": 1e-05, "loss": 0.5554, "step": 624 }, { "epoch": 0.18011527377521613, "grad_norm": 2.679832939952373, "learning_rate": 1e-05, "loss": 0.4127, "step": 625 }, { "epoch": 0.18040345821325648, "grad_norm": 2.756781505942509, "learning_rate": 1e-05, "loss": 0.564, "step": 626 }, { "epoch": 0.18069164265129684, "grad_norm": 1.6826096690716983, "learning_rate": 1e-05, "loss": 0.5281, "step": 627 }, { "epoch": 0.1809798270893372, "grad_norm": 3.861133291619558, "learning_rate": 1e-05, "loss": 0.3834, "step": 628 }, { "epoch": 0.1812680115273775, "grad_norm": 2.485939817416258, "learning_rate": 1e-05, "loss": 0.5667, "step": 629 }, { "epoch": 0.18155619596541786, "grad_norm": 1.6639736644632237, "learning_rate": 1e-05, "loss": 0.5406, "step": 630 }, { "epoch": 0.1818443804034582, "grad_norm": 2.519406384304628, "learning_rate": 1e-05, "loss": 0.4043, "step": 631 }, { "epoch": 0.18213256484149856, "grad_norm": 4.059356699784603, "learning_rate": 1e-05, "loss": 0.528, "step": 632 }, { "epoch": 0.1824207492795389, "grad_norm": 2.1055667595745526, "learning_rate": 1e-05, "loss": 0.5426, "step": 633 }, { "epoch": 0.18270893371757926, "grad_norm": 4.327703295132949, "learning_rate": 1e-05, "loss": 0.3954, "step": 634 }, { "epoch": 0.1829971181556196, "grad_norm": 1.8548168277556574, "learning_rate": 1e-05, "loss": 0.5559, "step": 635 }, { "epoch": 0.18328530259365994, "grad_norm": 1.3956688231830576, "learning_rate": 1e-05, "loss": 0.5349, "step": 636 }, { "epoch": 0.1835734870317003, "grad_norm": 2.9016659878557407, "learning_rate": 1e-05, "loss": 0.403, "step": 637 }, { "epoch": 0.18386167146974064, "grad_norm": 3.005895623836115, "learning_rate": 1e-05, "loss": 0.5422, "step": 638 }, { "epoch": 0.184149855907781, "grad_norm": 2.0525646021136716, "learning_rate": 1e-05, "loss": 0.5513, "step": 639 }, { "epoch": 0.1844380403458213, "grad_norm": 1.6391771026267206, "learning_rate": 1e-05, "loss": 0.3948, "step": 640 }, { "epoch": 0.18472622478386166, "grad_norm": 1.7490421062019743, "learning_rate": 1e-05, "loss": 0.5699, "step": 641 }, { "epoch": 0.18501440922190202, "grad_norm": 1.5702471437958372, "learning_rate": 1e-05, "loss": 0.5545, "step": 642 }, { "epoch": 0.18530259365994237, "grad_norm": 4.152208977372896, "learning_rate": 1e-05, "loss": 0.3948, "step": 643 }, { "epoch": 0.18559077809798272, "grad_norm": 2.2787741438401916, "learning_rate": 1e-05, "loss": 0.5362, "step": 644 }, { "epoch": 0.18587896253602307, "grad_norm": 1.1631130314710791, "learning_rate": 1e-05, "loss": 0.5307, "step": 645 }, { "epoch": 0.1861671469740634, "grad_norm": 2.5882277274941075, "learning_rate": 1e-05, "loss": 0.377, "step": 646 }, { "epoch": 0.18645533141210374, "grad_norm": 2.2812164584511123, "learning_rate": 1e-05, "loss": 0.5317, "step": 647 }, { "epoch": 0.1867435158501441, "grad_norm": 1.672678780235995, "learning_rate": 1e-05, "loss": 0.5354, "step": 648 }, { "epoch": 0.18703170028818444, "grad_norm": 1.818573544529323, "learning_rate": 1e-05, "loss": 0.4098, "step": 649 }, { "epoch": 0.1873198847262248, "grad_norm": 2.2179155359228626, "learning_rate": 1e-05, "loss": 0.5467, "step": 650 }, { "epoch": 0.18760806916426512, "grad_norm": 1.8033963460692448, "learning_rate": 1e-05, "loss": 0.5473, "step": 651 }, { "epoch": 0.18789625360230547, "grad_norm": 3.473690365508948, "learning_rate": 1e-05, "loss": 0.3873, "step": 652 }, { "epoch": 0.18818443804034582, "grad_norm": 2.389456829919027, "learning_rate": 1e-05, "loss": 0.529, "step": 653 }, { "epoch": 0.18847262247838617, "grad_norm": 1.6131730909394986, "learning_rate": 1e-05, "loss": 0.5501, "step": 654 }, { "epoch": 0.18876080691642652, "grad_norm": 1.8623033968265517, "learning_rate": 1e-05, "loss": 0.374, "step": 655 }, { "epoch": 0.18904899135446687, "grad_norm": 2.3707633059771918, "learning_rate": 1e-05, "loss": 0.5654, "step": 656 }, { "epoch": 0.1893371757925072, "grad_norm": 1.68411902183965, "learning_rate": 1e-05, "loss": 0.546, "step": 657 }, { "epoch": 0.18962536023054755, "grad_norm": 6.49456396269767, "learning_rate": 1e-05, "loss": 0.3892, "step": 658 }, { "epoch": 0.1899135446685879, "grad_norm": 3.5905436001612623, "learning_rate": 1e-05, "loss": 0.5414, "step": 659 }, { "epoch": 0.19020172910662825, "grad_norm": 1.5502358348297345, "learning_rate": 1e-05, "loss": 0.5596, "step": 660 }, { "epoch": 0.1904899135446686, "grad_norm": 3.0716995823954765, "learning_rate": 1e-05, "loss": 0.3834, "step": 661 }, { "epoch": 0.19077809798270892, "grad_norm": 2.932377541022994, "learning_rate": 1e-05, "loss": 0.5263, "step": 662 }, { "epoch": 0.19106628242074927, "grad_norm": 2.019552377168776, "learning_rate": 1e-05, "loss": 0.538, "step": 663 }, { "epoch": 0.19135446685878962, "grad_norm": 1.8477246530970768, "learning_rate": 1e-05, "loss": 0.4037, "step": 664 }, { "epoch": 0.19164265129682997, "grad_norm": 2.4470624112951285, "learning_rate": 1e-05, "loss": 0.5781, "step": 665 }, { "epoch": 0.19193083573487033, "grad_norm": 1.410434793166915, "learning_rate": 1e-05, "loss": 0.5366, "step": 666 }, { "epoch": 0.19221902017291068, "grad_norm": 3.4079539947667468, "learning_rate": 1e-05, "loss": 0.3929, "step": 667 }, { "epoch": 0.192507204610951, "grad_norm": 3.2474549702210043, "learning_rate": 1e-05, "loss": 0.5414, "step": 668 }, { "epoch": 0.19279538904899135, "grad_norm": 2.062451109518947, "learning_rate": 1e-05, "loss": 0.5455, "step": 669 }, { "epoch": 0.1930835734870317, "grad_norm": 3.3242299042170345, "learning_rate": 1e-05, "loss": 0.3868, "step": 670 }, { "epoch": 0.19337175792507205, "grad_norm": 2.7024222046537263, "learning_rate": 1e-05, "loss": 0.5606, "step": 671 }, { "epoch": 0.1936599423631124, "grad_norm": 1.7649507400502558, "learning_rate": 1e-05, "loss": 0.5599, "step": 672 }, { "epoch": 0.19394812680115273, "grad_norm": 2.783929652797413, "learning_rate": 1e-05, "loss": 0.3847, "step": 673 }, { "epoch": 0.19423631123919308, "grad_norm": 3.241842986522019, "learning_rate": 1e-05, "loss": 0.5442, "step": 674 }, { "epoch": 0.19452449567723343, "grad_norm": 1.3303643604091253, "learning_rate": 1e-05, "loss": 0.5389, "step": 675 }, { "epoch": 0.19481268011527378, "grad_norm": 1.195358403648017, "learning_rate": 1e-05, "loss": 0.3667, "step": 676 }, { "epoch": 0.19510086455331413, "grad_norm": 3.1427074417202316, "learning_rate": 1e-05, "loss": 0.5527, "step": 677 }, { "epoch": 0.19538904899135448, "grad_norm": 1.4903596488788504, "learning_rate": 1e-05, "loss": 0.5401, "step": 678 }, { "epoch": 0.1956772334293948, "grad_norm": 1.893268281849623, "learning_rate": 1e-05, "loss": 0.3845, "step": 679 }, { "epoch": 0.19596541786743515, "grad_norm": 3.178108430580706, "learning_rate": 1e-05, "loss": 0.5808, "step": 680 }, { "epoch": 0.1962536023054755, "grad_norm": 1.314861960044517, "learning_rate": 1e-05, "loss": 0.5211, "step": 681 }, { "epoch": 0.19654178674351586, "grad_norm": 1.6646688411495776, "learning_rate": 1e-05, "loss": 0.3811, "step": 682 }, { "epoch": 0.1968299711815562, "grad_norm": 1.731434277680317, "learning_rate": 1e-05, "loss": 0.547, "step": 683 }, { "epoch": 0.19711815561959653, "grad_norm": 1.7992172856050568, "learning_rate": 1e-05, "loss": 0.5394, "step": 684 }, { "epoch": 0.19740634005763688, "grad_norm": 2.3807111350441916, "learning_rate": 1e-05, "loss": 0.4015, "step": 685 }, { "epoch": 0.19769452449567723, "grad_norm": 2.7222627650122377, "learning_rate": 1e-05, "loss": 0.5471, "step": 686 }, { "epoch": 0.19798270893371758, "grad_norm": 1.6445933398287433, "learning_rate": 1e-05, "loss": 0.534, "step": 687 }, { "epoch": 0.19827089337175793, "grad_norm": 1.6151151672539203, "learning_rate": 1e-05, "loss": 0.372, "step": 688 }, { "epoch": 0.19855907780979828, "grad_norm": 1.670692914013598, "learning_rate": 1e-05, "loss": 0.576, "step": 689 }, { "epoch": 0.1988472622478386, "grad_norm": 1.6555833971116176, "learning_rate": 1e-05, "loss": 0.5351, "step": 690 }, { "epoch": 0.19913544668587896, "grad_norm": 3.2247248878775276, "learning_rate": 1e-05, "loss": 0.3772, "step": 691 }, { "epoch": 0.1994236311239193, "grad_norm": 1.928176484072151, "learning_rate": 1e-05, "loss": 0.5423, "step": 692 }, { "epoch": 0.19971181556195966, "grad_norm": 1.6063135640174004, "learning_rate": 1e-05, "loss": 0.545, "step": 693 }, { "epoch": 0.2, "grad_norm": 1.871140642527986, "learning_rate": 1e-05, "loss": 0.375, "step": 694 }, { "epoch": 0.20028818443804033, "grad_norm": 1.8679099541746644, "learning_rate": 1e-05, "loss": 0.5682, "step": 695 }, { "epoch": 0.20057636887608069, "grad_norm": 1.7962709463328976, "learning_rate": 1e-05, "loss": 0.531, "step": 696 }, { "epoch": 0.20086455331412104, "grad_norm": 2.493526817760628, "learning_rate": 1e-05, "loss": 0.3761, "step": 697 }, { "epoch": 0.2011527377521614, "grad_norm": 2.435720278720426, "learning_rate": 1e-05, "loss": 0.542, "step": 698 }, { "epoch": 0.20144092219020174, "grad_norm": 1.2839405537920643, "learning_rate": 1e-05, "loss": 0.5419, "step": 699 }, { "epoch": 0.2017291066282421, "grad_norm": 3.593143484480332, "learning_rate": 1e-05, "loss": 0.3861, "step": 700 }, { "epoch": 0.2020172910662824, "grad_norm": 3.4186791670908727, "learning_rate": 1e-05, "loss": 0.5457, "step": 701 }, { "epoch": 0.20230547550432276, "grad_norm": 1.863503051365215, "learning_rate": 1e-05, "loss": 0.5403, "step": 702 }, { "epoch": 0.2025936599423631, "grad_norm": 2.34662159725905, "learning_rate": 1e-05, "loss": 0.3919, "step": 703 }, { "epoch": 0.20288184438040346, "grad_norm": 3.256613608855492, "learning_rate": 1e-05, "loss": 0.5304, "step": 704 }, { "epoch": 0.20317002881844382, "grad_norm": 1.3295205271030683, "learning_rate": 1e-05, "loss": 0.5202, "step": 705 }, { "epoch": 0.20345821325648414, "grad_norm": 3.7719510069622966, "learning_rate": 1e-05, "loss": 0.3744, "step": 706 }, { "epoch": 0.2037463976945245, "grad_norm": 1.9907034061327187, "learning_rate": 1e-05, "loss": 0.5341, "step": 707 }, { "epoch": 0.20403458213256484, "grad_norm": 2.0807541314953215, "learning_rate": 1e-05, "loss": 0.5247, "step": 708 }, { "epoch": 0.2043227665706052, "grad_norm": 2.1837956868707886, "learning_rate": 1e-05, "loss": 0.3812, "step": 709 }, { "epoch": 0.20461095100864554, "grad_norm": 2.347923727511402, "learning_rate": 1e-05, "loss": 0.5407, "step": 710 }, { "epoch": 0.2048991354466859, "grad_norm": 1.8202318162086493, "learning_rate": 1e-05, "loss": 0.5271, "step": 711 }, { "epoch": 0.20518731988472622, "grad_norm": 4.588594008908393, "learning_rate": 1e-05, "loss": 0.3766, "step": 712 }, { "epoch": 0.20547550432276657, "grad_norm": 2.5052719320350194, "learning_rate": 1e-05, "loss": 0.491, "step": 713 }, { "epoch": 0.20576368876080692, "grad_norm": 2.272802999770661, "learning_rate": 1e-05, "loss": 0.5417, "step": 714 }, { "epoch": 0.20605187319884727, "grad_norm": 2.5405417145197604, "learning_rate": 1e-05, "loss": 0.3714, "step": 715 }, { "epoch": 0.20634005763688762, "grad_norm": 2.576715752621875, "learning_rate": 1e-05, "loss": 0.5217, "step": 716 }, { "epoch": 0.20662824207492794, "grad_norm": 2.7334756717737028, "learning_rate": 1e-05, "loss": 0.5387, "step": 717 }, { "epoch": 0.2069164265129683, "grad_norm": 2.4819037880568433, "learning_rate": 1e-05, "loss": 0.3714, "step": 718 }, { "epoch": 0.20720461095100864, "grad_norm": 2.2184880201269226, "learning_rate": 1e-05, "loss": 0.5229, "step": 719 }, { "epoch": 0.207492795389049, "grad_norm": 1.5608036614364995, "learning_rate": 1e-05, "loss": 0.5182, "step": 720 }, { "epoch": 0.20778097982708935, "grad_norm": 1.8292187821414867, "learning_rate": 1e-05, "loss": 0.3629, "step": 721 }, { "epoch": 0.2080691642651297, "grad_norm": 2.266258143468976, "learning_rate": 1e-05, "loss": 0.5391, "step": 722 }, { "epoch": 0.20835734870317002, "grad_norm": 1.6281311577197308, "learning_rate": 1e-05, "loss": 0.5395, "step": 723 }, { "epoch": 0.20864553314121037, "grad_norm": 2.8517201813458475, "learning_rate": 1e-05, "loss": 0.3878, "step": 724 }, { "epoch": 0.20893371757925072, "grad_norm": 2.12071000582212, "learning_rate": 1e-05, "loss": 0.544, "step": 725 }, { "epoch": 0.20922190201729107, "grad_norm": 1.4786087983239724, "learning_rate": 1e-05, "loss": 0.5314, "step": 726 }, { "epoch": 0.20951008645533142, "grad_norm": 1.5577610909440798, "learning_rate": 1e-05, "loss": 0.3701, "step": 727 }, { "epoch": 0.20979827089337175, "grad_norm": 3.1641406948540847, "learning_rate": 1e-05, "loss": 0.5592, "step": 728 }, { "epoch": 0.2100864553314121, "grad_norm": 1.4810639216076793, "learning_rate": 1e-05, "loss": 0.5247, "step": 729 }, { "epoch": 0.21037463976945245, "grad_norm": 3.1653545929532205, "learning_rate": 1e-05, "loss": 0.3695, "step": 730 }, { "epoch": 0.2106628242074928, "grad_norm": 2.1220789892916563, "learning_rate": 1e-05, "loss": 0.5401, "step": 731 }, { "epoch": 0.21095100864553315, "grad_norm": 2.251796845084247, "learning_rate": 1e-05, "loss": 0.5468, "step": 732 }, { "epoch": 0.21123919308357347, "grad_norm": 1.8026942421448902, "learning_rate": 1e-05, "loss": 0.3785, "step": 733 }, { "epoch": 0.21152737752161382, "grad_norm": 2.674342159575944, "learning_rate": 1e-05, "loss": 0.5194, "step": 734 }, { "epoch": 0.21181556195965417, "grad_norm": 1.6456776112403628, "learning_rate": 1e-05, "loss": 0.5291, "step": 735 }, { "epoch": 0.21210374639769453, "grad_norm": 12.680453282507559, "learning_rate": 1e-05, "loss": 0.3911, "step": 736 }, { "epoch": 0.21239193083573488, "grad_norm": 2.691332725684379, "learning_rate": 1e-05, "loss": 0.5383, "step": 737 }, { "epoch": 0.21268011527377523, "grad_norm": 1.5529169372421723, "learning_rate": 1e-05, "loss": 0.5354, "step": 738 }, { "epoch": 0.21296829971181555, "grad_norm": 6.518363469400075, "learning_rate": 1e-05, "loss": 0.3985, "step": 739 }, { "epoch": 0.2132564841498559, "grad_norm": 2.3740170317253706, "learning_rate": 1e-05, "loss": 0.5679, "step": 740 }, { "epoch": 0.21354466858789625, "grad_norm": 1.7889993945011, "learning_rate": 1e-05, "loss": 0.5279, "step": 741 }, { "epoch": 0.2138328530259366, "grad_norm": 3.515910998913752, "learning_rate": 1e-05, "loss": 0.3865, "step": 742 }, { "epoch": 0.21412103746397695, "grad_norm": 2.548307289511562, "learning_rate": 1e-05, "loss": 0.5277, "step": 743 }, { "epoch": 0.21440922190201728, "grad_norm": 1.3488652772286054, "learning_rate": 1e-05, "loss": 0.5174, "step": 744 }, { "epoch": 0.21469740634005763, "grad_norm": 4.249962624071357, "learning_rate": 1e-05, "loss": 0.3561, "step": 745 }, { "epoch": 0.21498559077809798, "grad_norm": 2.4901332890788437, "learning_rate": 1e-05, "loss": 0.5815, "step": 746 }, { "epoch": 0.21527377521613833, "grad_norm": 1.3670992784434364, "learning_rate": 1e-05, "loss": 0.5222, "step": 747 }, { "epoch": 0.21556195965417868, "grad_norm": 2.110011740407682, "learning_rate": 1e-05, "loss": 0.378, "step": 748 }, { "epoch": 0.21585014409221903, "grad_norm": 2.1655089030291013, "learning_rate": 1e-05, "loss": 0.5121, "step": 749 }, { "epoch": 0.21613832853025935, "grad_norm": 1.4820566524783922, "learning_rate": 1e-05, "loss": 0.5469, "step": 750 }, { "epoch": 0.21613832853025935, "eval_websight_new_IoU": 0.01727907732129097, "eval_websight_new_MAE_x": 286.7799987792969, "eval_websight_new_MAE_y": 523.760009765625, "eval_websight_new_coord_probability": 0.7429283261299133, "eval_websight_new_correct_token_probability": 0.0011671152897179127, "eval_websight_new_inside_bbox": 0.019999999552965164, "eval_websight_new_loss": 0.3837199807167053, "eval_websight_new_range_100_probability": 0.09132597595453262, "eval_websight_new_range_20_probability": 0.01654745265841484, "eval_websight_new_range_5_probability": 0.003870067186653614, "eval_websight_new_runtime": 33.8429, "eval_websight_new_samples_per_second": 1.477, "eval_websight_new_steps_per_second": 0.207, "step": 750 }, { "epoch": 0.21613832853025935, "eval_seeclick_IoU": 0.03237491846084595, "eval_seeclick_MAE_x": 473.1000061035156, "eval_seeclick_MAE_y": 515.7999877929688, "eval_seeclick_coord_probability": 0.7453280687332153, "eval_seeclick_correct_token_probability": 0.001050356775522232, "eval_seeclick_inside_bbox": 0.05999999865889549, "eval_seeclick_loss": 0.4347634017467499, "eval_seeclick_range_100_probability": 0.11753538995981216, "eval_seeclick_range_20_probability": 0.03401518613100052, "eval_seeclick_range_5_probability": 0.006139020901173353, "eval_seeclick_runtime": 42.062, "eval_seeclick_samples_per_second": 1.189, "eval_seeclick_steps_per_second": 0.166, "step": 750 }, { "epoch": 0.21613832853025935, "eval_icons_IoU": 0.003064297139644623, "eval_icons_MAE_x": 473.5400085449219, "eval_icons_MAE_y": 634.4000244140625, "eval_icons_coord_probability": 0.7481094598770142, "eval_icons_correct_token_probability": 0.005065434146672487, "eval_icons_inside_bbox": 0.0, "eval_icons_loss": 0.5070137977600098, "eval_icons_range_100_probability": 0.2729437053203583, "eval_icons_range_20_probability": 0.09965262562036514, "eval_icons_range_5_probability": 0.02936328575015068, "eval_icons_runtime": 43.2884, "eval_icons_samples_per_second": 1.155, "eval_icons_steps_per_second": 0.162, "step": 750 }, { "epoch": 0.2164265129682997, "grad_norm": 2.1888763443405295, "learning_rate": 1e-05, "loss": 0.3598, "step": 751 }, { "epoch": 0.21671469740634006, "grad_norm": 2.2774890567448027, "learning_rate": 1e-05, "loss": 0.5035, "step": 752 }, { "epoch": 0.2170028818443804, "grad_norm": 1.4442456121352483, "learning_rate": 1e-05, "loss": 0.5184, "step": 753 }, { "epoch": 0.21729106628242076, "grad_norm": 3.785601337006597, "learning_rate": 1e-05, "loss": 0.3839, "step": 754 }, { "epoch": 0.21757925072046108, "grad_norm": 3.1321856047605956, "learning_rate": 1e-05, "loss": 0.5039, "step": 755 }, { "epoch": 0.21786743515850143, "grad_norm": 1.9440011815842064, "learning_rate": 1e-05, "loss": 0.5302, "step": 756 }, { "epoch": 0.21815561959654178, "grad_norm": 2.7511348005259344, "learning_rate": 1e-05, "loss": 0.3573, "step": 757 }, { "epoch": 0.21844380403458213, "grad_norm": 2.775030827757665, "learning_rate": 1e-05, "loss": 0.5631, "step": 758 }, { "epoch": 0.21873198847262249, "grad_norm": 1.933287538834539, "learning_rate": 1e-05, "loss": 0.5257, "step": 759 }, { "epoch": 0.21902017291066284, "grad_norm": 1.6546472853128393, "learning_rate": 1e-05, "loss": 0.3678, "step": 760 }, { "epoch": 0.21930835734870316, "grad_norm": 2.604311079856137, "learning_rate": 1e-05, "loss": 0.5335, "step": 761 }, { "epoch": 0.2195965417867435, "grad_norm": 2.2191616680988573, "learning_rate": 1e-05, "loss": 0.5456, "step": 762 }, { "epoch": 0.21988472622478386, "grad_norm": 2.3295803403283553, "learning_rate": 1e-05, "loss": 0.3866, "step": 763 }, { "epoch": 0.2201729106628242, "grad_norm": 1.9475593674895146, "learning_rate": 1e-05, "loss": 0.5194, "step": 764 }, { "epoch": 0.22046109510086456, "grad_norm": 1.6865455352852274, "learning_rate": 1e-05, "loss": 0.5289, "step": 765 }, { "epoch": 0.22074927953890489, "grad_norm": 1.7278367147952058, "learning_rate": 1e-05, "loss": 0.3569, "step": 766 }, { "epoch": 0.22103746397694524, "grad_norm": 2.9832195025090593, "learning_rate": 1e-05, "loss": 0.5357, "step": 767 }, { "epoch": 0.2213256484149856, "grad_norm": 1.3242254653617855, "learning_rate": 1e-05, "loss": 0.5151, "step": 768 }, { "epoch": 0.22161383285302594, "grad_norm": 1.8377308740842033, "learning_rate": 1e-05, "loss": 0.373, "step": 769 }, { "epoch": 0.2219020172910663, "grad_norm": 2.2036161876977163, "learning_rate": 1e-05, "loss": 0.5327, "step": 770 }, { "epoch": 0.22219020172910664, "grad_norm": 1.6665259044580085, "learning_rate": 1e-05, "loss": 0.5187, "step": 771 }, { "epoch": 0.22247838616714696, "grad_norm": 2.606597787113097, "learning_rate": 1e-05, "loss": 0.3599, "step": 772 }, { "epoch": 0.22276657060518731, "grad_norm": 3.7691354716746854, "learning_rate": 1e-05, "loss": 0.5282, "step": 773 }, { "epoch": 0.22305475504322766, "grad_norm": 1.926087044611918, "learning_rate": 1e-05, "loss": 0.5186, "step": 774 }, { "epoch": 0.22334293948126802, "grad_norm": 2.1872704449275466, "learning_rate": 1e-05, "loss": 0.3543, "step": 775 }, { "epoch": 0.22363112391930837, "grad_norm": 2.442419185379171, "learning_rate": 1e-05, "loss": 0.5499, "step": 776 }, { "epoch": 0.2239193083573487, "grad_norm": 1.7615743180506984, "learning_rate": 1e-05, "loss": 0.5129, "step": 777 }, { "epoch": 0.22420749279538904, "grad_norm": 3.5936095177088605, "learning_rate": 1e-05, "loss": 0.3595, "step": 778 }, { "epoch": 0.2244956772334294, "grad_norm": 3.9778261145163514, "learning_rate": 1e-05, "loss": 0.5125, "step": 779 }, { "epoch": 0.22478386167146974, "grad_norm": 1.9736950314580592, "learning_rate": 1e-05, "loss": 0.5193, "step": 780 }, { "epoch": 0.2250720461095101, "grad_norm": 2.964728492787288, "learning_rate": 1e-05, "loss": 0.3669, "step": 781 }, { "epoch": 0.22536023054755044, "grad_norm": 2.784188417504383, "learning_rate": 1e-05, "loss": 0.5407, "step": 782 }, { "epoch": 0.22564841498559077, "grad_norm": 3.202046534108722, "learning_rate": 1e-05, "loss": 0.5128, "step": 783 }, { "epoch": 0.22593659942363112, "grad_norm": 2.0773279346175006, "learning_rate": 1e-05, "loss": 0.3715, "step": 784 }, { "epoch": 0.22622478386167147, "grad_norm": 2.7721339436817463, "learning_rate": 1e-05, "loss": 0.5678, "step": 785 }, { "epoch": 0.22651296829971182, "grad_norm": 2.052925896005973, "learning_rate": 1e-05, "loss": 0.5251, "step": 786 }, { "epoch": 0.22680115273775217, "grad_norm": 1.7600006840401639, "learning_rate": 1e-05, "loss": 0.3686, "step": 787 }, { "epoch": 0.2270893371757925, "grad_norm": 3.076061233667846, "learning_rate": 1e-05, "loss": 0.5062, "step": 788 }, { "epoch": 0.22737752161383284, "grad_norm": 1.8007049947721312, "learning_rate": 1e-05, "loss": 0.5221, "step": 789 }, { "epoch": 0.2276657060518732, "grad_norm": 3.1757006191847856, "learning_rate": 1e-05, "loss": 0.3777, "step": 790 }, { "epoch": 0.22795389048991355, "grad_norm": 2.836473562759461, "learning_rate": 1e-05, "loss": 0.5497, "step": 791 }, { "epoch": 0.2282420749279539, "grad_norm": 1.9915125344875975, "learning_rate": 1e-05, "loss": 0.517, "step": 792 }, { "epoch": 0.22853025936599425, "grad_norm": 1.9379971324561747, "learning_rate": 1e-05, "loss": 0.3648, "step": 793 }, { "epoch": 0.22881844380403457, "grad_norm": 2.420860422194887, "learning_rate": 1e-05, "loss": 0.5482, "step": 794 }, { "epoch": 0.22910662824207492, "grad_norm": 2.1897133743195734, "learning_rate": 1e-05, "loss": 0.5273, "step": 795 }, { "epoch": 0.22939481268011527, "grad_norm": 1.319229454100792, "learning_rate": 1e-05, "loss": 0.3784, "step": 796 }, { "epoch": 0.22968299711815562, "grad_norm": 2.118156858440375, "learning_rate": 1e-05, "loss": 0.5125, "step": 797 }, { "epoch": 0.22997118155619597, "grad_norm": 1.448708704573136, "learning_rate": 1e-05, "loss": 0.5189, "step": 798 }, { "epoch": 0.2302593659942363, "grad_norm": 2.0463003698498, "learning_rate": 1e-05, "loss": 0.3895, "step": 799 }, { "epoch": 0.23054755043227665, "grad_norm": 2.006341116386733, "learning_rate": 1e-05, "loss": 0.5205, "step": 800 }, { "epoch": 0.230835734870317, "grad_norm": 2.0493659703648808, "learning_rate": 1e-05, "loss": 0.5181, "step": 801 }, { "epoch": 0.23112391930835735, "grad_norm": 1.4723595443050466, "learning_rate": 1e-05, "loss": 0.3702, "step": 802 }, { "epoch": 0.2314121037463977, "grad_norm": 1.9647835931699873, "learning_rate": 1e-05, "loss": 0.5322, "step": 803 }, { "epoch": 0.23170028818443805, "grad_norm": 1.5165203723354632, "learning_rate": 1e-05, "loss": 0.524, "step": 804 }, { "epoch": 0.23198847262247838, "grad_norm": 3.1727305578969958, "learning_rate": 1e-05, "loss": 0.3691, "step": 805 }, { "epoch": 0.23227665706051873, "grad_norm": 2.455598470170328, "learning_rate": 1e-05, "loss": 0.5294, "step": 806 }, { "epoch": 0.23256484149855908, "grad_norm": 1.6062968633545127, "learning_rate": 1e-05, "loss": 0.5264, "step": 807 }, { "epoch": 0.23285302593659943, "grad_norm": 1.6464516538182972, "learning_rate": 1e-05, "loss": 0.3525, "step": 808 }, { "epoch": 0.23314121037463978, "grad_norm": 2.9692256786690727, "learning_rate": 1e-05, "loss": 0.5277, "step": 809 }, { "epoch": 0.2334293948126801, "grad_norm": 1.6283622611780912, "learning_rate": 1e-05, "loss": 0.5228, "step": 810 }, { "epoch": 0.23371757925072045, "grad_norm": 3.452754600099684, "learning_rate": 1e-05, "loss": 0.352, "step": 811 }, { "epoch": 0.2340057636887608, "grad_norm": 2.7458062624189483, "learning_rate": 1e-05, "loss": 0.5292, "step": 812 }, { "epoch": 0.23429394812680115, "grad_norm": 1.4757636230943205, "learning_rate": 1e-05, "loss": 0.5178, "step": 813 }, { "epoch": 0.2345821325648415, "grad_norm": 3.401862068032278, "learning_rate": 1e-05, "loss": 0.3441, "step": 814 }, { "epoch": 0.23487031700288186, "grad_norm": 2.4224097077513997, "learning_rate": 1e-05, "loss": 0.5594, "step": 815 }, { "epoch": 0.23515850144092218, "grad_norm": 1.5672899266594948, "learning_rate": 1e-05, "loss": 0.5232, "step": 816 }, { "epoch": 0.23544668587896253, "grad_norm": 1.956384688894368, "learning_rate": 1e-05, "loss": 0.3676, "step": 817 }, { "epoch": 0.23573487031700288, "grad_norm": 3.716738993954289, "learning_rate": 1e-05, "loss": 0.5148, "step": 818 }, { "epoch": 0.23602305475504323, "grad_norm": 1.9951514137318538, "learning_rate": 1e-05, "loss": 0.5198, "step": 819 }, { "epoch": 0.23631123919308358, "grad_norm": 3.4242391430302153, "learning_rate": 1e-05, "loss": 0.3363, "step": 820 }, { "epoch": 0.2365994236311239, "grad_norm": 2.772418569518423, "learning_rate": 1e-05, "loss": 0.5342, "step": 821 }, { "epoch": 0.23688760806916426, "grad_norm": 1.4757115320734255, "learning_rate": 1e-05, "loss": 0.5151, "step": 822 }, { "epoch": 0.2371757925072046, "grad_norm": 2.1011737845303204, "learning_rate": 1e-05, "loss": 0.3586, "step": 823 }, { "epoch": 0.23746397694524496, "grad_norm": 2.9805427737176577, "learning_rate": 1e-05, "loss": 0.5272, "step": 824 }, { "epoch": 0.2377521613832853, "grad_norm": 1.3117274089431927, "learning_rate": 1e-05, "loss": 0.5079, "step": 825 }, { "epoch": 0.23804034582132566, "grad_norm": 2.3328559839201892, "learning_rate": 1e-05, "loss": 0.3454, "step": 826 }, { "epoch": 0.23832853025936598, "grad_norm": 3.856422134773823, "learning_rate": 1e-05, "loss": 0.5304, "step": 827 }, { "epoch": 0.23861671469740633, "grad_norm": 1.4266057337400306, "learning_rate": 1e-05, "loss": 0.5173, "step": 828 }, { "epoch": 0.23890489913544669, "grad_norm": 1.8049556866780567, "learning_rate": 1e-05, "loss": 0.3554, "step": 829 }, { "epoch": 0.23919308357348704, "grad_norm": 3.0646174126487233, "learning_rate": 1e-05, "loss": 0.5417, "step": 830 }, { "epoch": 0.2394812680115274, "grad_norm": 1.5027819303170922, "learning_rate": 1e-05, "loss": 0.5095, "step": 831 }, { "epoch": 0.2397694524495677, "grad_norm": 2.8189905860249995, "learning_rate": 1e-05, "loss": 0.374, "step": 832 }, { "epoch": 0.24005763688760806, "grad_norm": 2.7766436345114687, "learning_rate": 1e-05, "loss": 0.5262, "step": 833 }, { "epoch": 0.2403458213256484, "grad_norm": 1.6962941988775544, "learning_rate": 1e-05, "loss": 0.5094, "step": 834 }, { "epoch": 0.24063400576368876, "grad_norm": 3.9779864695789833, "learning_rate": 1e-05, "loss": 0.3477, "step": 835 }, { "epoch": 0.24092219020172911, "grad_norm": 3.946522752119048, "learning_rate": 1e-05, "loss": 0.5166, "step": 836 }, { "epoch": 0.24121037463976946, "grad_norm": 2.3943887033548426, "learning_rate": 1e-05, "loss": 0.5168, "step": 837 }, { "epoch": 0.2414985590778098, "grad_norm": 2.329503948048036, "learning_rate": 1e-05, "loss": 0.35, "step": 838 }, { "epoch": 0.24178674351585014, "grad_norm": 2.4187939347443494, "learning_rate": 1e-05, "loss": 0.4892, "step": 839 }, { "epoch": 0.2420749279538905, "grad_norm": 1.47487900374655, "learning_rate": 1e-05, "loss": 0.5119, "step": 840 }, { "epoch": 0.24236311239193084, "grad_norm": 3.800528926250578, "learning_rate": 1e-05, "loss": 0.3457, "step": 841 }, { "epoch": 0.2426512968299712, "grad_norm": 2.5069874848836817, "learning_rate": 1e-05, "loss": 0.4978, "step": 842 }, { "epoch": 0.24293948126801151, "grad_norm": 1.8611740241971595, "learning_rate": 1e-05, "loss": 0.517, "step": 843 }, { "epoch": 0.24322766570605187, "grad_norm": 1.8766061195321277, "learning_rate": 1e-05, "loss": 0.3647, "step": 844 }, { "epoch": 0.24351585014409222, "grad_norm": 4.214162592822582, "learning_rate": 1e-05, "loss": 0.5036, "step": 845 }, { "epoch": 0.24380403458213257, "grad_norm": 2.2303791341440635, "learning_rate": 1e-05, "loss": 0.5178, "step": 846 }, { "epoch": 0.24409221902017292, "grad_norm": 2.0840787909368825, "learning_rate": 1e-05, "loss": 0.3632, "step": 847 }, { "epoch": 0.24438040345821327, "grad_norm": 2.0604828244399993, "learning_rate": 1e-05, "loss": 0.5498, "step": 848 }, { "epoch": 0.2446685878962536, "grad_norm": 1.784944626914137, "learning_rate": 1e-05, "loss": 0.5183, "step": 849 }, { "epoch": 0.24495677233429394, "grad_norm": 2.0313924719542897, "learning_rate": 1e-05, "loss": 0.3522, "step": 850 }, { "epoch": 0.2452449567723343, "grad_norm": 2.52825075621921, "learning_rate": 1e-05, "loss": 0.5492, "step": 851 }, { "epoch": 0.24553314121037464, "grad_norm": 1.3334641465216632, "learning_rate": 1e-05, "loss": 0.5258, "step": 852 }, { "epoch": 0.245821325648415, "grad_norm": 2.2192424001626936, "learning_rate": 1e-05, "loss": 0.364, "step": 853 }, { "epoch": 0.24610951008645532, "grad_norm": 2.460333177583637, "learning_rate": 1e-05, "loss": 0.5014, "step": 854 }, { "epoch": 0.24639769452449567, "grad_norm": 2.606972913728769, "learning_rate": 1e-05, "loss": 0.5121, "step": 855 }, { "epoch": 0.24668587896253602, "grad_norm": 1.4715149103233114, "learning_rate": 1e-05, "loss": 0.3742, "step": 856 }, { "epoch": 0.24697406340057637, "grad_norm": 1.8883963445055671, "learning_rate": 1e-05, "loss": 0.5134, "step": 857 }, { "epoch": 0.24726224783861672, "grad_norm": 1.5747373487704703, "learning_rate": 1e-05, "loss": 0.5187, "step": 858 }, { "epoch": 0.24755043227665707, "grad_norm": 1.6956743231061768, "learning_rate": 1e-05, "loss": 0.3711, "step": 859 }, { "epoch": 0.2478386167146974, "grad_norm": 2.0715775798740377, "learning_rate": 1e-05, "loss": 0.5167, "step": 860 }, { "epoch": 0.24812680115273775, "grad_norm": 1.4829408931933024, "learning_rate": 1e-05, "loss": 0.5145, "step": 861 }, { "epoch": 0.2484149855907781, "grad_norm": 3.0141285576727066, "learning_rate": 1e-05, "loss": 0.3511, "step": 862 }, { "epoch": 0.24870317002881845, "grad_norm": 2.4860482609325785, "learning_rate": 1e-05, "loss": 0.534, "step": 863 }, { "epoch": 0.2489913544668588, "grad_norm": 1.2180242319881303, "learning_rate": 1e-05, "loss": 0.5122, "step": 864 }, { "epoch": 0.24927953890489912, "grad_norm": 1.5460133411288588, "learning_rate": 1e-05, "loss": 0.3442, "step": 865 }, { "epoch": 0.24956772334293947, "grad_norm": 2.270173704845945, "learning_rate": 1e-05, "loss": 0.5334, "step": 866 }, { "epoch": 0.24985590778097982, "grad_norm": 1.704269971401325, "learning_rate": 1e-05, "loss": 0.5026, "step": 867 }, { "epoch": 0.2501440922190202, "grad_norm": 1.4900878720973898, "learning_rate": 1e-05, "loss": 0.3544, "step": 868 }, { "epoch": 0.2504322766570605, "grad_norm": 2.5843336862710884, "learning_rate": 1e-05, "loss": 0.5405, "step": 869 }, { "epoch": 0.2507204610951009, "grad_norm": 1.3856792728436378, "learning_rate": 1e-05, "loss": 0.5126, "step": 870 }, { "epoch": 0.25100864553314123, "grad_norm": 1.772133582598731, "learning_rate": 1e-05, "loss": 0.3522, "step": 871 }, { "epoch": 0.2512968299711816, "grad_norm": 2.5978131685660655, "learning_rate": 1e-05, "loss": 0.5394, "step": 872 }, { "epoch": 0.2515850144092219, "grad_norm": 1.4779713020696061, "learning_rate": 1e-05, "loss": 0.5172, "step": 873 }, { "epoch": 0.2518731988472622, "grad_norm": 2.9554944616355, "learning_rate": 1e-05, "loss": 0.3455, "step": 874 }, { "epoch": 0.2521613832853026, "grad_norm": 3.6095839901225832, "learning_rate": 1e-05, "loss": 0.5261, "step": 875 }, { "epoch": 0.2524495677233429, "grad_norm": 1.863377495387552, "learning_rate": 1e-05, "loss": 0.5096, "step": 876 }, { "epoch": 0.2527377521613833, "grad_norm": 2.6368724875045957, "learning_rate": 1e-05, "loss": 0.3461, "step": 877 }, { "epoch": 0.25302593659942363, "grad_norm": 2.23478728416295, "learning_rate": 1e-05, "loss": 0.5454, "step": 878 }, { "epoch": 0.253314121037464, "grad_norm": 2.1140083394089895, "learning_rate": 1e-05, "loss": 0.5199, "step": 879 }, { "epoch": 0.25360230547550433, "grad_norm": 1.6564815866196856, "learning_rate": 1e-05, "loss": 0.3376, "step": 880 }, { "epoch": 0.2538904899135447, "grad_norm": 2.2729324647369564, "learning_rate": 1e-05, "loss": 0.5389, "step": 881 }, { "epoch": 0.25417867435158503, "grad_norm": 1.917905469875635, "learning_rate": 1e-05, "loss": 0.5209, "step": 882 }, { "epoch": 0.2544668587896254, "grad_norm": 2.495268518614722, "learning_rate": 1e-05, "loss": 0.3325, "step": 883 }, { "epoch": 0.2547550432276657, "grad_norm": 2.20632661090717, "learning_rate": 1e-05, "loss": 0.5398, "step": 884 }, { "epoch": 0.25504322766570603, "grad_norm": 2.978182106779898, "learning_rate": 1e-05, "loss": 0.5031, "step": 885 }, { "epoch": 0.2553314121037464, "grad_norm": 2.023394065311876, "learning_rate": 1e-05, "loss": 0.3531, "step": 886 }, { "epoch": 0.25561959654178673, "grad_norm": 2.9384319821079714, "learning_rate": 1e-05, "loss": 0.5277, "step": 887 }, { "epoch": 0.2559077809798271, "grad_norm": 2.305370048844002, "learning_rate": 1e-05, "loss": 0.5057, "step": 888 }, { "epoch": 0.25619596541786743, "grad_norm": 2.6623452134444583, "learning_rate": 1e-05, "loss": 0.3657, "step": 889 }, { "epoch": 0.2564841498559078, "grad_norm": 2.5367479962613975, "learning_rate": 1e-05, "loss": 0.5066, "step": 890 }, { "epoch": 0.25677233429394813, "grad_norm": 1.419286418468924, "learning_rate": 1e-05, "loss": 0.5091, "step": 891 }, { "epoch": 0.2570605187319885, "grad_norm": 2.1384303010117605, "learning_rate": 1e-05, "loss": 0.3489, "step": 892 }, { "epoch": 0.25734870317002884, "grad_norm": 3.2342022984795125, "learning_rate": 1e-05, "loss": 0.527, "step": 893 }, { "epoch": 0.2576368876080692, "grad_norm": 2.0204922629256608, "learning_rate": 1e-05, "loss": 0.5069, "step": 894 }, { "epoch": 0.2579250720461095, "grad_norm": 4.76560770896601, "learning_rate": 1e-05, "loss": 0.3782, "step": 895 }, { "epoch": 0.25821325648414983, "grad_norm": 3.2405572098774535, "learning_rate": 1e-05, "loss": 0.512, "step": 896 }, { "epoch": 0.2585014409221902, "grad_norm": 1.799772119453473, "learning_rate": 1e-05, "loss": 0.5146, "step": 897 }, { "epoch": 0.25878962536023054, "grad_norm": 1.8846168408199326, "learning_rate": 1e-05, "loss": 0.3511, "step": 898 }, { "epoch": 0.2590778097982709, "grad_norm": 3.2201950166329483, "learning_rate": 1e-05, "loss": 0.5116, "step": 899 }, { "epoch": 0.25936599423631124, "grad_norm": 2.538171058228147, "learning_rate": 1e-05, "loss": 0.5338, "step": 900 }, { "epoch": 0.2596541786743516, "grad_norm": 3.35038891535301, "learning_rate": 1e-05, "loss": 0.3602, "step": 901 }, { "epoch": 0.25994236311239194, "grad_norm": 2.3003666047019924, "learning_rate": 1e-05, "loss": 0.5488, "step": 902 }, { "epoch": 0.2602305475504323, "grad_norm": 1.6378904104692014, "learning_rate": 1e-05, "loss": 0.5089, "step": 903 }, { "epoch": 0.26051873198847264, "grad_norm": 1.9298438110931269, "learning_rate": 1e-05, "loss": 0.3465, "step": 904 }, { "epoch": 0.260806916426513, "grad_norm": 2.4619128302226434, "learning_rate": 1e-05, "loss": 0.528, "step": 905 }, { "epoch": 0.2610951008645533, "grad_norm": 1.3476313349250084, "learning_rate": 1e-05, "loss": 0.5095, "step": 906 }, { "epoch": 0.26138328530259364, "grad_norm": 2.7887628111368166, "learning_rate": 1e-05, "loss": 0.3477, "step": 907 }, { "epoch": 0.261671469740634, "grad_norm": 2.365357342524135, "learning_rate": 1e-05, "loss": 0.4832, "step": 908 }, { "epoch": 0.26195965417867434, "grad_norm": 2.3627181380185625, "learning_rate": 1e-05, "loss": 0.5089, "step": 909 }, { "epoch": 0.2622478386167147, "grad_norm": 2.876074731526754, "learning_rate": 1e-05, "loss": 0.3697, "step": 910 }, { "epoch": 0.26253602305475504, "grad_norm": 2.6280887251899174, "learning_rate": 1e-05, "loss": 0.5383, "step": 911 }, { "epoch": 0.2628242074927954, "grad_norm": 1.6832899182461254, "learning_rate": 1e-05, "loss": 0.5016, "step": 912 }, { "epoch": 0.26311239193083574, "grad_norm": 1.1162671633980008, "learning_rate": 1e-05, "loss": 0.3379, "step": 913 }, { "epoch": 0.2634005763688761, "grad_norm": 3.0120401514277373, "learning_rate": 1e-05, "loss": 0.528, "step": 914 }, { "epoch": 0.26368876080691644, "grad_norm": 1.416544071898569, "learning_rate": 1e-05, "loss": 0.5118, "step": 915 }, { "epoch": 0.2639769452449568, "grad_norm": 1.7925956476565545, "learning_rate": 1e-05, "loss": 0.3528, "step": 916 }, { "epoch": 0.2642651296829971, "grad_norm": 2.6421410122111126, "learning_rate": 1e-05, "loss": 0.5276, "step": 917 }, { "epoch": 0.26455331412103744, "grad_norm": 1.4449283784010145, "learning_rate": 1e-05, "loss": 0.5097, "step": 918 }, { "epoch": 0.2648414985590778, "grad_norm": 1.62279020218201, "learning_rate": 1e-05, "loss": 0.3468, "step": 919 }, { "epoch": 0.26512968299711814, "grad_norm": 2.412267330061665, "learning_rate": 1e-05, "loss": 0.5339, "step": 920 }, { "epoch": 0.2654178674351585, "grad_norm": 2.012484193733642, "learning_rate": 1e-05, "loss": 0.501, "step": 921 }, { "epoch": 0.26570605187319885, "grad_norm": 1.3534784316568522, "learning_rate": 1e-05, "loss": 0.339, "step": 922 }, { "epoch": 0.2659942363112392, "grad_norm": 2.7367124146915134, "learning_rate": 1e-05, "loss": 0.5254, "step": 923 }, { "epoch": 0.26628242074927955, "grad_norm": 1.9157433438308276, "learning_rate": 1e-05, "loss": 0.5042, "step": 924 }, { "epoch": 0.2665706051873199, "grad_norm": 2.141332264753019, "learning_rate": 1e-05, "loss": 0.3566, "step": 925 }, { "epoch": 0.26685878962536025, "grad_norm": 2.1711057484674297, "learning_rate": 1e-05, "loss": 0.5131, "step": 926 }, { "epoch": 0.2671469740634006, "grad_norm": 1.4690749120533202, "learning_rate": 1e-05, "loss": 0.5162, "step": 927 }, { "epoch": 0.2674351585014409, "grad_norm": 2.6789176677358673, "learning_rate": 1e-05, "loss": 0.3531, "step": 928 }, { "epoch": 0.26772334293948125, "grad_norm": 2.264548478937063, "learning_rate": 1e-05, "loss": 0.5097, "step": 929 }, { "epoch": 0.2680115273775216, "grad_norm": 1.5144619403127813, "learning_rate": 1e-05, "loss": 0.5055, "step": 930 }, { "epoch": 0.26829971181556195, "grad_norm": 2.552370856032433, "learning_rate": 1e-05, "loss": 0.3574, "step": 931 }, { "epoch": 0.2685878962536023, "grad_norm": 3.0566838022283416, "learning_rate": 1e-05, "loss": 0.5114, "step": 932 }, { "epoch": 0.26887608069164265, "grad_norm": 1.9664727540315903, "learning_rate": 1e-05, "loss": 0.5135, "step": 933 }, { "epoch": 0.269164265129683, "grad_norm": 1.9116619780341022, "learning_rate": 1e-05, "loss": 0.3532, "step": 934 }, { "epoch": 0.26945244956772335, "grad_norm": 2.749209246235332, "learning_rate": 1e-05, "loss": 0.5303, "step": 935 }, { "epoch": 0.2697406340057637, "grad_norm": 1.9429900975170704, "learning_rate": 1e-05, "loss": 0.5016, "step": 936 }, { "epoch": 0.27002881844380405, "grad_norm": 2.1919804364456263, "learning_rate": 1e-05, "loss": 0.3363, "step": 937 }, { "epoch": 0.2703170028818444, "grad_norm": 2.9447218226456044, "learning_rate": 1e-05, "loss": 0.542, "step": 938 }, { "epoch": 0.2706051873198847, "grad_norm": 1.5941809809269842, "learning_rate": 1e-05, "loss": 0.5145, "step": 939 }, { "epoch": 0.27089337175792505, "grad_norm": 1.7321234967773047, "learning_rate": 1e-05, "loss": 0.357, "step": 940 }, { "epoch": 0.2711815561959654, "grad_norm": 3.3328482893388967, "learning_rate": 1e-05, "loss": 0.4793, "step": 941 }, { "epoch": 0.27146974063400575, "grad_norm": 1.5301144291370694, "learning_rate": 1e-05, "loss": 0.512, "step": 942 }, { "epoch": 0.2717579250720461, "grad_norm": 2.162147900076396, "learning_rate": 1e-05, "loss": 0.3378, "step": 943 }, { "epoch": 0.27204610951008645, "grad_norm": 2.82530953683193, "learning_rate": 1e-05, "loss": 0.5237, "step": 944 }, { "epoch": 0.2723342939481268, "grad_norm": 1.8044675645639783, "learning_rate": 1e-05, "loss": 0.5173, "step": 945 }, { "epoch": 0.27262247838616716, "grad_norm": 5.141915343195708, "learning_rate": 1e-05, "loss": 0.3404, "step": 946 }, { "epoch": 0.2729106628242075, "grad_norm": 2.1772464536272325, "learning_rate": 1e-05, "loss": 0.502, "step": 947 }, { "epoch": 0.27319884726224786, "grad_norm": 1.4401664870261714, "learning_rate": 1e-05, "loss": 0.5154, "step": 948 }, { "epoch": 0.2734870317002882, "grad_norm": 3.0831382894278683, "learning_rate": 1e-05, "loss": 0.3639, "step": 949 }, { "epoch": 0.2737752161383285, "grad_norm": 2.6477007846939737, "learning_rate": 1e-05, "loss": 0.4888, "step": 950 }, { "epoch": 0.27406340057636885, "grad_norm": 2.0297943635857134, "learning_rate": 1e-05, "loss": 0.5256, "step": 951 }, { "epoch": 0.2743515850144092, "grad_norm": 1.8996304186524549, "learning_rate": 1e-05, "loss": 0.3546, "step": 952 }, { "epoch": 0.27463976945244956, "grad_norm": 2.7855096498536134, "learning_rate": 1e-05, "loss": 0.5055, "step": 953 }, { "epoch": 0.2749279538904899, "grad_norm": 1.5800546792173475, "learning_rate": 1e-05, "loss": 0.5026, "step": 954 }, { "epoch": 0.27521613832853026, "grad_norm": 5.7576587032094455, "learning_rate": 1e-05, "loss": 0.3315, "step": 955 }, { "epoch": 0.2755043227665706, "grad_norm": 4.201893465596969, "learning_rate": 1e-05, "loss": 0.5388, "step": 956 }, { "epoch": 0.27579250720461096, "grad_norm": 1.4216292133657598, "learning_rate": 1e-05, "loss": 0.5087, "step": 957 }, { "epoch": 0.2760806916426513, "grad_norm": 2.226321484319974, "learning_rate": 1e-05, "loss": 0.3545, "step": 958 }, { "epoch": 0.27636887608069166, "grad_norm": 3.313341371873014, "learning_rate": 1e-05, "loss": 0.5253, "step": 959 }, { "epoch": 0.276657060518732, "grad_norm": 1.6604053305299187, "learning_rate": 1e-05, "loss": 0.5076, "step": 960 }, { "epoch": 0.2769452449567723, "grad_norm": 3.371233896844695, "learning_rate": 1e-05, "loss": 0.3361, "step": 961 }, { "epoch": 0.27723342939481266, "grad_norm": 3.2121544076481245, "learning_rate": 1e-05, "loss": 0.5218, "step": 962 }, { "epoch": 0.277521613832853, "grad_norm": 1.5130474234128468, "learning_rate": 1e-05, "loss": 0.5112, "step": 963 }, { "epoch": 0.27780979827089336, "grad_norm": 2.5496314917510325, "learning_rate": 1e-05, "loss": 0.3398, "step": 964 }, { "epoch": 0.2780979827089337, "grad_norm": 3.987921686601577, "learning_rate": 1e-05, "loss": 0.5268, "step": 965 }, { "epoch": 0.27838616714697406, "grad_norm": 2.2533295714840587, "learning_rate": 1e-05, "loss": 0.5015, "step": 966 }, { "epoch": 0.2786743515850144, "grad_norm": 1.5513596680091195, "learning_rate": 1e-05, "loss": 0.3512, "step": 967 }, { "epoch": 0.27896253602305476, "grad_norm": 2.013799560979713, "learning_rate": 1e-05, "loss": 0.5339, "step": 968 }, { "epoch": 0.2792507204610951, "grad_norm": 1.7163970279679512, "learning_rate": 1e-05, "loss": 0.516, "step": 969 }, { "epoch": 0.27953890489913547, "grad_norm": 1.740456061909977, "learning_rate": 1e-05, "loss": 0.3458, "step": 970 }, { "epoch": 0.2798270893371758, "grad_norm": 2.600900511052735, "learning_rate": 1e-05, "loss": 0.5783, "step": 971 }, { "epoch": 0.2801152737752161, "grad_norm": 1.6952006028519035, "learning_rate": 1e-05, "loss": 0.5113, "step": 972 }, { "epoch": 0.28040345821325646, "grad_norm": 1.6348837324966576, "learning_rate": 1e-05, "loss": 0.3416, "step": 973 }, { "epoch": 0.2806916426512968, "grad_norm": 3.203856767932054, "learning_rate": 1e-05, "loss": 0.4982, "step": 974 }, { "epoch": 0.28097982708933716, "grad_norm": 1.4188826070149956, "learning_rate": 1e-05, "loss": 0.4952, "step": 975 }, { "epoch": 0.2812680115273775, "grad_norm": 1.5009416983479824, "learning_rate": 1e-05, "loss": 0.3347, "step": 976 }, { "epoch": 0.28155619596541787, "grad_norm": 2.277957231446711, "learning_rate": 1e-05, "loss": 0.4995, "step": 977 }, { "epoch": 0.2818443804034582, "grad_norm": 1.3294390252675674, "learning_rate": 1e-05, "loss": 0.5116, "step": 978 }, { "epoch": 0.28213256484149857, "grad_norm": 1.4958690683527438, "learning_rate": 1e-05, "loss": 0.3253, "step": 979 }, { "epoch": 0.2824207492795389, "grad_norm": 2.4378365852311936, "learning_rate": 1e-05, "loss": 0.4753, "step": 980 }, { "epoch": 0.28270893371757927, "grad_norm": 1.6091312154798967, "learning_rate": 1e-05, "loss": 0.5081, "step": 981 }, { "epoch": 0.2829971181556196, "grad_norm": 2.1393277030619946, "learning_rate": 1e-05, "loss": 0.3604, "step": 982 }, { "epoch": 0.2832853025936599, "grad_norm": 2.1222333160060027, "learning_rate": 1e-05, "loss": 0.4901, "step": 983 }, { "epoch": 0.28357348703170027, "grad_norm": 1.4829404798184205, "learning_rate": 1e-05, "loss": 0.5092, "step": 984 }, { "epoch": 0.2838616714697406, "grad_norm": 1.757665907427271, "learning_rate": 1e-05, "loss": 0.3501, "step": 985 }, { "epoch": 0.28414985590778097, "grad_norm": 2.106202963340238, "learning_rate": 1e-05, "loss": 0.5245, "step": 986 }, { "epoch": 0.2844380403458213, "grad_norm": 1.4356411481647915, "learning_rate": 1e-05, "loss": 0.5126, "step": 987 }, { "epoch": 0.28472622478386167, "grad_norm": 1.8367029706409999, "learning_rate": 1e-05, "loss": 0.3274, "step": 988 }, { "epoch": 0.285014409221902, "grad_norm": 2.2418570564991716, "learning_rate": 1e-05, "loss": 0.5185, "step": 989 }, { "epoch": 0.28530259365994237, "grad_norm": 2.498112838150171, "learning_rate": 1e-05, "loss": 0.5071, "step": 990 }, { "epoch": 0.2855907780979827, "grad_norm": 2.4362164461254325, "learning_rate": 1e-05, "loss": 0.368, "step": 991 }, { "epoch": 0.2858789625360231, "grad_norm": 2.3850112864480444, "learning_rate": 1e-05, "loss": 0.5489, "step": 992 }, { "epoch": 0.2861671469740634, "grad_norm": 2.1723865818093855, "learning_rate": 1e-05, "loss": 0.4988, "step": 993 }, { "epoch": 0.2864553314121037, "grad_norm": 1.5193362742613035, "learning_rate": 1e-05, "loss": 0.349, "step": 994 }, { "epoch": 0.28674351585014407, "grad_norm": 2.3047313648767527, "learning_rate": 1e-05, "loss": 0.4986, "step": 995 }, { "epoch": 0.2870317002881844, "grad_norm": 1.849037996593631, "learning_rate": 1e-05, "loss": 0.5007, "step": 996 }, { "epoch": 0.2873198847262248, "grad_norm": 1.5160434534484937, "learning_rate": 1e-05, "loss": 0.3342, "step": 997 }, { "epoch": 0.2876080691642651, "grad_norm": 3.656074368240668, "learning_rate": 1e-05, "loss": 0.5135, "step": 998 }, { "epoch": 0.2878962536023055, "grad_norm": 1.7931777660285688, "learning_rate": 1e-05, "loss": 0.5173, "step": 999 }, { "epoch": 0.2881844380403458, "grad_norm": 1.7995325426446684, "learning_rate": 1e-05, "loss": 0.3335, "step": 1000 }, { "epoch": 0.2881844380403458, "eval_websight_new_IoU": 0.017799582332372665, "eval_websight_new_MAE_x": 286.7799987792969, "eval_websight_new_MAE_y": 626.0999755859375, "eval_websight_new_coord_probability": 0.7444506883621216, "eval_websight_new_correct_token_probability": 0.00132176058832556, "eval_websight_new_inside_bbox": 0.019999999552965164, "eval_websight_new_loss": 0.34697380661964417, "eval_websight_new_range_100_probability": 0.08497080951929092, "eval_websight_new_range_20_probability": 0.016971057280898094, "eval_websight_new_range_5_probability": 0.0037731751799583435, "eval_websight_new_runtime": 34.1093, "eval_websight_new_samples_per_second": 1.466, "eval_websight_new_steps_per_second": 0.205, "step": 1000 }, { "epoch": 0.2881844380403458, "eval_seeclick_IoU": 0.03177700936794281, "eval_seeclick_MAE_x": 473.1000061035156, "eval_seeclick_MAE_y": 420.239990234375, "eval_seeclick_coord_probability": 0.7443803548812866, "eval_seeclick_correct_token_probability": 0.0007717523258179426, "eval_seeclick_inside_bbox": 0.07999999821186066, "eval_seeclick_loss": 0.41178855299949646, "eval_seeclick_range_100_probability": 0.13013127446174622, "eval_seeclick_range_20_probability": 0.050893329083919525, "eval_seeclick_range_5_probability": 0.0074063134379684925, "eval_seeclick_runtime": 42.0879, "eval_seeclick_samples_per_second": 1.188, "eval_seeclick_steps_per_second": 0.166, "step": 1000 }, { "epoch": 0.2881844380403458, "eval_icons_IoU": 0.0025627599097788334, "eval_icons_MAE_x": 473.5400085449219, "eval_icons_MAE_y": 447.05999755859375, "eval_icons_coord_probability": 0.7488621473312378, "eval_icons_correct_token_probability": 0.0066015468910336494, "eval_icons_inside_bbox": 0.0, "eval_icons_loss": 0.4837815463542938, "eval_icons_range_100_probability": 0.2755989730358124, "eval_icons_range_20_probability": 0.10048653930425644, "eval_icons_range_5_probability": 0.0324726328253746, "eval_icons_runtime": 43.1872, "eval_icons_samples_per_second": 1.158, "eval_icons_steps_per_second": 0.162, "step": 1000 }, { "epoch": 0.2884726224783862, "grad_norm": 2.338705947652742, "learning_rate": 1e-05, "loss": 0.5499, "step": 1001 }, { "epoch": 0.2887608069164265, "grad_norm": 3.19564411545996, "learning_rate": 1e-05, "loss": 0.4956, "step": 1002 }, { "epoch": 0.2890489913544669, "grad_norm": 1.791379287140292, "learning_rate": 1e-05, "loss": 0.3211, "step": 1003 }, { "epoch": 0.28933717579250723, "grad_norm": 2.542255608805061, "learning_rate": 1e-05, "loss": 0.5143, "step": 1004 }, { "epoch": 0.2896253602305475, "grad_norm": 1.5656152479121874, "learning_rate": 1e-05, "loss": 0.503, "step": 1005 }, { "epoch": 0.2899135446685879, "grad_norm": 1.875361162771763, "learning_rate": 1e-05, "loss": 0.3259, "step": 1006 }, { "epoch": 0.2902017291066282, "grad_norm": 2.5654391299631047, "learning_rate": 1e-05, "loss": 0.4782, "step": 1007 }, { "epoch": 0.2904899135446686, "grad_norm": 1.687344227505042, "learning_rate": 1e-05, "loss": 0.5132, "step": 1008 }, { "epoch": 0.2907780979827089, "grad_norm": 1.4118840040804839, "learning_rate": 1e-05, "loss": 0.3425, "step": 1009 }, { "epoch": 0.2910662824207493, "grad_norm": 3.0684612023857323, "learning_rate": 1e-05, "loss": 0.4977, "step": 1010 }, { "epoch": 0.29135446685878963, "grad_norm": 2.3222928133076612, "learning_rate": 1e-05, "loss": 0.5029, "step": 1011 }, { "epoch": 0.29164265129683, "grad_norm": 1.4926787652225055, "learning_rate": 1e-05, "loss": 0.3512, "step": 1012 }, { "epoch": 0.29193083573487033, "grad_norm": 2.225007072873174, "learning_rate": 1e-05, "loss": 0.4875, "step": 1013 }, { "epoch": 0.2922190201729107, "grad_norm": 1.3432831467936621, "learning_rate": 1e-05, "loss": 0.5122, "step": 1014 }, { "epoch": 0.29250720461095103, "grad_norm": 1.4827838645606013, "learning_rate": 1e-05, "loss": 0.3436, "step": 1015 }, { "epoch": 0.29279538904899133, "grad_norm": 2.3026347865131322, "learning_rate": 1e-05, "loss": 0.553, "step": 1016 }, { "epoch": 0.2930835734870317, "grad_norm": 2.0832470548307267, "learning_rate": 1e-05, "loss": 0.5061, "step": 1017 }, { "epoch": 0.29337175792507203, "grad_norm": 1.5728771509894899, "learning_rate": 1e-05, "loss": 0.3495, "step": 1018 }, { "epoch": 0.2936599423631124, "grad_norm": 2.7257448830204827, "learning_rate": 1e-05, "loss": 0.5238, "step": 1019 }, { "epoch": 0.29394812680115273, "grad_norm": 2.2189118689950584, "learning_rate": 1e-05, "loss": 0.5084, "step": 1020 }, { "epoch": 0.2942363112391931, "grad_norm": 2.915582284166355, "learning_rate": 1e-05, "loss": 0.3676, "step": 1021 }, { "epoch": 0.29452449567723343, "grad_norm": 2.502896874980711, "learning_rate": 1e-05, "loss": 0.4961, "step": 1022 }, { "epoch": 0.2948126801152738, "grad_norm": 2.1008151251217253, "learning_rate": 1e-05, "loss": 0.5092, "step": 1023 }, { "epoch": 0.29510086455331414, "grad_norm": 1.710263389551021, "learning_rate": 1e-05, "loss": 0.3457, "step": 1024 }, { "epoch": 0.2953890489913545, "grad_norm": 3.556721941846693, "learning_rate": 1e-05, "loss": 0.4959, "step": 1025 }, { "epoch": 0.29567723342939484, "grad_norm": 1.4887559265779642, "learning_rate": 1e-05, "loss": 0.5115, "step": 1026 }, { "epoch": 0.29596541786743513, "grad_norm": 1.8539460109675998, "learning_rate": 1e-05, "loss": 0.3516, "step": 1027 }, { "epoch": 0.2962536023054755, "grad_norm": 2.3425154740643555, "learning_rate": 1e-05, "loss": 0.4887, "step": 1028 }, { "epoch": 0.29654178674351583, "grad_norm": 1.1950489518835659, "learning_rate": 1e-05, "loss": 0.4877, "step": 1029 }, { "epoch": 0.2968299711815562, "grad_norm": 1.325524273952331, "learning_rate": 1e-05, "loss": 0.3462, "step": 1030 }, { "epoch": 0.29711815561959654, "grad_norm": 3.6218677604895317, "learning_rate": 1e-05, "loss": 0.5156, "step": 1031 }, { "epoch": 0.2974063400576369, "grad_norm": 1.3493498512154727, "learning_rate": 1e-05, "loss": 0.495, "step": 1032 }, { "epoch": 0.29769452449567724, "grad_norm": 2.143130694787277, "learning_rate": 1e-05, "loss": 0.3186, "step": 1033 }, { "epoch": 0.2979827089337176, "grad_norm": 2.1363154839009724, "learning_rate": 1e-05, "loss": 0.5064, "step": 1034 }, { "epoch": 0.29827089337175794, "grad_norm": 1.699553612367511, "learning_rate": 1e-05, "loss": 0.4945, "step": 1035 }, { "epoch": 0.2985590778097983, "grad_norm": 2.0868848043456336, "learning_rate": 1e-05, "loss": 0.3304, "step": 1036 }, { "epoch": 0.29884726224783864, "grad_norm": 2.2575186930196534, "learning_rate": 1e-05, "loss": 0.5199, "step": 1037 }, { "epoch": 0.29913544668587894, "grad_norm": 1.7017161362946382, "learning_rate": 1e-05, "loss": 0.5047, "step": 1038 }, { "epoch": 0.2994236311239193, "grad_norm": 2.37203629858315, "learning_rate": 1e-05, "loss": 0.3492, "step": 1039 }, { "epoch": 0.29971181556195964, "grad_norm": 2.665105073295838, "learning_rate": 1e-05, "loss": 0.534, "step": 1040 }, { "epoch": 0.3, "grad_norm": 2.048127299268337, "learning_rate": 1e-05, "loss": 0.4985, "step": 1041 }, { "epoch": 0.30028818443804034, "grad_norm": 2.546255498986328, "learning_rate": 1e-05, "loss": 0.3369, "step": 1042 }, { "epoch": 0.3005763688760807, "grad_norm": 3.327265663756822, "learning_rate": 1e-05, "loss": 0.4978, "step": 1043 }, { "epoch": 0.30086455331412104, "grad_norm": 1.5376186167854946, "learning_rate": 1e-05, "loss": 0.5036, "step": 1044 }, { "epoch": 0.3011527377521614, "grad_norm": 2.326408967350416, "learning_rate": 1e-05, "loss": 0.3184, "step": 1045 }, { "epoch": 0.30144092219020174, "grad_norm": 2.327117884034042, "learning_rate": 1e-05, "loss": 0.4875, "step": 1046 }, { "epoch": 0.3017291066282421, "grad_norm": 1.8993417820337035, "learning_rate": 1e-05, "loss": 0.5083, "step": 1047 }, { "epoch": 0.30201729106628245, "grad_norm": 1.986867371144744, "learning_rate": 1e-05, "loss": 0.3337, "step": 1048 }, { "epoch": 0.30230547550432274, "grad_norm": 2.468355040160504, "learning_rate": 1e-05, "loss": 0.5177, "step": 1049 }, { "epoch": 0.3025936599423631, "grad_norm": 1.3565315890460994, "learning_rate": 1e-05, "loss": 0.4926, "step": 1050 }, { "epoch": 0.30288184438040344, "grad_norm": 3.4523937722407285, "learning_rate": 1e-05, "loss": 0.3439, "step": 1051 }, { "epoch": 0.3031700288184438, "grad_norm": 2.843193735309629, "learning_rate": 1e-05, "loss": 0.4768, "step": 1052 }, { "epoch": 0.30345821325648414, "grad_norm": 1.2883347670704832, "learning_rate": 1e-05, "loss": 0.5009, "step": 1053 }, { "epoch": 0.3037463976945245, "grad_norm": 1.5757669538464156, "learning_rate": 1e-05, "loss": 0.3412, "step": 1054 }, { "epoch": 0.30403458213256485, "grad_norm": 2.2483605576330694, "learning_rate": 1e-05, "loss": 0.4609, "step": 1055 }, { "epoch": 0.3043227665706052, "grad_norm": 1.737514661382086, "learning_rate": 1e-05, "loss": 0.4975, "step": 1056 }, { "epoch": 0.30461095100864555, "grad_norm": 2.941418614338563, "learning_rate": 1e-05, "loss": 0.3121, "step": 1057 }, { "epoch": 0.3048991354466859, "grad_norm": 3.4503887886225684, "learning_rate": 1e-05, "loss": 0.5021, "step": 1058 }, { "epoch": 0.30518731988472625, "grad_norm": 2.3020992021384186, "learning_rate": 1e-05, "loss": 0.4961, "step": 1059 }, { "epoch": 0.30547550432276654, "grad_norm": 1.606250972485056, "learning_rate": 1e-05, "loss": 0.3234, "step": 1060 }, { "epoch": 0.3057636887608069, "grad_norm": 3.204938244559325, "learning_rate": 1e-05, "loss": 0.5312, "step": 1061 }, { "epoch": 0.30605187319884725, "grad_norm": 1.7501247432174412, "learning_rate": 1e-05, "loss": 0.4954, "step": 1062 }, { "epoch": 0.3063400576368876, "grad_norm": 2.462337105028575, "learning_rate": 1e-05, "loss": 0.3362, "step": 1063 }, { "epoch": 0.30662824207492795, "grad_norm": 2.9218099597484413, "learning_rate": 1e-05, "loss": 0.4813, "step": 1064 }, { "epoch": 0.3069164265129683, "grad_norm": 1.3206159124427526, "learning_rate": 1e-05, "loss": 0.4995, "step": 1065 }, { "epoch": 0.30720461095100865, "grad_norm": 2.3711038543775724, "learning_rate": 1e-05, "loss": 0.3198, "step": 1066 }, { "epoch": 0.307492795389049, "grad_norm": 2.869326574113226, "learning_rate": 1e-05, "loss": 0.5007, "step": 1067 }, { "epoch": 0.30778097982708935, "grad_norm": 1.6101459790709531, "learning_rate": 1e-05, "loss": 0.4893, "step": 1068 }, { "epoch": 0.3080691642651297, "grad_norm": 2.0601913564631342, "learning_rate": 1e-05, "loss": 0.3208, "step": 1069 }, { "epoch": 0.30835734870317005, "grad_norm": 2.0557299549071004, "learning_rate": 1e-05, "loss": 0.4669, "step": 1070 }, { "epoch": 0.30864553314121035, "grad_norm": 1.6732954854682147, "learning_rate": 1e-05, "loss": 0.486, "step": 1071 }, { "epoch": 0.3089337175792507, "grad_norm": 1.6402686890449796, "learning_rate": 1e-05, "loss": 0.3418, "step": 1072 }, { "epoch": 0.30922190201729105, "grad_norm": 2.569262783943642, "learning_rate": 1e-05, "loss": 0.5016, "step": 1073 }, { "epoch": 0.3095100864553314, "grad_norm": 1.3435841034096272, "learning_rate": 1e-05, "loss": 0.4929, "step": 1074 }, { "epoch": 0.30979827089337175, "grad_norm": 1.9873277200581418, "learning_rate": 1e-05, "loss": 0.335, "step": 1075 }, { "epoch": 0.3100864553314121, "grad_norm": 3.0261184720776995, "learning_rate": 1e-05, "loss": 0.4929, "step": 1076 }, { "epoch": 0.31037463976945245, "grad_norm": 1.4833336101948864, "learning_rate": 1e-05, "loss": 0.4887, "step": 1077 }, { "epoch": 0.3106628242074928, "grad_norm": 1.6388746026047194, "learning_rate": 1e-05, "loss": 0.3443, "step": 1078 }, { "epoch": 0.31095100864553316, "grad_norm": 2.17100085821424, "learning_rate": 1e-05, "loss": 0.5004, "step": 1079 }, { "epoch": 0.3112391930835735, "grad_norm": 1.5009698236618427, "learning_rate": 1e-05, "loss": 0.5029, "step": 1080 }, { "epoch": 0.31152737752161386, "grad_norm": 1.3224191389643551, "learning_rate": 1e-05, "loss": 0.3305, "step": 1081 }, { "epoch": 0.31181556195965415, "grad_norm": 2.363374841564539, "learning_rate": 1e-05, "loss": 0.4831, "step": 1082 }, { "epoch": 0.3121037463976945, "grad_norm": 1.835170709153892, "learning_rate": 1e-05, "loss": 0.5032, "step": 1083 }, { "epoch": 0.31239193083573485, "grad_norm": 1.7927680679006808, "learning_rate": 1e-05, "loss": 0.3288, "step": 1084 }, { "epoch": 0.3126801152737752, "grad_norm": 2.442094038750008, "learning_rate": 1e-05, "loss": 0.4659, "step": 1085 }, { "epoch": 0.31296829971181556, "grad_norm": 1.2804687823000351, "learning_rate": 1e-05, "loss": 0.4878, "step": 1086 }, { "epoch": 0.3132564841498559, "grad_norm": 1.6619241894603738, "learning_rate": 1e-05, "loss": 0.3325, "step": 1087 }, { "epoch": 0.31354466858789626, "grad_norm": 2.6481437855016523, "learning_rate": 1e-05, "loss": 0.5162, "step": 1088 }, { "epoch": 0.3138328530259366, "grad_norm": 2.0292716630329117, "learning_rate": 1e-05, "loss": 0.5062, "step": 1089 }, { "epoch": 0.31412103746397696, "grad_norm": 1.6257203678204113, "learning_rate": 1e-05, "loss": 0.3118, "step": 1090 }, { "epoch": 0.3144092219020173, "grad_norm": 2.616220849824178, "learning_rate": 1e-05, "loss": 0.508, "step": 1091 }, { "epoch": 0.31469740634005766, "grad_norm": 2.431223788935733, "learning_rate": 1e-05, "loss": 0.5045, "step": 1092 }, { "epoch": 0.31498559077809796, "grad_norm": 1.6089488492295851, "learning_rate": 1e-05, "loss": 0.3272, "step": 1093 }, { "epoch": 0.3152737752161383, "grad_norm": 2.5294353517048087, "learning_rate": 1e-05, "loss": 0.4893, "step": 1094 }, { "epoch": 0.31556195965417866, "grad_norm": 1.8357982310621475, "learning_rate": 1e-05, "loss": 0.5081, "step": 1095 }, { "epoch": 0.315850144092219, "grad_norm": 1.4577791992451414, "learning_rate": 1e-05, "loss": 0.3463, "step": 1096 }, { "epoch": 0.31613832853025936, "grad_norm": 2.384433196506135, "learning_rate": 1e-05, "loss": 0.5092, "step": 1097 }, { "epoch": 0.3164265129682997, "grad_norm": 2.123313500223575, "learning_rate": 1e-05, "loss": 0.4831, "step": 1098 }, { "epoch": 0.31671469740634006, "grad_norm": 1.6293099403350886, "learning_rate": 1e-05, "loss": 0.351, "step": 1099 }, { "epoch": 0.3170028818443804, "grad_norm": 2.472562235724969, "learning_rate": 1e-05, "loss": 0.5227, "step": 1100 }, { "epoch": 0.31729106628242076, "grad_norm": 1.6770690157243782, "learning_rate": 1e-05, "loss": 0.4898, "step": 1101 }, { "epoch": 0.3175792507204611, "grad_norm": 2.3972959029136014, "learning_rate": 1e-05, "loss": 0.317, "step": 1102 }, { "epoch": 0.31786743515850147, "grad_norm": 2.664672422405968, "learning_rate": 1e-05, "loss": 0.4909, "step": 1103 }, { "epoch": 0.31815561959654176, "grad_norm": 1.5753150829159668, "learning_rate": 1e-05, "loss": 0.4941, "step": 1104 }, { "epoch": 0.3184438040345821, "grad_norm": 1.933326489331015, "learning_rate": 1e-05, "loss": 0.3533, "step": 1105 }, { "epoch": 0.31873198847262246, "grad_norm": 2.420775407156173, "learning_rate": 1e-05, "loss": 0.4497, "step": 1106 }, { "epoch": 0.3190201729106628, "grad_norm": 1.7918076644983714, "learning_rate": 1e-05, "loss": 0.4898, "step": 1107 }, { "epoch": 0.31930835734870316, "grad_norm": 1.2780275917344892, "learning_rate": 1e-05, "loss": 0.3087, "step": 1108 }, { "epoch": 0.3195965417867435, "grad_norm": 3.4097096363935027, "learning_rate": 1e-05, "loss": 0.4985, "step": 1109 }, { "epoch": 0.31988472622478387, "grad_norm": 1.5942865708242027, "learning_rate": 1e-05, "loss": 0.4861, "step": 1110 }, { "epoch": 0.3201729106628242, "grad_norm": 2.932396606270518, "learning_rate": 1e-05, "loss": 0.3253, "step": 1111 }, { "epoch": 0.32046109510086457, "grad_norm": 1.976735231620706, "learning_rate": 1e-05, "loss": 0.4609, "step": 1112 }, { "epoch": 0.3207492795389049, "grad_norm": 2.756056048048722, "learning_rate": 1e-05, "loss": 0.4999, "step": 1113 }, { "epoch": 0.32103746397694527, "grad_norm": 1.7328647439434661, "learning_rate": 1e-05, "loss": 0.3196, "step": 1114 }, { "epoch": 0.32132564841498557, "grad_norm": 2.262601047104278, "learning_rate": 1e-05, "loss": 0.5254, "step": 1115 }, { "epoch": 0.3216138328530259, "grad_norm": 1.6506093692491353, "learning_rate": 1e-05, "loss": 0.4981, "step": 1116 }, { "epoch": 0.32190201729106627, "grad_norm": 3.2703496210059018, "learning_rate": 1e-05, "loss": 0.3375, "step": 1117 }, { "epoch": 0.3221902017291066, "grad_norm": 4.868071164616914, "learning_rate": 1e-05, "loss": 0.5018, "step": 1118 }, { "epoch": 0.32247838616714697, "grad_norm": 1.6085911689626369, "learning_rate": 1e-05, "loss": 0.4998, "step": 1119 }, { "epoch": 0.3227665706051873, "grad_norm": 3.4486472045757277, "learning_rate": 1e-05, "loss": 0.3183, "step": 1120 }, { "epoch": 0.32305475504322767, "grad_norm": 2.6922668896039514, "learning_rate": 1e-05, "loss": 0.498, "step": 1121 }, { "epoch": 0.323342939481268, "grad_norm": 1.6744870417845208, "learning_rate": 1e-05, "loss": 0.4932, "step": 1122 }, { "epoch": 0.32363112391930837, "grad_norm": 2.323000645504488, "learning_rate": 1e-05, "loss": 0.3225, "step": 1123 }, { "epoch": 0.3239193083573487, "grad_norm": 3.6972537266210423, "learning_rate": 1e-05, "loss": 0.4985, "step": 1124 }, { "epoch": 0.3242074927953891, "grad_norm": 3.1444825583795826, "learning_rate": 1e-05, "loss": 0.4948, "step": 1125 }, { "epoch": 0.32449567723342937, "grad_norm": 2.5453979516419074, "learning_rate": 1e-05, "loss": 0.3321, "step": 1126 }, { "epoch": 0.3247838616714697, "grad_norm": 2.2607852092163196, "learning_rate": 1e-05, "loss": 0.4742, "step": 1127 }, { "epoch": 0.32507204610951007, "grad_norm": 2.1275514433154257, "learning_rate": 1e-05, "loss": 0.4905, "step": 1128 }, { "epoch": 0.3253602305475504, "grad_norm": 1.637432413303326, "learning_rate": 1e-05, "loss": 0.3423, "step": 1129 }, { "epoch": 0.3256484149855908, "grad_norm": 3.034997933911647, "learning_rate": 1e-05, "loss": 0.5107, "step": 1130 }, { "epoch": 0.3259365994236311, "grad_norm": 1.4552193149878974, "learning_rate": 1e-05, "loss": 0.4914, "step": 1131 }, { "epoch": 0.3262247838616715, "grad_norm": 4.1907660132530005, "learning_rate": 1e-05, "loss": 0.3091, "step": 1132 }, { "epoch": 0.3265129682997118, "grad_norm": 2.008765562856227, "learning_rate": 1e-05, "loss": 0.5149, "step": 1133 }, { "epoch": 0.3268011527377522, "grad_norm": 2.1363305914665207, "learning_rate": 1e-05, "loss": 0.4963, "step": 1134 }, { "epoch": 0.3270893371757925, "grad_norm": 1.9298203593679761, "learning_rate": 1e-05, "loss": 0.3166, "step": 1135 }, { "epoch": 0.3273775216138329, "grad_norm": 2.430956307266632, "learning_rate": 1e-05, "loss": 0.464, "step": 1136 }, { "epoch": 0.3276657060518732, "grad_norm": 2.2524622025954537, "learning_rate": 1e-05, "loss": 0.4901, "step": 1137 }, { "epoch": 0.3279538904899135, "grad_norm": 2.688017518403025, "learning_rate": 1e-05, "loss": 0.3251, "step": 1138 }, { "epoch": 0.3282420749279539, "grad_norm": 2.0325657283588567, "learning_rate": 1e-05, "loss": 0.5062, "step": 1139 }, { "epoch": 0.3285302593659942, "grad_norm": 1.7806061781593912, "learning_rate": 1e-05, "loss": 0.4768, "step": 1140 }, { "epoch": 0.3288184438040346, "grad_norm": 2.156722268754788, "learning_rate": 1e-05, "loss": 0.3264, "step": 1141 }, { "epoch": 0.32910662824207493, "grad_norm": 2.7039124237656016, "learning_rate": 1e-05, "loss": 0.5116, "step": 1142 }, { "epoch": 0.3293948126801153, "grad_norm": 1.846803783065928, "learning_rate": 1e-05, "loss": 0.4927, "step": 1143 }, { "epoch": 0.32968299711815563, "grad_norm": 1.9812623456518241, "learning_rate": 1e-05, "loss": 0.3019, "step": 1144 }, { "epoch": 0.329971181556196, "grad_norm": 2.6877800273170434, "learning_rate": 1e-05, "loss": 0.5018, "step": 1145 }, { "epoch": 0.33025936599423633, "grad_norm": 1.8582044690750283, "learning_rate": 1e-05, "loss": 0.503, "step": 1146 }, { "epoch": 0.3305475504322767, "grad_norm": 2.5289453697355757, "learning_rate": 1e-05, "loss": 0.3375, "step": 1147 }, { "epoch": 0.330835734870317, "grad_norm": 3.94007933818287, "learning_rate": 1e-05, "loss": 0.4854, "step": 1148 }, { "epoch": 0.33112391930835733, "grad_norm": 2.3006899651143424, "learning_rate": 1e-05, "loss": 0.4887, "step": 1149 }, { "epoch": 0.3314121037463977, "grad_norm": 1.9633104945925637, "learning_rate": 1e-05, "loss": 0.334, "step": 1150 }, { "epoch": 0.33170028818443803, "grad_norm": 2.523047665002316, "learning_rate": 1e-05, "loss": 0.5205, "step": 1151 }, { "epoch": 0.3319884726224784, "grad_norm": 1.73966685174154, "learning_rate": 1e-05, "loss": 0.4991, "step": 1152 }, { "epoch": 0.33227665706051873, "grad_norm": 1.8066746398505633, "learning_rate": 1e-05, "loss": 0.3437, "step": 1153 }, { "epoch": 0.3325648414985591, "grad_norm": 2.283090155831804, "learning_rate": 1e-05, "loss": 0.5052, "step": 1154 }, { "epoch": 0.33285302593659943, "grad_norm": 1.5571410132639631, "learning_rate": 1e-05, "loss": 0.4896, "step": 1155 }, { "epoch": 0.3331412103746398, "grad_norm": 1.3363582344794764, "learning_rate": 1e-05, "loss": 0.3343, "step": 1156 }, { "epoch": 0.33342939481268014, "grad_norm": 2.992721080077841, "learning_rate": 1e-05, "loss": 0.4711, "step": 1157 }, { "epoch": 0.3337175792507205, "grad_norm": 1.816968349529476, "learning_rate": 1e-05, "loss": 0.4921, "step": 1158 }, { "epoch": 0.3340057636887608, "grad_norm": 1.6857815219001941, "learning_rate": 1e-05, "loss": 0.3426, "step": 1159 }, { "epoch": 0.33429394812680113, "grad_norm": 3.0959603364657204, "learning_rate": 1e-05, "loss": 0.513, "step": 1160 }, { "epoch": 0.3345821325648415, "grad_norm": 1.4472170523508918, "learning_rate": 1e-05, "loss": 0.4973, "step": 1161 }, { "epoch": 0.33487031700288183, "grad_norm": 2.283479435772585, "learning_rate": 1e-05, "loss": 0.3309, "step": 1162 }, { "epoch": 0.3351585014409222, "grad_norm": 2.50287615214733, "learning_rate": 1e-05, "loss": 0.493, "step": 1163 }, { "epoch": 0.33544668587896254, "grad_norm": 2.1825900702502294, "learning_rate": 1e-05, "loss": 0.4941, "step": 1164 }, { "epoch": 0.3357348703170029, "grad_norm": 1.977525837420698, "learning_rate": 1e-05, "loss": 0.3375, "step": 1165 }, { "epoch": 0.33602305475504324, "grad_norm": 2.6235561088202357, "learning_rate": 1e-05, "loss": 0.4653, "step": 1166 }, { "epoch": 0.3363112391930836, "grad_norm": 1.8806137088648773, "learning_rate": 1e-05, "loss": 0.4834, "step": 1167 }, { "epoch": 0.33659942363112394, "grad_norm": 2.076944516350051, "learning_rate": 1e-05, "loss": 0.3133, "step": 1168 }, { "epoch": 0.3368876080691643, "grad_norm": 2.830054702700364, "learning_rate": 1e-05, "loss": 0.5036, "step": 1169 }, { "epoch": 0.3371757925072046, "grad_norm": 1.8246097676492032, "learning_rate": 1e-05, "loss": 0.4886, "step": 1170 }, { "epoch": 0.33746397694524494, "grad_norm": 1.9362712545579375, "learning_rate": 1e-05, "loss": 0.3026, "step": 1171 }, { "epoch": 0.3377521613832853, "grad_norm": 2.4855666695575493, "learning_rate": 1e-05, "loss": 0.4988, "step": 1172 }, { "epoch": 0.33804034582132564, "grad_norm": 1.6627595920456808, "learning_rate": 1e-05, "loss": 0.492, "step": 1173 }, { "epoch": 0.338328530259366, "grad_norm": 1.9276841704652814, "learning_rate": 1e-05, "loss": 0.3227, "step": 1174 }, { "epoch": 0.33861671469740634, "grad_norm": 2.576002040137353, "learning_rate": 1e-05, "loss": 0.5259, "step": 1175 }, { "epoch": 0.3389048991354467, "grad_norm": 1.9066436871077268, "learning_rate": 1e-05, "loss": 0.4944, "step": 1176 }, { "epoch": 0.33919308357348704, "grad_norm": 1.5441205950825012, "learning_rate": 1e-05, "loss": 0.337, "step": 1177 }, { "epoch": 0.3394812680115274, "grad_norm": 3.2639640612206944, "learning_rate": 1e-05, "loss": 0.4634, "step": 1178 }, { "epoch": 0.33976945244956774, "grad_norm": 2.258471482981334, "learning_rate": 1e-05, "loss": 0.5032, "step": 1179 }, { "epoch": 0.3400576368876081, "grad_norm": 2.008940650731972, "learning_rate": 1e-05, "loss": 0.3102, "step": 1180 }, { "epoch": 0.3403458213256484, "grad_norm": 3.0963746293113514, "learning_rate": 1e-05, "loss": 0.5367, "step": 1181 }, { "epoch": 0.34063400576368874, "grad_norm": 1.915204271376301, "learning_rate": 1e-05, "loss": 0.4929, "step": 1182 }, { "epoch": 0.3409221902017291, "grad_norm": 2.286096240621281, "learning_rate": 1e-05, "loss": 0.3084, "step": 1183 }, { "epoch": 0.34121037463976944, "grad_norm": 2.6863781439408223, "learning_rate": 1e-05, "loss": 0.4711, "step": 1184 }, { "epoch": 0.3414985590778098, "grad_norm": 1.8449861890807961, "learning_rate": 1e-05, "loss": 0.4948, "step": 1185 }, { "epoch": 0.34178674351585014, "grad_norm": 2.7888935869922826, "learning_rate": 1e-05, "loss": 0.3196, "step": 1186 }, { "epoch": 0.3420749279538905, "grad_norm": 2.709673365782931, "learning_rate": 1e-05, "loss": 0.5203, "step": 1187 }, { "epoch": 0.34236311239193085, "grad_norm": 1.7317494751780376, "learning_rate": 1e-05, "loss": 0.4959, "step": 1188 }, { "epoch": 0.3426512968299712, "grad_norm": 2.1063969485892895, "learning_rate": 1e-05, "loss": 0.3123, "step": 1189 }, { "epoch": 0.34293948126801155, "grad_norm": 2.2464240545060257, "learning_rate": 1e-05, "loss": 0.5109, "step": 1190 }, { "epoch": 0.3432276657060519, "grad_norm": 2.4253077947170296, "learning_rate": 1e-05, "loss": 0.4969, "step": 1191 }, { "epoch": 0.3435158501440922, "grad_norm": 2.4332739078321004, "learning_rate": 1e-05, "loss": 0.334, "step": 1192 }, { "epoch": 0.34380403458213254, "grad_norm": 3.482343282441452, "learning_rate": 1e-05, "loss": 0.4952, "step": 1193 }, { "epoch": 0.3440922190201729, "grad_norm": 2.38694438541596, "learning_rate": 1e-05, "loss": 0.485, "step": 1194 }, { "epoch": 0.34438040345821325, "grad_norm": 2.410215883752191, "learning_rate": 1e-05, "loss": 0.3394, "step": 1195 }, { "epoch": 0.3446685878962536, "grad_norm": 2.3249140368948633, "learning_rate": 1e-05, "loss": 0.5061, "step": 1196 }, { "epoch": 0.34495677233429395, "grad_norm": 1.8467063337590413, "learning_rate": 1e-05, "loss": 0.4908, "step": 1197 }, { "epoch": 0.3452449567723343, "grad_norm": 1.762787342750634, "learning_rate": 1e-05, "loss": 0.3195, "step": 1198 }, { "epoch": 0.34553314121037465, "grad_norm": 2.3874762172329436, "learning_rate": 1e-05, "loss": 0.4993, "step": 1199 }, { "epoch": 0.345821325648415, "grad_norm": 2.8834715847733507, "learning_rate": 1e-05, "loss": 0.4904, "step": 1200 }, { "epoch": 0.34610951008645535, "grad_norm": 2.3666481840738345, "learning_rate": 1e-05, "loss": 0.3251, "step": 1201 }, { "epoch": 0.3463976945244957, "grad_norm": 3.6847967110324222, "learning_rate": 1e-05, "loss": 0.5112, "step": 1202 }, { "epoch": 0.346685878962536, "grad_norm": 1.5959837004659907, "learning_rate": 1e-05, "loss": 0.4939, "step": 1203 }, { "epoch": 0.34697406340057635, "grad_norm": 2.1624166756889753, "learning_rate": 1e-05, "loss": 0.3097, "step": 1204 }, { "epoch": 0.3472622478386167, "grad_norm": 2.3874538395666125, "learning_rate": 1e-05, "loss": 0.4892, "step": 1205 }, { "epoch": 0.34755043227665705, "grad_norm": 2.2499510745749514, "learning_rate": 1e-05, "loss": 0.4915, "step": 1206 }, { "epoch": 0.3478386167146974, "grad_norm": 1.404668578845065, "learning_rate": 1e-05, "loss": 0.3006, "step": 1207 }, { "epoch": 0.34812680115273775, "grad_norm": 2.7956025858497378, "learning_rate": 1e-05, "loss": 0.4777, "step": 1208 }, { "epoch": 0.3484149855907781, "grad_norm": 2.379588347293559, "learning_rate": 1e-05, "loss": 0.4961, "step": 1209 }, { "epoch": 0.34870317002881845, "grad_norm": 2.0538953562342837, "learning_rate": 1e-05, "loss": 0.3216, "step": 1210 }, { "epoch": 0.3489913544668588, "grad_norm": 2.355086879363187, "learning_rate": 1e-05, "loss": 0.4709, "step": 1211 }, { "epoch": 0.34927953890489916, "grad_norm": 2.073100433132766, "learning_rate": 1e-05, "loss": 0.4815, "step": 1212 }, { "epoch": 0.3495677233429395, "grad_norm": 1.673457990794969, "learning_rate": 1e-05, "loss": 0.3215, "step": 1213 }, { "epoch": 0.3498559077809798, "grad_norm": 2.6904057157051153, "learning_rate": 1e-05, "loss": 0.491, "step": 1214 }, { "epoch": 0.35014409221902015, "grad_norm": 1.6598497495637436, "learning_rate": 1e-05, "loss": 0.4948, "step": 1215 }, { "epoch": 0.3504322766570605, "grad_norm": 1.6461641354690129, "learning_rate": 1e-05, "loss": 0.3062, "step": 1216 }, { "epoch": 0.35072046109510085, "grad_norm": 2.972128860288283, "learning_rate": 1e-05, "loss": 0.4939, "step": 1217 }, { "epoch": 0.3510086455331412, "grad_norm": 1.9455660845083287, "learning_rate": 1e-05, "loss": 0.4873, "step": 1218 }, { "epoch": 0.35129682997118156, "grad_norm": 2.7234024735995908, "learning_rate": 1e-05, "loss": 0.337, "step": 1219 }, { "epoch": 0.3515850144092219, "grad_norm": 2.8288095617482347, "learning_rate": 1e-05, "loss": 0.4854, "step": 1220 }, { "epoch": 0.35187319884726226, "grad_norm": 1.4670768600406208, "learning_rate": 1e-05, "loss": 0.4928, "step": 1221 }, { "epoch": 0.3521613832853026, "grad_norm": 2.138084040796824, "learning_rate": 1e-05, "loss": 0.3109, "step": 1222 }, { "epoch": 0.35244956772334296, "grad_norm": 3.625345475281414, "learning_rate": 1e-05, "loss": 0.5104, "step": 1223 }, { "epoch": 0.3527377521613833, "grad_norm": 1.5645525491115844, "learning_rate": 1e-05, "loss": 0.4739, "step": 1224 }, { "epoch": 0.3530259365994236, "grad_norm": 1.388550786833075, "learning_rate": 1e-05, "loss": 0.3304, "step": 1225 }, { "epoch": 0.35331412103746396, "grad_norm": 2.4455139912586077, "learning_rate": 1e-05, "loss": 0.4635, "step": 1226 }, { "epoch": 0.3536023054755043, "grad_norm": 1.865557352159103, "learning_rate": 1e-05, "loss": 0.488, "step": 1227 }, { "epoch": 0.35389048991354466, "grad_norm": 2.5138800299778428, "learning_rate": 1e-05, "loss": 0.3237, "step": 1228 }, { "epoch": 0.354178674351585, "grad_norm": 2.7204347633921686, "learning_rate": 1e-05, "loss": 0.4777, "step": 1229 }, { "epoch": 0.35446685878962536, "grad_norm": 1.8944054861645894, "learning_rate": 1e-05, "loss": 0.4904, "step": 1230 }, { "epoch": 0.3547550432276657, "grad_norm": 1.7954201368104827, "learning_rate": 1e-05, "loss": 0.328, "step": 1231 }, { "epoch": 0.35504322766570606, "grad_norm": 2.920565212870612, "learning_rate": 1e-05, "loss": 0.464, "step": 1232 }, { "epoch": 0.3553314121037464, "grad_norm": 2.724892083275897, "learning_rate": 1e-05, "loss": 0.5022, "step": 1233 }, { "epoch": 0.35561959654178676, "grad_norm": 1.3347016429534064, "learning_rate": 1e-05, "loss": 0.2885, "step": 1234 }, { "epoch": 0.3559077809798271, "grad_norm": 3.2167881848599342, "learning_rate": 1e-05, "loss": 0.4999, "step": 1235 }, { "epoch": 0.3561959654178674, "grad_norm": 1.87428169441781, "learning_rate": 1e-05, "loss": 0.4836, "step": 1236 }, { "epoch": 0.35648414985590776, "grad_norm": 2.8143313945360267, "learning_rate": 1e-05, "loss": 0.3185, "step": 1237 }, { "epoch": 0.3567723342939481, "grad_norm": 2.3846807789170965, "learning_rate": 1e-05, "loss": 0.4839, "step": 1238 }, { "epoch": 0.35706051873198846, "grad_norm": 2.035876155842584, "learning_rate": 1e-05, "loss": 0.4875, "step": 1239 }, { "epoch": 0.3573487031700288, "grad_norm": 1.9676162943433577, "learning_rate": 1e-05, "loss": 0.3067, "step": 1240 }, { "epoch": 0.35763688760806917, "grad_norm": 3.1917087666871975, "learning_rate": 1e-05, "loss": 0.4951, "step": 1241 }, { "epoch": 0.3579250720461095, "grad_norm": 1.6257615101345382, "learning_rate": 1e-05, "loss": 0.4707, "step": 1242 }, { "epoch": 0.35821325648414987, "grad_norm": 2.759397431881197, "learning_rate": 1e-05, "loss": 0.3148, "step": 1243 }, { "epoch": 0.3585014409221902, "grad_norm": 2.6349765334068067, "learning_rate": 1e-05, "loss": 0.4737, "step": 1244 }, { "epoch": 0.35878962536023057, "grad_norm": 1.5678831740735477, "learning_rate": 1e-05, "loss": 0.4899, "step": 1245 }, { "epoch": 0.3590778097982709, "grad_norm": 1.6145035667601524, "learning_rate": 1e-05, "loss": 0.302, "step": 1246 }, { "epoch": 0.3593659942363112, "grad_norm": 2.9853048113082368, "learning_rate": 1e-05, "loss": 0.5149, "step": 1247 }, { "epoch": 0.35965417867435157, "grad_norm": 1.357609329275561, "learning_rate": 1e-05, "loss": 0.4912, "step": 1248 }, { "epoch": 0.3599423631123919, "grad_norm": 2.121872333122013, "learning_rate": 1e-05, "loss": 0.322, "step": 1249 }, { "epoch": 0.36023054755043227, "grad_norm": 2.145641816572467, "learning_rate": 1e-05, "loss": 0.4628, "step": 1250 }, { "epoch": 0.36023054755043227, "eval_websight_new_IoU": 0.017749059945344925, "eval_websight_new_MAE_x": 286.7799987792969, "eval_websight_new_MAE_y": 621.280029296875, "eval_websight_new_coord_probability": 0.7440074682235718, "eval_websight_new_correct_token_probability": 0.0020548468455672264, "eval_websight_new_inside_bbox": 0.03999999910593033, "eval_websight_new_loss": 0.3225766122341156, "eval_websight_new_range_100_probability": 0.09060951322317123, "eval_websight_new_range_20_probability": 0.016936045140028, "eval_websight_new_range_5_probability": 0.004216435365378857, "eval_websight_new_runtime": 34.1266, "eval_websight_new_samples_per_second": 1.465, "eval_websight_new_steps_per_second": 0.205, "step": 1250 }, { "epoch": 0.36023054755043227, "eval_seeclick_IoU": 0.03234250470995903, "eval_seeclick_MAE_x": 473.1000061035156, "eval_seeclick_MAE_y": 587.97998046875, "eval_seeclick_coord_probability": 0.7468330264091492, "eval_seeclick_correct_token_probability": 0.0008442556136287749, "eval_seeclick_inside_bbox": 0.07999999821186066, "eval_seeclick_loss": 0.39571070671081543, "eval_seeclick_range_100_probability": 0.12909740209579468, "eval_seeclick_range_20_probability": 0.0456046499311924, "eval_seeclick_range_5_probability": 0.00795473251491785, "eval_seeclick_runtime": 43.2477, "eval_seeclick_samples_per_second": 1.156, "eval_seeclick_steps_per_second": 0.162, "step": 1250 }, { "epoch": 0.36023054755043227, "eval_icons_IoU": 0.003231856506317854, "eval_icons_MAE_x": 473.5400085449219, "eval_icons_MAE_y": 375.55999755859375, "eval_icons_coord_probability": 0.7471686601638794, "eval_icons_correct_token_probability": 0.0060055917128920555, "eval_icons_inside_bbox": 0.0, "eval_icons_loss": 0.4698539972305298, "eval_icons_range_100_probability": 0.27581825852394104, "eval_icons_range_20_probability": 0.10117445886135101, "eval_icons_range_5_probability": 0.03151170164346695, "eval_icons_runtime": 42.8179, "eval_icons_samples_per_second": 1.168, "eval_icons_steps_per_second": 0.163, "step": 1250 }, { "epoch": 0.3605187319884726, "grad_norm": 2.9267574931825977, "learning_rate": 1e-05, "loss": 0.4939, "step": 1251 }, { "epoch": 0.36080691642651297, "grad_norm": 1.4161651526852423, "learning_rate": 1e-05, "loss": 0.3099, "step": 1252 }, { "epoch": 0.3610951008645533, "grad_norm": 2.9329296516871315, "learning_rate": 1e-05, "loss": 0.4942, "step": 1253 }, { "epoch": 0.36138328530259367, "grad_norm": 1.7443290807635572, "learning_rate": 1e-05, "loss": 0.4993, "step": 1254 }, { "epoch": 0.361671469740634, "grad_norm": 1.5849987171284718, "learning_rate": 1e-05, "loss": 0.3201, "step": 1255 }, { "epoch": 0.3619596541786744, "grad_norm": 3.388739534462795, "learning_rate": 1e-05, "loss": 0.4998, "step": 1256 }, { "epoch": 0.3622478386167147, "grad_norm": 2.7104139892986048, "learning_rate": 1e-05, "loss": 0.4875, "step": 1257 }, { "epoch": 0.362536023054755, "grad_norm": 2.159282660602506, "learning_rate": 1e-05, "loss": 0.3115, "step": 1258 }, { "epoch": 0.36282420749279537, "grad_norm": 2.5475589208738225, "learning_rate": 1e-05, "loss": 0.471, "step": 1259 }, { "epoch": 0.3631123919308357, "grad_norm": 1.7243894630942112, "learning_rate": 1e-05, "loss": 0.4764, "step": 1260 }, { "epoch": 0.36340057636887607, "grad_norm": 1.777580820937449, "learning_rate": 1e-05, "loss": 0.3162, "step": 1261 }, { "epoch": 0.3636887608069164, "grad_norm": 3.2518330570615053, "learning_rate": 1e-05, "loss": 0.4846, "step": 1262 }, { "epoch": 0.3639769452449568, "grad_norm": 2.302001017306, "learning_rate": 1e-05, "loss": 0.5045, "step": 1263 }, { "epoch": 0.3642651296829971, "grad_norm": 2.528883284010198, "learning_rate": 1e-05, "loss": 0.3067, "step": 1264 }, { "epoch": 0.3645533141210375, "grad_norm": 3.2737716372913637, "learning_rate": 1e-05, "loss": 0.5047, "step": 1265 }, { "epoch": 0.3648414985590778, "grad_norm": 1.8572245666074538, "learning_rate": 1e-05, "loss": 0.4921, "step": 1266 }, { "epoch": 0.3651296829971182, "grad_norm": 2.419883123172451, "learning_rate": 1e-05, "loss": 0.3014, "step": 1267 }, { "epoch": 0.3654178674351585, "grad_norm": 2.380087061074094, "learning_rate": 1e-05, "loss": 0.4939, "step": 1268 }, { "epoch": 0.3657060518731988, "grad_norm": 1.455830805212864, "learning_rate": 1e-05, "loss": 0.4874, "step": 1269 }, { "epoch": 0.3659942363112392, "grad_norm": 1.7106818638178742, "learning_rate": 1e-05, "loss": 0.3375, "step": 1270 }, { "epoch": 0.3662824207492795, "grad_norm": 2.938412142230153, "learning_rate": 1e-05, "loss": 0.5121, "step": 1271 }, { "epoch": 0.3665706051873199, "grad_norm": 1.944071545920773, "learning_rate": 1e-05, "loss": 0.4717, "step": 1272 }, { "epoch": 0.3668587896253602, "grad_norm": 2.127957365004338, "learning_rate": 1e-05, "loss": 0.3184, "step": 1273 }, { "epoch": 0.3671469740634006, "grad_norm": 2.570902478592638, "learning_rate": 1e-05, "loss": 0.4889, "step": 1274 }, { "epoch": 0.36743515850144093, "grad_norm": 1.657296767571961, "learning_rate": 1e-05, "loss": 0.4818, "step": 1275 }, { "epoch": 0.3677233429394813, "grad_norm": 1.830962402830421, "learning_rate": 1e-05, "loss": 0.307, "step": 1276 }, { "epoch": 0.36801152737752163, "grad_norm": 3.2433727313856098, "learning_rate": 1e-05, "loss": 0.5087, "step": 1277 }, { "epoch": 0.368299711815562, "grad_norm": 3.2623484670700353, "learning_rate": 1e-05, "loss": 0.4897, "step": 1278 }, { "epoch": 0.36858789625360233, "grad_norm": 1.8457997670091872, "learning_rate": 1e-05, "loss": 0.3077, "step": 1279 }, { "epoch": 0.3688760806916426, "grad_norm": 3.4046358478665413, "learning_rate": 1e-05, "loss": 0.5126, "step": 1280 }, { "epoch": 0.369164265129683, "grad_norm": 2.1078191717348997, "learning_rate": 1e-05, "loss": 0.4961, "step": 1281 }, { "epoch": 0.36945244956772333, "grad_norm": 2.6760556258353914, "learning_rate": 1e-05, "loss": 0.2994, "step": 1282 }, { "epoch": 0.3697406340057637, "grad_norm": 2.7779267593627672, "learning_rate": 1e-05, "loss": 0.4815, "step": 1283 }, { "epoch": 0.37002881844380403, "grad_norm": 1.9673132430282592, "learning_rate": 1e-05, "loss": 0.4755, "step": 1284 }, { "epoch": 0.3703170028818444, "grad_norm": 1.6916980677277118, "learning_rate": 1e-05, "loss": 0.3235, "step": 1285 }, { "epoch": 0.37060518731988473, "grad_norm": 2.2892499234506247, "learning_rate": 1e-05, "loss": 0.4651, "step": 1286 }, { "epoch": 0.3708933717579251, "grad_norm": 1.9612395551381552, "learning_rate": 1e-05, "loss": 0.4843, "step": 1287 }, { "epoch": 0.37118155619596543, "grad_norm": 2.3321421597678538, "learning_rate": 1e-05, "loss": 0.3176, "step": 1288 }, { "epoch": 0.3714697406340058, "grad_norm": 2.775891285608071, "learning_rate": 1e-05, "loss": 0.4586, "step": 1289 }, { "epoch": 0.37175792507204614, "grad_norm": 1.5916999326429562, "learning_rate": 1e-05, "loss": 0.4813, "step": 1290 }, { "epoch": 0.37204610951008643, "grad_norm": 1.3546781839457918, "learning_rate": 1e-05, "loss": 0.3079, "step": 1291 }, { "epoch": 0.3723342939481268, "grad_norm": 2.034301394316897, "learning_rate": 1e-05, "loss": 0.4496, "step": 1292 }, { "epoch": 0.37262247838616713, "grad_norm": 2.5830915613085708, "learning_rate": 1e-05, "loss": 0.5008, "step": 1293 }, { "epoch": 0.3729106628242075, "grad_norm": 1.4480034078466113, "learning_rate": 1e-05, "loss": 0.3008, "step": 1294 }, { "epoch": 0.37319884726224783, "grad_norm": 2.6955945052568198, "learning_rate": 1e-05, "loss": 0.4906, "step": 1295 }, { "epoch": 0.3734870317002882, "grad_norm": 1.3768118074381839, "learning_rate": 1e-05, "loss": 0.4803, "step": 1296 }, { "epoch": 0.37377521613832854, "grad_norm": 2.2906236875197687, "learning_rate": 1e-05, "loss": 0.3053, "step": 1297 }, { "epoch": 0.3740634005763689, "grad_norm": 3.1020107631734337, "learning_rate": 1e-05, "loss": 0.5018, "step": 1298 }, { "epoch": 0.37435158501440924, "grad_norm": 3.2403181026251033, "learning_rate": 1e-05, "loss": 0.4919, "step": 1299 }, { "epoch": 0.3746397694524496, "grad_norm": 1.8184864154877236, "learning_rate": 1e-05, "loss": 0.3025, "step": 1300 }, { "epoch": 0.37492795389048994, "grad_norm": 2.117689221882779, "learning_rate": 1e-05, "loss": 0.4782, "step": 1301 }, { "epoch": 0.37521613832853024, "grad_norm": 2.611897522826078, "learning_rate": 1e-05, "loss": 0.4858, "step": 1302 }, { "epoch": 0.3755043227665706, "grad_norm": 1.6906029962405396, "learning_rate": 1e-05, "loss": 0.3183, "step": 1303 }, { "epoch": 0.37579250720461094, "grad_norm": 2.740084728664951, "learning_rate": 1e-05, "loss": 0.5032, "step": 1304 }, { "epoch": 0.3760806916426513, "grad_norm": 1.9411042343379972, "learning_rate": 1e-05, "loss": 0.4928, "step": 1305 }, { "epoch": 0.37636887608069164, "grad_norm": 2.422741433303496, "learning_rate": 1e-05, "loss": 0.3027, "step": 1306 }, { "epoch": 0.376657060518732, "grad_norm": 2.2839314934791886, "learning_rate": 1e-05, "loss": 0.4648, "step": 1307 }, { "epoch": 0.37694524495677234, "grad_norm": 2.115259153820208, "learning_rate": 1e-05, "loss": 0.4783, "step": 1308 }, { "epoch": 0.3772334293948127, "grad_norm": 2.1455241632791027, "learning_rate": 1e-05, "loss": 0.3167, "step": 1309 }, { "epoch": 0.37752161383285304, "grad_norm": 3.1279634221134094, "learning_rate": 1e-05, "loss": 0.4851, "step": 1310 }, { "epoch": 0.3778097982708934, "grad_norm": 1.9221366164413791, "learning_rate": 1e-05, "loss": 0.4933, "step": 1311 }, { "epoch": 0.37809798270893374, "grad_norm": 1.4491711126374, "learning_rate": 1e-05, "loss": 0.3365, "step": 1312 }, { "epoch": 0.37838616714697404, "grad_norm": 2.748317882630484, "learning_rate": 1e-05, "loss": 0.4923, "step": 1313 }, { "epoch": 0.3786743515850144, "grad_norm": 1.6488192337586374, "learning_rate": 1e-05, "loss": 0.4873, "step": 1314 }, { "epoch": 0.37896253602305474, "grad_norm": 1.8069141908922857, "learning_rate": 1e-05, "loss": 0.2988, "step": 1315 }, { "epoch": 0.3792507204610951, "grad_norm": 2.497952537050359, "learning_rate": 1e-05, "loss": 0.4943, "step": 1316 }, { "epoch": 0.37953890489913544, "grad_norm": 1.668255912182165, "learning_rate": 1e-05, "loss": 0.4814, "step": 1317 }, { "epoch": 0.3798270893371758, "grad_norm": 2.6220758634279675, "learning_rate": 1e-05, "loss": 0.2971, "step": 1318 }, { "epoch": 0.38011527377521614, "grad_norm": 3.064543845303531, "learning_rate": 1e-05, "loss": 0.4628, "step": 1319 }, { "epoch": 0.3804034582132565, "grad_norm": 2.1629034925763686, "learning_rate": 1e-05, "loss": 0.4831, "step": 1320 }, { "epoch": 0.38069164265129685, "grad_norm": 3.4293882979735906, "learning_rate": 1e-05, "loss": 0.3098, "step": 1321 }, { "epoch": 0.3809798270893372, "grad_norm": 3.153726299390492, "learning_rate": 1e-05, "loss": 0.498, "step": 1322 }, { "epoch": 0.38126801152737755, "grad_norm": 1.8266526696789749, "learning_rate": 1e-05, "loss": 0.4866, "step": 1323 }, { "epoch": 0.38155619596541784, "grad_norm": 1.8106958910859294, "learning_rate": 1e-05, "loss": 0.2919, "step": 1324 }, { "epoch": 0.3818443804034582, "grad_norm": 2.4230832410767764, "learning_rate": 1e-05, "loss": 0.5213, "step": 1325 }, { "epoch": 0.38213256484149855, "grad_norm": 1.7093502651446373, "learning_rate": 1e-05, "loss": 0.4743, "step": 1326 }, { "epoch": 0.3824207492795389, "grad_norm": 1.6023156819322093, "learning_rate": 1e-05, "loss": 0.2964, "step": 1327 }, { "epoch": 0.38270893371757925, "grad_norm": 2.7148544933679832, "learning_rate": 1e-05, "loss": 0.4697, "step": 1328 }, { "epoch": 0.3829971181556196, "grad_norm": 1.592202006561088, "learning_rate": 1e-05, "loss": 0.4831, "step": 1329 }, { "epoch": 0.38328530259365995, "grad_norm": 1.5708819141233699, "learning_rate": 1e-05, "loss": 0.3069, "step": 1330 }, { "epoch": 0.3835734870317003, "grad_norm": 2.193907768911475, "learning_rate": 1e-05, "loss": 0.4949, "step": 1331 }, { "epoch": 0.38386167146974065, "grad_norm": 1.82617787309074, "learning_rate": 1e-05, "loss": 0.4841, "step": 1332 }, { "epoch": 0.384149855907781, "grad_norm": 1.649677103567239, "learning_rate": 1e-05, "loss": 0.3103, "step": 1333 }, { "epoch": 0.38443804034582135, "grad_norm": 3.814639114475869, "learning_rate": 1e-05, "loss": 0.4801, "step": 1334 }, { "epoch": 0.38472622478386165, "grad_norm": 2.798330556491224, "learning_rate": 1e-05, "loss": 0.4931, "step": 1335 }, { "epoch": 0.385014409221902, "grad_norm": 2.3047698188749006, "learning_rate": 1e-05, "loss": 0.3029, "step": 1336 }, { "epoch": 0.38530259365994235, "grad_norm": 2.4107253949315135, "learning_rate": 1e-05, "loss": 0.4941, "step": 1337 }, { "epoch": 0.3855907780979827, "grad_norm": 2.340576269044811, "learning_rate": 1e-05, "loss": 0.4791, "step": 1338 }, { "epoch": 0.38587896253602305, "grad_norm": 2.3449681373393587, "learning_rate": 1e-05, "loss": 0.2941, "step": 1339 }, { "epoch": 0.3861671469740634, "grad_norm": 2.396207023795265, "learning_rate": 1e-05, "loss": 0.4957, "step": 1340 }, { "epoch": 0.38645533141210375, "grad_norm": 2.2731384422383605, "learning_rate": 1e-05, "loss": 0.4772, "step": 1341 }, { "epoch": 0.3867435158501441, "grad_norm": 2.1533263214602636, "learning_rate": 1e-05, "loss": 0.3438, "step": 1342 }, { "epoch": 0.38703170028818445, "grad_norm": 2.2200595691980127, "learning_rate": 1e-05, "loss": 0.4787, "step": 1343 }, { "epoch": 0.3873198847262248, "grad_norm": 1.7112106300037293, "learning_rate": 1e-05, "loss": 0.4826, "step": 1344 }, { "epoch": 0.38760806916426516, "grad_norm": 1.9886221802591015, "learning_rate": 1e-05, "loss": 0.2897, "step": 1345 }, { "epoch": 0.38789625360230545, "grad_norm": 4.093146799119056, "learning_rate": 1e-05, "loss": 0.5198, "step": 1346 }, { "epoch": 0.3881844380403458, "grad_norm": 1.893178744352946, "learning_rate": 1e-05, "loss": 0.4722, "step": 1347 }, { "epoch": 0.38847262247838615, "grad_norm": 2.3039484364608693, "learning_rate": 1e-05, "loss": 0.3054, "step": 1348 }, { "epoch": 0.3887608069164265, "grad_norm": 2.6947301278259728, "learning_rate": 1e-05, "loss": 0.497, "step": 1349 }, { "epoch": 0.38904899135446686, "grad_norm": 2.1200112126908897, "learning_rate": 1e-05, "loss": 0.4719, "step": 1350 }, { "epoch": 0.3893371757925072, "grad_norm": 1.4961252933768772, "learning_rate": 1e-05, "loss": 0.3102, "step": 1351 }, { "epoch": 0.38962536023054756, "grad_norm": 2.141053986186834, "learning_rate": 1e-05, "loss": 0.4916, "step": 1352 }, { "epoch": 0.3899135446685879, "grad_norm": 3.1611691572610483, "learning_rate": 1e-05, "loss": 0.4837, "step": 1353 }, { "epoch": 0.39020172910662826, "grad_norm": 1.6877173749956633, "learning_rate": 1e-05, "loss": 0.3019, "step": 1354 }, { "epoch": 0.3904899135446686, "grad_norm": 2.357194862463654, "learning_rate": 1e-05, "loss": 0.4183, "step": 1355 }, { "epoch": 0.39077809798270896, "grad_norm": 1.7644790755685829, "learning_rate": 1e-05, "loss": 0.491, "step": 1356 }, { "epoch": 0.39106628242074926, "grad_norm": 1.748201040487368, "learning_rate": 1e-05, "loss": 0.3265, "step": 1357 }, { "epoch": 0.3913544668587896, "grad_norm": 2.7881708335472766, "learning_rate": 1e-05, "loss": 0.4923, "step": 1358 }, { "epoch": 0.39164265129682996, "grad_norm": 1.9879518805417644, "learning_rate": 1e-05, "loss": 0.4709, "step": 1359 }, { "epoch": 0.3919308357348703, "grad_norm": 1.5462275623714095, "learning_rate": 1e-05, "loss": 0.3173, "step": 1360 }, { "epoch": 0.39221902017291066, "grad_norm": 2.245245037347222, "learning_rate": 1e-05, "loss": 0.5091, "step": 1361 }, { "epoch": 0.392507204610951, "grad_norm": 2.0100895309071953, "learning_rate": 1e-05, "loss": 0.4904, "step": 1362 }, { "epoch": 0.39279538904899136, "grad_norm": 2.0630055451528286, "learning_rate": 1e-05, "loss": 0.3033, "step": 1363 }, { "epoch": 0.3930835734870317, "grad_norm": 2.5843217211195606, "learning_rate": 1e-05, "loss": 0.4676, "step": 1364 }, { "epoch": 0.39337175792507206, "grad_norm": 1.790497685082556, "learning_rate": 1e-05, "loss": 0.4872, "step": 1365 }, { "epoch": 0.3936599423631124, "grad_norm": 3.0926875606313944, "learning_rate": 1e-05, "loss": 0.3062, "step": 1366 }, { "epoch": 0.39394812680115276, "grad_norm": 3.684659310424777, "learning_rate": 1e-05, "loss": 0.4969, "step": 1367 }, { "epoch": 0.39423631123919306, "grad_norm": 2.6374079921845732, "learning_rate": 1e-05, "loss": 0.481, "step": 1368 }, { "epoch": 0.3945244956772334, "grad_norm": 2.529947938300062, "learning_rate": 1e-05, "loss": 0.2729, "step": 1369 }, { "epoch": 0.39481268011527376, "grad_norm": 2.6403102299203063, "learning_rate": 1e-05, "loss": 0.5024, "step": 1370 }, { "epoch": 0.3951008645533141, "grad_norm": 2.4139843294062597, "learning_rate": 1e-05, "loss": 0.4853, "step": 1371 }, { "epoch": 0.39538904899135446, "grad_norm": 1.919988233786583, "learning_rate": 1e-05, "loss": 0.2881, "step": 1372 }, { "epoch": 0.3956772334293948, "grad_norm": 2.8129664017254736, "learning_rate": 1e-05, "loss": 0.4699, "step": 1373 }, { "epoch": 0.39596541786743517, "grad_norm": 1.634030689856549, "learning_rate": 1e-05, "loss": 0.4757, "step": 1374 }, { "epoch": 0.3962536023054755, "grad_norm": 3.424058629062118, "learning_rate": 1e-05, "loss": 0.3152, "step": 1375 }, { "epoch": 0.39654178674351587, "grad_norm": 3.112838442542796, "learning_rate": 1e-05, "loss": 0.5075, "step": 1376 }, { "epoch": 0.3968299711815562, "grad_norm": 2.1388924889196996, "learning_rate": 1e-05, "loss": 0.4765, "step": 1377 }, { "epoch": 0.39711815561959657, "grad_norm": 2.3876696427532207, "learning_rate": 1e-05, "loss": 0.3046, "step": 1378 }, { "epoch": 0.39740634005763686, "grad_norm": 2.7980681050344045, "learning_rate": 1e-05, "loss": 0.5026, "step": 1379 }, { "epoch": 0.3976945244956772, "grad_norm": 1.6533707032971585, "learning_rate": 1e-05, "loss": 0.4778, "step": 1380 }, { "epoch": 0.39798270893371757, "grad_norm": 3.6450302461981057, "learning_rate": 1e-05, "loss": 0.3155, "step": 1381 }, { "epoch": 0.3982708933717579, "grad_norm": 2.9253563119175645, "learning_rate": 1e-05, "loss": 0.5041, "step": 1382 }, { "epoch": 0.39855907780979827, "grad_norm": 1.9576576125421563, "learning_rate": 1e-05, "loss": 0.4854, "step": 1383 }, { "epoch": 0.3988472622478386, "grad_norm": 1.992720781724827, "learning_rate": 1e-05, "loss": 0.2959, "step": 1384 }, { "epoch": 0.39913544668587897, "grad_norm": 3.110274141952733, "learning_rate": 1e-05, "loss": 0.5109, "step": 1385 }, { "epoch": 0.3994236311239193, "grad_norm": 1.4837612858925824, "learning_rate": 1e-05, "loss": 0.4757, "step": 1386 }, { "epoch": 0.39971181556195967, "grad_norm": 1.5299937302980349, "learning_rate": 1e-05, "loss": 0.2849, "step": 1387 }, { "epoch": 0.4, "grad_norm": 2.3616359226141648, "learning_rate": 1e-05, "loss": 0.5097, "step": 1388 }, { "epoch": 0.4002881844380404, "grad_norm": 2.2733226729071183, "learning_rate": 1e-05, "loss": 0.4905, "step": 1389 }, { "epoch": 0.40057636887608067, "grad_norm": 1.9042849334434573, "learning_rate": 1e-05, "loss": 0.3066, "step": 1390 }, { "epoch": 0.400864553314121, "grad_norm": 3.129236639675653, "learning_rate": 1e-05, "loss": 0.4525, "step": 1391 }, { "epoch": 0.40115273775216137, "grad_norm": 1.60501688777708, "learning_rate": 1e-05, "loss": 0.4824, "step": 1392 }, { "epoch": 0.4014409221902017, "grad_norm": 2.4101207174232164, "learning_rate": 1e-05, "loss": 0.3235, "step": 1393 }, { "epoch": 0.40172910662824207, "grad_norm": 2.5438807558281376, "learning_rate": 1e-05, "loss": 0.4584, "step": 1394 }, { "epoch": 0.4020172910662824, "grad_norm": 2.0730066361358466, "learning_rate": 1e-05, "loss": 0.4803, "step": 1395 }, { "epoch": 0.4023054755043228, "grad_norm": 1.6283441998766814, "learning_rate": 1e-05, "loss": 0.2952, "step": 1396 }, { "epoch": 0.4025936599423631, "grad_norm": 3.131426981929381, "learning_rate": 1e-05, "loss": 0.4915, "step": 1397 }, { "epoch": 0.4028818443804035, "grad_norm": 2.184896822021097, "learning_rate": 1e-05, "loss": 0.4905, "step": 1398 }, { "epoch": 0.4031700288184438, "grad_norm": 1.9986154852514275, "learning_rate": 1e-05, "loss": 0.292, "step": 1399 }, { "epoch": 0.4034582132564842, "grad_norm": 2.522326409736831, "learning_rate": 1e-05, "loss": 0.4719, "step": 1400 }, { "epoch": 0.4037463976945245, "grad_norm": 1.7452401996021782, "learning_rate": 1e-05, "loss": 0.4777, "step": 1401 }, { "epoch": 0.4040345821325648, "grad_norm": 1.6123031076560754, "learning_rate": 1e-05, "loss": 0.32, "step": 1402 }, { "epoch": 0.4043227665706052, "grad_norm": 2.912808244320673, "learning_rate": 1e-05, "loss": 0.5106, "step": 1403 }, { "epoch": 0.4046109510086455, "grad_norm": 1.5889190205139632, "learning_rate": 1e-05, "loss": 0.4701, "step": 1404 }, { "epoch": 0.4048991354466859, "grad_norm": 2.6005923423253963, "learning_rate": 1e-05, "loss": 0.3034, "step": 1405 }, { "epoch": 0.4051873198847262, "grad_norm": 2.370651320021173, "learning_rate": 1e-05, "loss": 0.5104, "step": 1406 }, { "epoch": 0.4054755043227666, "grad_norm": 1.8283032148070109, "learning_rate": 1e-05, "loss": 0.4757, "step": 1407 }, { "epoch": 0.40576368876080693, "grad_norm": 1.7580438693746454, "learning_rate": 1e-05, "loss": 0.294, "step": 1408 }, { "epoch": 0.4060518731988473, "grad_norm": 2.563673768843064, "learning_rate": 1e-05, "loss": 0.4781, "step": 1409 }, { "epoch": 0.40634005763688763, "grad_norm": 2.047885702235358, "learning_rate": 1e-05, "loss": 0.4847, "step": 1410 }, { "epoch": 0.406628242074928, "grad_norm": 1.603045143656897, "learning_rate": 1e-05, "loss": 0.2888, "step": 1411 }, { "epoch": 0.4069164265129683, "grad_norm": 2.5886615969240387, "learning_rate": 1e-05, "loss": 0.4748, "step": 1412 }, { "epoch": 0.4072046109510086, "grad_norm": 1.6802958158922825, "learning_rate": 1e-05, "loss": 0.479, "step": 1413 }, { "epoch": 0.407492795389049, "grad_norm": 1.619522368911882, "learning_rate": 1e-05, "loss": 0.2841, "step": 1414 }, { "epoch": 0.40778097982708933, "grad_norm": 2.434723005486242, "learning_rate": 1e-05, "loss": 0.4428, "step": 1415 }, { "epoch": 0.4080691642651297, "grad_norm": 2.3743119348863813, "learning_rate": 1e-05, "loss": 0.4708, "step": 1416 }, { "epoch": 0.40835734870317003, "grad_norm": 1.5103269192763225, "learning_rate": 1e-05, "loss": 0.2837, "step": 1417 }, { "epoch": 0.4086455331412104, "grad_norm": 2.6875133844687644, "learning_rate": 1e-05, "loss": 0.4426, "step": 1418 }, { "epoch": 0.40893371757925073, "grad_norm": 1.742967920823814, "learning_rate": 1e-05, "loss": 0.4762, "step": 1419 }, { "epoch": 0.4092219020172911, "grad_norm": 1.5959741836972012, "learning_rate": 1e-05, "loss": 0.3049, "step": 1420 }, { "epoch": 0.40951008645533143, "grad_norm": 2.534248957154458, "learning_rate": 1e-05, "loss": 0.4601, "step": 1421 }, { "epoch": 0.4097982708933718, "grad_norm": 1.9664757858549118, "learning_rate": 1e-05, "loss": 0.482, "step": 1422 }, { "epoch": 0.4100864553314121, "grad_norm": 1.5027571897809393, "learning_rate": 1e-05, "loss": 0.2945, "step": 1423 }, { "epoch": 0.41037463976945243, "grad_norm": 2.952620010745475, "learning_rate": 1e-05, "loss": 0.5281, "step": 1424 }, { "epoch": 0.4106628242074928, "grad_norm": 1.9273716488939223, "learning_rate": 1e-05, "loss": 0.477, "step": 1425 }, { "epoch": 0.41095100864553313, "grad_norm": 1.9634936397917053, "learning_rate": 1e-05, "loss": 0.3252, "step": 1426 }, { "epoch": 0.4112391930835735, "grad_norm": 2.693940835759522, "learning_rate": 1e-05, "loss": 0.4795, "step": 1427 }, { "epoch": 0.41152737752161384, "grad_norm": 2.4222487408861277, "learning_rate": 1e-05, "loss": 0.4759, "step": 1428 }, { "epoch": 0.4118155619596542, "grad_norm": 2.1830351365182246, "learning_rate": 1e-05, "loss": 0.306, "step": 1429 }, { "epoch": 0.41210374639769454, "grad_norm": 2.4769551938384864, "learning_rate": 1e-05, "loss": 0.4586, "step": 1430 }, { "epoch": 0.4123919308357349, "grad_norm": 2.161859985283154, "learning_rate": 1e-05, "loss": 0.4992, "step": 1431 }, { "epoch": 0.41268011527377524, "grad_norm": 1.9311564431354833, "learning_rate": 1e-05, "loss": 0.3019, "step": 1432 }, { "epoch": 0.4129682997118156, "grad_norm": 2.545808915398012, "learning_rate": 1e-05, "loss": 0.488, "step": 1433 }, { "epoch": 0.4132564841498559, "grad_norm": 2.0614808574437897, "learning_rate": 1e-05, "loss": 0.4795, "step": 1434 }, { "epoch": 0.41354466858789624, "grad_norm": 1.56057299395756, "learning_rate": 1e-05, "loss": 0.3019, "step": 1435 }, { "epoch": 0.4138328530259366, "grad_norm": 2.2399026840728262, "learning_rate": 1e-05, "loss": 0.4724, "step": 1436 }, { "epoch": 0.41412103746397694, "grad_norm": 2.105709306541403, "learning_rate": 1e-05, "loss": 0.4693, "step": 1437 }, { "epoch": 0.4144092219020173, "grad_norm": 2.162936762973091, "learning_rate": 1e-05, "loss": 0.3169, "step": 1438 }, { "epoch": 0.41469740634005764, "grad_norm": 2.5542633457686272, "learning_rate": 1e-05, "loss": 0.4678, "step": 1439 }, { "epoch": 0.414985590778098, "grad_norm": 1.4167261304479106, "learning_rate": 1e-05, "loss": 0.4745, "step": 1440 }, { "epoch": 0.41527377521613834, "grad_norm": 1.61449310937752, "learning_rate": 1e-05, "loss": 0.291, "step": 1441 }, { "epoch": 0.4155619596541787, "grad_norm": 2.460741009092124, "learning_rate": 1e-05, "loss": 0.4947, "step": 1442 }, { "epoch": 0.41585014409221904, "grad_norm": 1.2816142509425972, "learning_rate": 1e-05, "loss": 0.4696, "step": 1443 }, { "epoch": 0.4161383285302594, "grad_norm": 1.6593680240842172, "learning_rate": 1e-05, "loss": 0.2982, "step": 1444 }, { "epoch": 0.4164265129682997, "grad_norm": 2.304564977019813, "learning_rate": 1e-05, "loss": 0.428, "step": 1445 }, { "epoch": 0.41671469740634004, "grad_norm": 1.5971229954726822, "learning_rate": 1e-05, "loss": 0.4756, "step": 1446 }, { "epoch": 0.4170028818443804, "grad_norm": 1.752756272258912, "learning_rate": 1e-05, "loss": 0.2723, "step": 1447 }, { "epoch": 0.41729106628242074, "grad_norm": 2.2574850631520462, "learning_rate": 1e-05, "loss": 0.494, "step": 1448 }, { "epoch": 0.4175792507204611, "grad_norm": 1.5487360772919614, "learning_rate": 1e-05, "loss": 0.4746, "step": 1449 }, { "epoch": 0.41786743515850144, "grad_norm": 1.6966590813465043, "learning_rate": 1e-05, "loss": 0.3112, "step": 1450 }, { "epoch": 0.4181556195965418, "grad_norm": 3.2315311759123118, "learning_rate": 1e-05, "loss": 0.4753, "step": 1451 }, { "epoch": 0.41844380403458215, "grad_norm": 2.2574661317758937, "learning_rate": 1e-05, "loss": 0.4802, "step": 1452 }, { "epoch": 0.4187319884726225, "grad_norm": 3.9803998971006664, "learning_rate": 1e-05, "loss": 0.2902, "step": 1453 }, { "epoch": 0.41902017291066285, "grad_norm": 3.209780933951834, "learning_rate": 1e-05, "loss": 0.4602, "step": 1454 }, { "epoch": 0.41930835734870314, "grad_norm": 2.230305806397594, "learning_rate": 1e-05, "loss": 0.4744, "step": 1455 }, { "epoch": 0.4195965417867435, "grad_norm": 2.0377336323745716, "learning_rate": 1e-05, "loss": 0.314, "step": 1456 }, { "epoch": 0.41988472622478384, "grad_norm": 2.0423812560080714, "learning_rate": 1e-05, "loss": 0.4396, "step": 1457 }, { "epoch": 0.4201729106628242, "grad_norm": 2.058425649358438, "learning_rate": 1e-05, "loss": 0.4805, "step": 1458 }, { "epoch": 0.42046109510086455, "grad_norm": 2.107814332893319, "learning_rate": 1e-05, "loss": 0.282, "step": 1459 }, { "epoch": 0.4207492795389049, "grad_norm": 4.456406890207841, "learning_rate": 1e-05, "loss": 0.4971, "step": 1460 }, { "epoch": 0.42103746397694525, "grad_norm": 1.8378400632612824, "learning_rate": 1e-05, "loss": 0.4748, "step": 1461 }, { "epoch": 0.4213256484149856, "grad_norm": 2.11974453191249, "learning_rate": 1e-05, "loss": 0.3012, "step": 1462 }, { "epoch": 0.42161383285302595, "grad_norm": 2.686223251549633, "learning_rate": 1e-05, "loss": 0.503, "step": 1463 }, { "epoch": 0.4219020172910663, "grad_norm": 1.8509864605201591, "learning_rate": 1e-05, "loss": 0.4759, "step": 1464 }, { "epoch": 0.42219020172910665, "grad_norm": 2.6826435072967882, "learning_rate": 1e-05, "loss": 0.2867, "step": 1465 }, { "epoch": 0.42247838616714695, "grad_norm": 4.054905726717292, "learning_rate": 1e-05, "loss": 0.5015, "step": 1466 }, { "epoch": 0.4227665706051873, "grad_norm": 1.7056820602826313, "learning_rate": 1e-05, "loss": 0.4786, "step": 1467 }, { "epoch": 0.42305475504322765, "grad_norm": 2.373737807042705, "learning_rate": 1e-05, "loss": 0.2919, "step": 1468 }, { "epoch": 0.423342939481268, "grad_norm": 3.1455534455823435, "learning_rate": 1e-05, "loss": 0.4915, "step": 1469 }, { "epoch": 0.42363112391930835, "grad_norm": 1.5609259233210926, "learning_rate": 1e-05, "loss": 0.4792, "step": 1470 }, { "epoch": 0.4239193083573487, "grad_norm": 2.2781872873900832, "learning_rate": 1e-05, "loss": 0.2805, "step": 1471 }, { "epoch": 0.42420749279538905, "grad_norm": 2.5535708657805336, "learning_rate": 1e-05, "loss": 0.4633, "step": 1472 }, { "epoch": 0.4244956772334294, "grad_norm": 2.576172876923238, "learning_rate": 1e-05, "loss": 0.4713, "step": 1473 }, { "epoch": 0.42478386167146975, "grad_norm": 2.413693568839609, "learning_rate": 1e-05, "loss": 0.2942, "step": 1474 }, { "epoch": 0.4250720461095101, "grad_norm": 3.7683768285466663, "learning_rate": 1e-05, "loss": 0.5096, "step": 1475 }, { "epoch": 0.42536023054755046, "grad_norm": 1.4097587762945767, "learning_rate": 1e-05, "loss": 0.4767, "step": 1476 }, { "epoch": 0.42564841498559075, "grad_norm": 1.9611481825788895, "learning_rate": 1e-05, "loss": 0.3, "step": 1477 }, { "epoch": 0.4259365994236311, "grad_norm": 2.7697927302614125, "learning_rate": 1e-05, "loss": 0.4952, "step": 1478 }, { "epoch": 0.42622478386167145, "grad_norm": 2.0120709047065715, "learning_rate": 1e-05, "loss": 0.4861, "step": 1479 }, { "epoch": 0.4265129682997118, "grad_norm": 2.239333012962598, "learning_rate": 1e-05, "loss": 0.3059, "step": 1480 }, { "epoch": 0.42680115273775215, "grad_norm": 3.9713986125353244, "learning_rate": 1e-05, "loss": 0.5035, "step": 1481 }, { "epoch": 0.4270893371757925, "grad_norm": 2.135974392250055, "learning_rate": 1e-05, "loss": 0.4727, "step": 1482 }, { "epoch": 0.42737752161383286, "grad_norm": 1.7694741634549267, "learning_rate": 1e-05, "loss": 0.3134, "step": 1483 }, { "epoch": 0.4276657060518732, "grad_norm": 2.201081975407542, "learning_rate": 1e-05, "loss": 0.4837, "step": 1484 }, { "epoch": 0.42795389048991356, "grad_norm": 2.1049177714570804, "learning_rate": 1e-05, "loss": 0.4713, "step": 1485 }, { "epoch": 0.4282420749279539, "grad_norm": 1.828062091862777, "learning_rate": 1e-05, "loss": 0.3028, "step": 1486 }, { "epoch": 0.42853025936599426, "grad_norm": 3.388980148519981, "learning_rate": 1e-05, "loss": 0.4695, "step": 1487 }, { "epoch": 0.42881844380403455, "grad_norm": 2.071387382431709, "learning_rate": 1e-05, "loss": 0.4655, "step": 1488 }, { "epoch": 0.4291066282420749, "grad_norm": 2.473212913479042, "learning_rate": 1e-05, "loss": 0.2895, "step": 1489 }, { "epoch": 0.42939481268011526, "grad_norm": 3.5488756794281264, "learning_rate": 1e-05, "loss": 0.4806, "step": 1490 }, { "epoch": 0.4296829971181556, "grad_norm": 2.070394697000306, "learning_rate": 1e-05, "loss": 0.4697, "step": 1491 }, { "epoch": 0.42997118155619596, "grad_norm": 1.7693617947762856, "learning_rate": 1e-05, "loss": 0.2963, "step": 1492 }, { "epoch": 0.4302593659942363, "grad_norm": 2.2658425376271634, "learning_rate": 1e-05, "loss": 0.4988, "step": 1493 }, { "epoch": 0.43054755043227666, "grad_norm": 1.635647565764361, "learning_rate": 1e-05, "loss": 0.4618, "step": 1494 }, { "epoch": 0.430835734870317, "grad_norm": 2.844256003261886, "learning_rate": 1e-05, "loss": 0.2956, "step": 1495 }, { "epoch": 0.43112391930835736, "grad_norm": 1.8942081514772837, "learning_rate": 1e-05, "loss": 0.4516, "step": 1496 }, { "epoch": 0.4314121037463977, "grad_norm": 1.8172410700735635, "learning_rate": 1e-05, "loss": 0.4769, "step": 1497 }, { "epoch": 0.43170028818443806, "grad_norm": 2.007527346024837, "learning_rate": 1e-05, "loss": 0.2824, "step": 1498 }, { "epoch": 0.43198847262247836, "grad_norm": 3.226794105209606, "learning_rate": 1e-05, "loss": 0.5084, "step": 1499 }, { "epoch": 0.4322766570605187, "grad_norm": 1.906344602352076, "learning_rate": 1e-05, "loss": 0.4764, "step": 1500 }, { "epoch": 0.4322766570605187, "eval_websight_new_IoU": 0.018606476485729218, "eval_websight_new_MAE_x": 286.7799987792969, "eval_websight_new_MAE_y": 592.5800170898438, "eval_websight_new_coord_probability": 0.7464273571968079, "eval_websight_new_correct_token_probability": 0.0013086475664749742, "eval_websight_new_inside_bbox": 0.019999999552965164, "eval_websight_new_loss": 0.3021695017814636, "eval_websight_new_range_100_probability": 0.09368956834077835, "eval_websight_new_range_20_probability": 0.017098380252718925, "eval_websight_new_range_5_probability": 0.003493149532005191, "eval_websight_new_runtime": 34.573, "eval_websight_new_samples_per_second": 1.446, "eval_websight_new_steps_per_second": 0.202, "step": 1500 }, { "epoch": 0.4322766570605187, "eval_seeclick_IoU": 0.0315781868994236, "eval_seeclick_MAE_x": 473.1000061035156, "eval_seeclick_MAE_y": 555.760009765625, "eval_seeclick_coord_probability": 0.7461659908294678, "eval_seeclick_correct_token_probability": 0.0007614689529873431, "eval_seeclick_inside_bbox": 0.07999999821186066, "eval_seeclick_loss": 0.3881937861442566, "eval_seeclick_range_100_probability": 0.1348019391298294, "eval_seeclick_range_20_probability": 0.053533121943473816, "eval_seeclick_range_5_probability": 0.008099403232336044, "eval_seeclick_runtime": 41.6794, "eval_seeclick_samples_per_second": 1.2, "eval_seeclick_steps_per_second": 0.168, "step": 1500 }, { "epoch": 0.4322766570605187, "eval_icons_IoU": 0.002965953666716814, "eval_icons_MAE_x": 473.5400085449219, "eval_icons_MAE_y": 583.760009765625, "eval_icons_coord_probability": 0.747755229473114, "eval_icons_correct_token_probability": 0.00685780169442296, "eval_icons_inside_bbox": 0.0, "eval_icons_loss": 0.45823726058006287, "eval_icons_range_100_probability": 0.27524334192276, "eval_icons_range_20_probability": 0.09912868589162827, "eval_icons_range_5_probability": 0.030710894614458084, "eval_icons_runtime": 43.7339, "eval_icons_samples_per_second": 1.143, "eval_icons_steps_per_second": 0.16, "step": 1500 }, { "epoch": 0.43256484149855906, "grad_norm": 1.5872574925183616, "learning_rate": 1e-05, "loss": 0.2881, "step": 1501 }, { "epoch": 0.4328530259365994, "grad_norm": 2.513250812854745, "learning_rate": 1e-05, "loss": 0.4873, "step": 1502 }, { "epoch": 0.43314121037463976, "grad_norm": 1.4467163565839434, "learning_rate": 1e-05, "loss": 0.4754, "step": 1503 }, { "epoch": 0.4334293948126801, "grad_norm": 2.8143907431672854, "learning_rate": 1e-05, "loss": 0.2882, "step": 1504 }, { "epoch": 0.43371757925072046, "grad_norm": 3.1965223788772543, "learning_rate": 1e-05, "loss": 0.464, "step": 1505 }, { "epoch": 0.4340057636887608, "grad_norm": 1.9414117477439994, "learning_rate": 1e-05, "loss": 0.4703, "step": 1506 }, { "epoch": 0.43429394812680117, "grad_norm": 1.476265941565406, "learning_rate": 1e-05, "loss": 0.2882, "step": 1507 }, { "epoch": 0.4345821325648415, "grad_norm": 3.288435398577001, "learning_rate": 1e-05, "loss": 0.4863, "step": 1508 }, { "epoch": 0.43487031700288187, "grad_norm": 1.7781756587565274, "learning_rate": 1e-05, "loss": 0.4782, "step": 1509 }, { "epoch": 0.43515850144092216, "grad_norm": 1.5639576209977695, "learning_rate": 1e-05, "loss": 0.2804, "step": 1510 }, { "epoch": 0.4354466858789625, "grad_norm": 2.5849426529224067, "learning_rate": 1e-05, "loss": 0.4566, "step": 1511 }, { "epoch": 0.43573487031700286, "grad_norm": 2.224779390377121, "learning_rate": 1e-05, "loss": 0.4695, "step": 1512 }, { "epoch": 0.4360230547550432, "grad_norm": 2.518276835035036, "learning_rate": 1e-05, "loss": 0.2947, "step": 1513 }, { "epoch": 0.43631123919308357, "grad_norm": 2.16698257127677, "learning_rate": 1e-05, "loss": 0.4633, "step": 1514 }, { "epoch": 0.4365994236311239, "grad_norm": 2.0026538905015023, "learning_rate": 1e-05, "loss": 0.4654, "step": 1515 }, { "epoch": 0.43688760806916427, "grad_norm": 1.5740477738801, "learning_rate": 1e-05, "loss": 0.3039, "step": 1516 }, { "epoch": 0.4371757925072046, "grad_norm": 2.237092615867006, "learning_rate": 1e-05, "loss": 0.4779, "step": 1517 }, { "epoch": 0.43746397694524497, "grad_norm": 2.8136904641085607, "learning_rate": 1e-05, "loss": 0.4734, "step": 1518 }, { "epoch": 0.4377521613832853, "grad_norm": 1.4916193545442005, "learning_rate": 1e-05, "loss": 0.2616, "step": 1519 }, { "epoch": 0.43804034582132567, "grad_norm": 2.319559728956529, "learning_rate": 1e-05, "loss": 0.479, "step": 1520 }, { "epoch": 0.43832853025936597, "grad_norm": 1.9413915090878375, "learning_rate": 1e-05, "loss": 0.4712, "step": 1521 }, { "epoch": 0.4386167146974063, "grad_norm": 1.5467348645246979, "learning_rate": 1e-05, "loss": 0.2958, "step": 1522 }, { "epoch": 0.43890489913544667, "grad_norm": 3.514060979099786, "learning_rate": 1e-05, "loss": 0.4728, "step": 1523 }, { "epoch": 0.439193083573487, "grad_norm": 1.816045519131449, "learning_rate": 1e-05, "loss": 0.4772, "step": 1524 }, { "epoch": 0.43948126801152737, "grad_norm": 2.0350489231668996, "learning_rate": 1e-05, "loss": 0.3024, "step": 1525 }, { "epoch": 0.4397694524495677, "grad_norm": 3.2219810135777163, "learning_rate": 1e-05, "loss": 0.4523, "step": 1526 }, { "epoch": 0.4400576368876081, "grad_norm": 2.0807276821580425, "learning_rate": 1e-05, "loss": 0.4596, "step": 1527 }, { "epoch": 0.4403458213256484, "grad_norm": 2.09696549398811, "learning_rate": 1e-05, "loss": 0.3015, "step": 1528 }, { "epoch": 0.4406340057636888, "grad_norm": 3.4772008092345343, "learning_rate": 1e-05, "loss": 0.4883, "step": 1529 }, { "epoch": 0.4409221902017291, "grad_norm": 1.82544342216196, "learning_rate": 1e-05, "loss": 0.4658, "step": 1530 }, { "epoch": 0.4412103746397695, "grad_norm": 1.723235618489549, "learning_rate": 1e-05, "loss": 0.2821, "step": 1531 }, { "epoch": 0.44149855907780977, "grad_norm": 2.3441861237895467, "learning_rate": 1e-05, "loss": 0.4624, "step": 1532 }, { "epoch": 0.4417867435158501, "grad_norm": 1.6247000768593272, "learning_rate": 1e-05, "loss": 0.4691, "step": 1533 }, { "epoch": 0.4420749279538905, "grad_norm": 1.9745718388471114, "learning_rate": 1e-05, "loss": 0.2873, "step": 1534 }, { "epoch": 0.4423631123919308, "grad_norm": 3.324646745868719, "learning_rate": 1e-05, "loss": 0.4745, "step": 1535 }, { "epoch": 0.4426512968299712, "grad_norm": 1.8867451406839792, "learning_rate": 1e-05, "loss": 0.4717, "step": 1536 }, { "epoch": 0.4429394812680115, "grad_norm": 2.4174374781034964, "learning_rate": 1e-05, "loss": 0.2877, "step": 1537 }, { "epoch": 0.4432276657060519, "grad_norm": 2.5017887054244565, "learning_rate": 1e-05, "loss": 0.4598, "step": 1538 }, { "epoch": 0.4435158501440922, "grad_norm": 1.7850722546794648, "learning_rate": 1e-05, "loss": 0.4674, "step": 1539 }, { "epoch": 0.4438040345821326, "grad_norm": 2.148436440003259, "learning_rate": 1e-05, "loss": 0.2897, "step": 1540 }, { "epoch": 0.44409221902017293, "grad_norm": 3.2171685780383497, "learning_rate": 1e-05, "loss": 0.4662, "step": 1541 }, { "epoch": 0.4443804034582133, "grad_norm": 1.5096168271723365, "learning_rate": 1e-05, "loss": 0.4606, "step": 1542 }, { "epoch": 0.4446685878962536, "grad_norm": 1.8582772475076226, "learning_rate": 1e-05, "loss": 0.2873, "step": 1543 }, { "epoch": 0.4449567723342939, "grad_norm": 2.5279416610391325, "learning_rate": 1e-05, "loss": 0.4558, "step": 1544 }, { "epoch": 0.4452449567723343, "grad_norm": 1.9269150992686295, "learning_rate": 1e-05, "loss": 0.4652, "step": 1545 }, { "epoch": 0.44553314121037463, "grad_norm": 2.1094191576045493, "learning_rate": 1e-05, "loss": 0.2997, "step": 1546 }, { "epoch": 0.445821325648415, "grad_norm": 2.7335428758417453, "learning_rate": 1e-05, "loss": 0.4816, "step": 1547 }, { "epoch": 0.44610951008645533, "grad_norm": 1.4570127507970505, "learning_rate": 1e-05, "loss": 0.4741, "step": 1548 }, { "epoch": 0.4463976945244957, "grad_norm": 2.0189080812348483, "learning_rate": 1e-05, "loss": 0.2861, "step": 1549 }, { "epoch": 0.44668587896253603, "grad_norm": 3.269122993577927, "learning_rate": 1e-05, "loss": 0.4595, "step": 1550 }, { "epoch": 0.4469740634005764, "grad_norm": 1.5717104712633572, "learning_rate": 1e-05, "loss": 0.4653, "step": 1551 }, { "epoch": 0.44726224783861673, "grad_norm": 2.0530804084811005, "learning_rate": 1e-05, "loss": 0.2712, "step": 1552 }, { "epoch": 0.4475504322766571, "grad_norm": 2.2505052949895505, "learning_rate": 1e-05, "loss": 0.4901, "step": 1553 }, { "epoch": 0.4478386167146974, "grad_norm": 2.041908678963047, "learning_rate": 1e-05, "loss": 0.4785, "step": 1554 }, { "epoch": 0.44812680115273773, "grad_norm": 1.6502146780344642, "learning_rate": 1e-05, "loss": 0.2822, "step": 1555 }, { "epoch": 0.4484149855907781, "grad_norm": 2.09226285052814, "learning_rate": 1e-05, "loss": 0.4674, "step": 1556 }, { "epoch": 0.44870317002881843, "grad_norm": 1.682934145653075, "learning_rate": 1e-05, "loss": 0.4692, "step": 1557 }, { "epoch": 0.4489913544668588, "grad_norm": 1.7936155554218134, "learning_rate": 1e-05, "loss": 0.2736, "step": 1558 }, { "epoch": 0.44927953890489913, "grad_norm": 2.9374696279877224, "learning_rate": 1e-05, "loss": 0.4824, "step": 1559 }, { "epoch": 0.4495677233429395, "grad_norm": 2.018776048115124, "learning_rate": 1e-05, "loss": 0.4673, "step": 1560 }, { "epoch": 0.44985590778097984, "grad_norm": 1.8145829483936216, "learning_rate": 1e-05, "loss": 0.2616, "step": 1561 }, { "epoch": 0.4501440922190202, "grad_norm": 2.311170758105593, "learning_rate": 1e-05, "loss": 0.4661, "step": 1562 }, { "epoch": 0.45043227665706054, "grad_norm": 2.1575239049221206, "learning_rate": 1e-05, "loss": 0.4726, "step": 1563 }, { "epoch": 0.4507204610951009, "grad_norm": 1.398521466330883, "learning_rate": 1e-05, "loss": 0.2966, "step": 1564 }, { "epoch": 0.4510086455331412, "grad_norm": 3.0139786653151392, "learning_rate": 1e-05, "loss": 0.4256, "step": 1565 }, { "epoch": 0.45129682997118153, "grad_norm": 1.5206472061961696, "learning_rate": 1e-05, "loss": 0.4687, "step": 1566 }, { "epoch": 0.4515850144092219, "grad_norm": 2.160624694251746, "learning_rate": 1e-05, "loss": 0.3051, "step": 1567 }, { "epoch": 0.45187319884726224, "grad_norm": 2.5482743201834444, "learning_rate": 1e-05, "loss": 0.443, "step": 1568 }, { "epoch": 0.4521613832853026, "grad_norm": 2.0224049925494327, "learning_rate": 1e-05, "loss": 0.4585, "step": 1569 }, { "epoch": 0.45244956772334294, "grad_norm": 2.460064527237816, "learning_rate": 1e-05, "loss": 0.2961, "step": 1570 }, { "epoch": 0.4527377521613833, "grad_norm": 3.2057774393838154, "learning_rate": 1e-05, "loss": 0.4577, "step": 1571 }, { "epoch": 0.45302593659942364, "grad_norm": 1.7868727145251238, "learning_rate": 1e-05, "loss": 0.4715, "step": 1572 }, { "epoch": 0.453314121037464, "grad_norm": 2.29698423854333, "learning_rate": 1e-05, "loss": 0.2844, "step": 1573 }, { "epoch": 0.45360230547550434, "grad_norm": 2.917727972513402, "learning_rate": 1e-05, "loss": 0.4916, "step": 1574 }, { "epoch": 0.4538904899135447, "grad_norm": 1.6247585355469811, "learning_rate": 1e-05, "loss": 0.4669, "step": 1575 }, { "epoch": 0.454178674351585, "grad_norm": 1.5614244317128618, "learning_rate": 1e-05, "loss": 0.2599, "step": 1576 }, { "epoch": 0.45446685878962534, "grad_norm": 2.1206996263612243, "learning_rate": 1e-05, "loss": 0.4601, "step": 1577 }, { "epoch": 0.4547550432276657, "grad_norm": 2.047320752803363, "learning_rate": 1e-05, "loss": 0.4831, "step": 1578 }, { "epoch": 0.45504322766570604, "grad_norm": 2.1405983343150456, "learning_rate": 1e-05, "loss": 0.2845, "step": 1579 }, { "epoch": 0.4553314121037464, "grad_norm": 2.458083877361452, "learning_rate": 1e-05, "loss": 0.4693, "step": 1580 }, { "epoch": 0.45561959654178674, "grad_norm": 1.6355539726377584, "learning_rate": 1e-05, "loss": 0.4725, "step": 1581 }, { "epoch": 0.4559077809798271, "grad_norm": 1.878882098430483, "learning_rate": 1e-05, "loss": 0.2986, "step": 1582 }, { "epoch": 0.45619596541786744, "grad_norm": 3.4414263782465637, "learning_rate": 1e-05, "loss": 0.4829, "step": 1583 }, { "epoch": 0.4564841498559078, "grad_norm": 1.7070299972485468, "learning_rate": 1e-05, "loss": 0.4626, "step": 1584 }, { "epoch": 0.45677233429394815, "grad_norm": 1.4912568027028905, "learning_rate": 1e-05, "loss": 0.2611, "step": 1585 }, { "epoch": 0.4570605187319885, "grad_norm": 2.5139741315017567, "learning_rate": 1e-05, "loss": 0.4912, "step": 1586 }, { "epoch": 0.4573487031700288, "grad_norm": 1.7131976761887835, "learning_rate": 1e-05, "loss": 0.478, "step": 1587 }, { "epoch": 0.45763688760806914, "grad_norm": 1.5223364212938868, "learning_rate": 1e-05, "loss": 0.2757, "step": 1588 }, { "epoch": 0.4579250720461095, "grad_norm": 2.090334564468152, "learning_rate": 1e-05, "loss": 0.4376, "step": 1589 }, { "epoch": 0.45821325648414984, "grad_norm": 1.384471336032835, "learning_rate": 1e-05, "loss": 0.4659, "step": 1590 }, { "epoch": 0.4585014409221902, "grad_norm": 1.6695985549873653, "learning_rate": 1e-05, "loss": 0.275, "step": 1591 }, { "epoch": 0.45878962536023055, "grad_norm": 2.2964318697011254, "learning_rate": 1e-05, "loss": 0.4346, "step": 1592 }, { "epoch": 0.4590778097982709, "grad_norm": 1.8497059623298822, "learning_rate": 1e-05, "loss": 0.4704, "step": 1593 }, { "epoch": 0.45936599423631125, "grad_norm": 1.6835772981604133, "learning_rate": 1e-05, "loss": 0.2774, "step": 1594 }, { "epoch": 0.4596541786743516, "grad_norm": 2.7826512914461503, "learning_rate": 1e-05, "loss": 0.4566, "step": 1595 }, { "epoch": 0.45994236311239195, "grad_norm": 1.9990675916202083, "learning_rate": 1e-05, "loss": 0.4626, "step": 1596 }, { "epoch": 0.4602305475504323, "grad_norm": 1.6885590526293213, "learning_rate": 1e-05, "loss": 0.273, "step": 1597 }, { "epoch": 0.4605187319884726, "grad_norm": 3.0511730495487233, "learning_rate": 1e-05, "loss": 0.4417, "step": 1598 }, { "epoch": 0.46080691642651295, "grad_norm": 1.9622648574010686, "learning_rate": 1e-05, "loss": 0.467, "step": 1599 }, { "epoch": 0.4610951008645533, "grad_norm": 1.5492189056892085, "learning_rate": 1e-05, "loss": 0.2778, "step": 1600 }, { "epoch": 0.46138328530259365, "grad_norm": 2.616357665417584, "learning_rate": 1e-05, "loss": 0.4732, "step": 1601 }, { "epoch": 0.461671469740634, "grad_norm": 2.220624824895359, "learning_rate": 1e-05, "loss": 0.4628, "step": 1602 }, { "epoch": 0.46195965417867435, "grad_norm": 2.5573975084738882, "learning_rate": 1e-05, "loss": 0.2894, "step": 1603 }, { "epoch": 0.4622478386167147, "grad_norm": 2.4115610321105403, "learning_rate": 1e-05, "loss": 0.4605, "step": 1604 }, { "epoch": 0.46253602305475505, "grad_norm": 1.9242603208135884, "learning_rate": 1e-05, "loss": 0.4705, "step": 1605 }, { "epoch": 0.4628242074927954, "grad_norm": 2.0542088440136035, "learning_rate": 1e-05, "loss": 0.2821, "step": 1606 }, { "epoch": 0.46311239193083575, "grad_norm": 2.234799390002291, "learning_rate": 1e-05, "loss": 0.5011, "step": 1607 }, { "epoch": 0.4634005763688761, "grad_norm": 2.3923851771137374, "learning_rate": 1e-05, "loss": 0.4651, "step": 1608 }, { "epoch": 0.4636887608069164, "grad_norm": 1.4616520651039109, "learning_rate": 1e-05, "loss": 0.2639, "step": 1609 }, { "epoch": 0.46397694524495675, "grad_norm": 2.9292817123777364, "learning_rate": 1e-05, "loss": 0.4638, "step": 1610 }, { "epoch": 0.4642651296829971, "grad_norm": 1.7415947974614805, "learning_rate": 1e-05, "loss": 0.4762, "step": 1611 }, { "epoch": 0.46455331412103745, "grad_norm": 1.670760159121049, "learning_rate": 1e-05, "loss": 0.2658, "step": 1612 }, { "epoch": 0.4648414985590778, "grad_norm": 2.5213464652038535, "learning_rate": 1e-05, "loss": 0.4756, "step": 1613 }, { "epoch": 0.46512968299711815, "grad_norm": 1.628758078685135, "learning_rate": 1e-05, "loss": 0.465, "step": 1614 }, { "epoch": 0.4654178674351585, "grad_norm": 1.5732845415610626, "learning_rate": 1e-05, "loss": 0.2923, "step": 1615 }, { "epoch": 0.46570605187319886, "grad_norm": 3.030593908464025, "learning_rate": 1e-05, "loss": 0.4769, "step": 1616 }, { "epoch": 0.4659942363112392, "grad_norm": 1.6352362506907436, "learning_rate": 1e-05, "loss": 0.4661, "step": 1617 }, { "epoch": 0.46628242074927956, "grad_norm": 1.6462901234522942, "learning_rate": 1e-05, "loss": 0.2857, "step": 1618 }, { "epoch": 0.4665706051873199, "grad_norm": 2.305484901094868, "learning_rate": 1e-05, "loss": 0.4336, "step": 1619 }, { "epoch": 0.4668587896253602, "grad_norm": 2.557398249605915, "learning_rate": 1e-05, "loss": 0.4593, "step": 1620 }, { "epoch": 0.46714697406340056, "grad_norm": 2.1407733470243153, "learning_rate": 1e-05, "loss": 0.2638, "step": 1621 }, { "epoch": 0.4674351585014409, "grad_norm": 3.3332545163100904, "learning_rate": 1e-05, "loss": 0.4365, "step": 1622 }, { "epoch": 0.46772334293948126, "grad_norm": 1.9781546237826027, "learning_rate": 1e-05, "loss": 0.4655, "step": 1623 }, { "epoch": 0.4680115273775216, "grad_norm": 2.521960806890818, "learning_rate": 1e-05, "loss": 0.2799, "step": 1624 }, { "epoch": 0.46829971181556196, "grad_norm": 2.510635807760795, "learning_rate": 1e-05, "loss": 0.4609, "step": 1625 }, { "epoch": 0.4685878962536023, "grad_norm": 1.9715299569973497, "learning_rate": 1e-05, "loss": 0.4747, "step": 1626 }, { "epoch": 0.46887608069164266, "grad_norm": 1.5624894931516837, "learning_rate": 1e-05, "loss": 0.2825, "step": 1627 }, { "epoch": 0.469164265129683, "grad_norm": 2.340010305197747, "learning_rate": 1e-05, "loss": 0.4476, "step": 1628 }, { "epoch": 0.46945244956772336, "grad_norm": 2.181830008102016, "learning_rate": 1e-05, "loss": 0.4608, "step": 1629 }, { "epoch": 0.4697406340057637, "grad_norm": 2.3829939547019072, "learning_rate": 1e-05, "loss": 0.2766, "step": 1630 }, { "epoch": 0.470028818443804, "grad_norm": 2.1314038791907612, "learning_rate": 1e-05, "loss": 0.4594, "step": 1631 }, { "epoch": 0.47031700288184436, "grad_norm": 1.8878699816203297, "learning_rate": 1e-05, "loss": 0.468, "step": 1632 }, { "epoch": 0.4706051873198847, "grad_norm": 1.4414170428945374, "learning_rate": 1e-05, "loss": 0.2663, "step": 1633 }, { "epoch": 0.47089337175792506, "grad_norm": 3.1604308673524635, "learning_rate": 1e-05, "loss": 0.4356, "step": 1634 }, { "epoch": 0.4711815561959654, "grad_norm": 2.393902564518669, "learning_rate": 1e-05, "loss": 0.4609, "step": 1635 }, { "epoch": 0.47146974063400576, "grad_norm": 2.334488135606776, "learning_rate": 1e-05, "loss": 0.2752, "step": 1636 }, { "epoch": 0.4717579250720461, "grad_norm": 2.615335064480252, "learning_rate": 1e-05, "loss": 0.4731, "step": 1637 }, { "epoch": 0.47204610951008646, "grad_norm": 1.9090025764184533, "learning_rate": 1e-05, "loss": 0.4607, "step": 1638 }, { "epoch": 0.4723342939481268, "grad_norm": 1.5997617443798198, "learning_rate": 1e-05, "loss": 0.2916, "step": 1639 }, { "epoch": 0.47262247838616717, "grad_norm": 2.8454983572286365, "learning_rate": 1e-05, "loss": 0.4584, "step": 1640 }, { "epoch": 0.4729106628242075, "grad_norm": 2.415401984530564, "learning_rate": 1e-05, "loss": 0.4604, "step": 1641 }, { "epoch": 0.4731988472622478, "grad_norm": 2.0629566279765394, "learning_rate": 1e-05, "loss": 0.2785, "step": 1642 }, { "epoch": 0.47348703170028816, "grad_norm": 2.6809143836691964, "learning_rate": 1e-05, "loss": 0.4633, "step": 1643 }, { "epoch": 0.4737752161383285, "grad_norm": 1.6133613428462672, "learning_rate": 1e-05, "loss": 0.4627, "step": 1644 }, { "epoch": 0.47406340057636887, "grad_norm": 1.9295680584512247, "learning_rate": 1e-05, "loss": 0.2943, "step": 1645 }, { "epoch": 0.4743515850144092, "grad_norm": 2.5630263242961457, "learning_rate": 1e-05, "loss": 0.4583, "step": 1646 }, { "epoch": 0.47463976945244957, "grad_norm": 2.871829494714822, "learning_rate": 1e-05, "loss": 0.4664, "step": 1647 }, { "epoch": 0.4749279538904899, "grad_norm": 1.5701051155975592, "learning_rate": 1e-05, "loss": 0.2993, "step": 1648 }, { "epoch": 0.47521613832853027, "grad_norm": 2.1813422991508493, "learning_rate": 1e-05, "loss": 0.4104, "step": 1649 }, { "epoch": 0.4755043227665706, "grad_norm": 1.9274624400881828, "learning_rate": 1e-05, "loss": 0.465, "step": 1650 }, { "epoch": 0.47579250720461097, "grad_norm": 2.188434136825765, "learning_rate": 1e-05, "loss": 0.2711, "step": 1651 }, { "epoch": 0.4760806916426513, "grad_norm": 2.9606405106504594, "learning_rate": 1e-05, "loss": 0.4851, "step": 1652 }, { "epoch": 0.4763688760806916, "grad_norm": 1.8623058087282203, "learning_rate": 1e-05, "loss": 0.4622, "step": 1653 }, { "epoch": 0.47665706051873197, "grad_norm": 1.7217779197591614, "learning_rate": 1e-05, "loss": 0.2973, "step": 1654 }, { "epoch": 0.4769452449567723, "grad_norm": 4.474901292845818, "learning_rate": 1e-05, "loss": 0.4646, "step": 1655 }, { "epoch": 0.47723342939481267, "grad_norm": 1.77208449825262, "learning_rate": 1e-05, "loss": 0.4651, "step": 1656 }, { "epoch": 0.477521613832853, "grad_norm": 1.46297419544384, "learning_rate": 1e-05, "loss": 0.2826, "step": 1657 }, { "epoch": 0.47780979827089337, "grad_norm": 2.4772530667091313, "learning_rate": 1e-05, "loss": 0.4465, "step": 1658 }, { "epoch": 0.4780979827089337, "grad_norm": 2.231787991203313, "learning_rate": 1e-05, "loss": 0.4674, "step": 1659 }, { "epoch": 0.4783861671469741, "grad_norm": 2.0476087840540496, "learning_rate": 1e-05, "loss": 0.2902, "step": 1660 }, { "epoch": 0.4786743515850144, "grad_norm": 3.351809836897069, "learning_rate": 1e-05, "loss": 0.4738, "step": 1661 }, { "epoch": 0.4789625360230548, "grad_norm": 1.4230338456345113, "learning_rate": 1e-05, "loss": 0.4641, "step": 1662 }, { "epoch": 0.4792507204610951, "grad_norm": 2.4454093220425532, "learning_rate": 1e-05, "loss": 0.2929, "step": 1663 }, { "epoch": 0.4795389048991354, "grad_norm": 2.3791521392022865, "learning_rate": 1e-05, "loss": 0.4637, "step": 1664 }, { "epoch": 0.47982708933717577, "grad_norm": 1.746024248598935, "learning_rate": 1e-05, "loss": 0.4694, "step": 1665 }, { "epoch": 0.4801152737752161, "grad_norm": 2.4297254692858217, "learning_rate": 1e-05, "loss": 0.2814, "step": 1666 }, { "epoch": 0.4804034582132565, "grad_norm": 3.6694270809864418, "learning_rate": 1e-05, "loss": 0.4592, "step": 1667 }, { "epoch": 0.4806916426512968, "grad_norm": 1.4818977252200454, "learning_rate": 1e-05, "loss": 0.464, "step": 1668 }, { "epoch": 0.4809798270893372, "grad_norm": 2.401649865474904, "learning_rate": 1e-05, "loss": 0.2726, "step": 1669 }, { "epoch": 0.4812680115273775, "grad_norm": 2.8556667154383093, "learning_rate": 1e-05, "loss": 0.4213, "step": 1670 }, { "epoch": 0.4815561959654179, "grad_norm": 1.705405878147625, "learning_rate": 1e-05, "loss": 0.4577, "step": 1671 }, { "epoch": 0.48184438040345823, "grad_norm": 1.5484368382923146, "learning_rate": 1e-05, "loss": 0.286, "step": 1672 }, { "epoch": 0.4821325648414986, "grad_norm": 2.325960044809156, "learning_rate": 1e-05, "loss": 0.4655, "step": 1673 }, { "epoch": 0.48242074927953893, "grad_norm": 2.377847976610141, "learning_rate": 1e-05, "loss": 0.4642, "step": 1674 }, { "epoch": 0.4827089337175792, "grad_norm": 2.880226609097058, "learning_rate": 1e-05, "loss": 0.2817, "step": 1675 }, { "epoch": 0.4829971181556196, "grad_norm": 2.3133838506031754, "learning_rate": 1e-05, "loss": 0.4739, "step": 1676 }, { "epoch": 0.4832853025936599, "grad_norm": 1.5933475041603575, "learning_rate": 1e-05, "loss": 0.4686, "step": 1677 }, { "epoch": 0.4835734870317003, "grad_norm": 2.3560075097273594, "learning_rate": 1e-05, "loss": 0.2734, "step": 1678 }, { "epoch": 0.48386167146974063, "grad_norm": 3.2853984031850767, "learning_rate": 1e-05, "loss": 0.4232, "step": 1679 }, { "epoch": 0.484149855907781, "grad_norm": 1.7081589209357984, "learning_rate": 1e-05, "loss": 0.4635, "step": 1680 }, { "epoch": 0.48443804034582133, "grad_norm": 2.393075804247037, "learning_rate": 1e-05, "loss": 0.2859, "step": 1681 }, { "epoch": 0.4847262247838617, "grad_norm": 3.529738877729882, "learning_rate": 1e-05, "loss": 0.4492, "step": 1682 }, { "epoch": 0.48501440922190203, "grad_norm": 1.9093370333183477, "learning_rate": 1e-05, "loss": 0.45, "step": 1683 }, { "epoch": 0.4853025936599424, "grad_norm": 1.7693072755843469, "learning_rate": 1e-05, "loss": 0.263, "step": 1684 }, { "epoch": 0.48559077809798273, "grad_norm": 4.568922047885823, "learning_rate": 1e-05, "loss": 0.4148, "step": 1685 }, { "epoch": 0.48587896253602303, "grad_norm": 2.134908794525408, "learning_rate": 1e-05, "loss": 0.4702, "step": 1686 }, { "epoch": 0.4861671469740634, "grad_norm": 1.439007644118069, "learning_rate": 1e-05, "loss": 0.2985, "step": 1687 }, { "epoch": 0.48645533141210373, "grad_norm": 3.3774076129624273, "learning_rate": 1e-05, "loss": 0.4384, "step": 1688 }, { "epoch": 0.4867435158501441, "grad_norm": 1.915779414442965, "learning_rate": 1e-05, "loss": 0.4735, "step": 1689 }, { "epoch": 0.48703170028818443, "grad_norm": 1.4489762424037487, "learning_rate": 1e-05, "loss": 0.2735, "step": 1690 }, { "epoch": 0.4873198847262248, "grad_norm": 2.950035274296331, "learning_rate": 1e-05, "loss": 0.437, "step": 1691 }, { "epoch": 0.48760806916426513, "grad_norm": 1.8952923358414893, "learning_rate": 1e-05, "loss": 0.4547, "step": 1692 }, { "epoch": 0.4878962536023055, "grad_norm": 1.4305584083910485, "learning_rate": 1e-05, "loss": 0.2796, "step": 1693 }, { "epoch": 0.48818443804034584, "grad_norm": 2.4632837605887734, "learning_rate": 1e-05, "loss": 0.4292, "step": 1694 }, { "epoch": 0.4884726224783862, "grad_norm": 1.3955015462464664, "learning_rate": 1e-05, "loss": 0.4585, "step": 1695 }, { "epoch": 0.48876080691642654, "grad_norm": 2.2759160277281127, "learning_rate": 1e-05, "loss": 0.2604, "step": 1696 }, { "epoch": 0.48904899135446683, "grad_norm": 3.0743248862314223, "learning_rate": 1e-05, "loss": 0.4332, "step": 1697 }, { "epoch": 0.4893371757925072, "grad_norm": 1.4239389086894434, "learning_rate": 1e-05, "loss": 0.4519, "step": 1698 }, { "epoch": 0.48962536023054753, "grad_norm": 1.3716393538929104, "learning_rate": 1e-05, "loss": 0.2514, "step": 1699 }, { "epoch": 0.4899135446685879, "grad_norm": 2.363439394993871, "learning_rate": 1e-05, "loss": 0.4481, "step": 1700 }, { "epoch": 0.49020172910662824, "grad_norm": 1.9900927483142574, "learning_rate": 1e-05, "loss": 0.4683, "step": 1701 }, { "epoch": 0.4904899135446686, "grad_norm": 1.8185424866723199, "learning_rate": 1e-05, "loss": 0.2972, "step": 1702 }, { "epoch": 0.49077809798270894, "grad_norm": 3.3545570355044085, "learning_rate": 1e-05, "loss": 0.4683, "step": 1703 }, { "epoch": 0.4910662824207493, "grad_norm": 1.6559295493133095, "learning_rate": 1e-05, "loss": 0.4598, "step": 1704 }, { "epoch": 0.49135446685878964, "grad_norm": 1.4775431660851086, "learning_rate": 1e-05, "loss": 0.2662, "step": 1705 }, { "epoch": 0.49164265129683, "grad_norm": 3.486095927001384, "learning_rate": 1e-05, "loss": 0.4906, "step": 1706 }, { "epoch": 0.49193083573487034, "grad_norm": 2.219624772953442, "learning_rate": 1e-05, "loss": 0.4599, "step": 1707 }, { "epoch": 0.49221902017291064, "grad_norm": 1.9844505305072102, "learning_rate": 1e-05, "loss": 0.2641, "step": 1708 }, { "epoch": 0.492507204610951, "grad_norm": 2.820281651509339, "learning_rate": 1e-05, "loss": 0.4751, "step": 1709 }, { "epoch": 0.49279538904899134, "grad_norm": 1.8429376323004725, "learning_rate": 1e-05, "loss": 0.4615, "step": 1710 }, { "epoch": 0.4930835734870317, "grad_norm": 1.876894295065116, "learning_rate": 1e-05, "loss": 0.2637, "step": 1711 }, { "epoch": 0.49337175792507204, "grad_norm": 3.6828539449790103, "learning_rate": 1e-05, "loss": 0.4508, "step": 1712 }, { "epoch": 0.4936599423631124, "grad_norm": 1.49070007978551, "learning_rate": 1e-05, "loss": 0.4665, "step": 1713 }, { "epoch": 0.49394812680115274, "grad_norm": 1.8332686252713897, "learning_rate": 1e-05, "loss": 0.2685, "step": 1714 }, { "epoch": 0.4942363112391931, "grad_norm": 3.653039605835981, "learning_rate": 1e-05, "loss": 0.4509, "step": 1715 }, { "epoch": 0.49452449567723344, "grad_norm": 2.2361933476641727, "learning_rate": 1e-05, "loss": 0.4678, "step": 1716 }, { "epoch": 0.4948126801152738, "grad_norm": 2.2198315477674417, "learning_rate": 1e-05, "loss": 0.2864, "step": 1717 }, { "epoch": 0.49510086455331415, "grad_norm": 2.9662040776422613, "learning_rate": 1e-05, "loss": 0.4596, "step": 1718 }, { "epoch": 0.49538904899135444, "grad_norm": 1.8533439742640203, "learning_rate": 1e-05, "loss": 0.4601, "step": 1719 }, { "epoch": 0.4956772334293948, "grad_norm": 2.05752974142649, "learning_rate": 1e-05, "loss": 0.2892, "step": 1720 }, { "epoch": 0.49596541786743514, "grad_norm": 2.5147899712900736, "learning_rate": 1e-05, "loss": 0.4467, "step": 1721 }, { "epoch": 0.4962536023054755, "grad_norm": 1.4329665661640703, "learning_rate": 1e-05, "loss": 0.4569, "step": 1722 }, { "epoch": 0.49654178674351584, "grad_norm": 1.9863338498032903, "learning_rate": 1e-05, "loss": 0.2768, "step": 1723 }, { "epoch": 0.4968299711815562, "grad_norm": 2.7689961736244513, "learning_rate": 1e-05, "loss": 0.4393, "step": 1724 }, { "epoch": 0.49711815561959655, "grad_norm": 1.863142892863334, "learning_rate": 1e-05, "loss": 0.4576, "step": 1725 }, { "epoch": 0.4974063400576369, "grad_norm": 1.749129294111609, "learning_rate": 1e-05, "loss": 0.2481, "step": 1726 }, { "epoch": 0.49769452449567725, "grad_norm": 2.4098307488706117, "learning_rate": 1e-05, "loss": 0.4352, "step": 1727 }, { "epoch": 0.4979827089337176, "grad_norm": 1.7004384919576656, "learning_rate": 1e-05, "loss": 0.4631, "step": 1728 }, { "epoch": 0.49827089337175795, "grad_norm": 1.6741308399940007, "learning_rate": 1e-05, "loss": 0.2644, "step": 1729 }, { "epoch": 0.49855907780979825, "grad_norm": 2.2932672534152436, "learning_rate": 1e-05, "loss": 0.4367, "step": 1730 }, { "epoch": 0.4988472622478386, "grad_norm": 2.6349075799650565, "learning_rate": 1e-05, "loss": 0.4599, "step": 1731 }, { "epoch": 0.49913544668587895, "grad_norm": 1.8889506861333187, "learning_rate": 1e-05, "loss": 0.2671, "step": 1732 }, { "epoch": 0.4994236311239193, "grad_norm": 2.6226440853671735, "learning_rate": 1e-05, "loss": 0.465, "step": 1733 }, { "epoch": 0.49971181556195965, "grad_norm": 1.7201450504469826, "learning_rate": 1e-05, "loss": 0.4531, "step": 1734 }, { "epoch": 0.5, "grad_norm": 2.2914202968030986, "learning_rate": 1e-05, "loss": 0.2772, "step": 1735 }, { "epoch": 0.5002881844380404, "grad_norm": 2.50152200887206, "learning_rate": 1e-05, "loss": 0.4495, "step": 1736 }, { "epoch": 0.5005763688760807, "grad_norm": 1.693806283768107, "learning_rate": 1e-05, "loss": 0.4638, "step": 1737 }, { "epoch": 0.500864553314121, "grad_norm": 1.9028704339010702, "learning_rate": 1e-05, "loss": 0.3013, "step": 1738 }, { "epoch": 0.5011527377521614, "grad_norm": 2.5933183846887995, "learning_rate": 1e-05, "loss": 0.4828, "step": 1739 }, { "epoch": 0.5014409221902018, "grad_norm": 2.0412542395277455, "learning_rate": 1e-05, "loss": 0.4587, "step": 1740 }, { "epoch": 0.5017291066282421, "grad_norm": 1.9856149329061437, "learning_rate": 1e-05, "loss": 0.278, "step": 1741 }, { "epoch": 0.5020172910662825, "grad_norm": 2.4016386267429217, "learning_rate": 1e-05, "loss": 0.4721, "step": 1742 }, { "epoch": 0.5023054755043228, "grad_norm": 2.3613869701064427, "learning_rate": 1e-05, "loss": 0.4709, "step": 1743 }, { "epoch": 0.5025936599423632, "grad_norm": 1.8895381195885375, "learning_rate": 1e-05, "loss": 0.2667, "step": 1744 }, { "epoch": 0.5028818443804035, "grad_norm": 2.519056644274537, "learning_rate": 1e-05, "loss": 0.4173, "step": 1745 }, { "epoch": 0.5031700288184437, "grad_norm": 1.597784148194564, "learning_rate": 1e-05, "loss": 0.4478, "step": 1746 }, { "epoch": 0.5034582132564841, "grad_norm": 1.8673531027230266, "learning_rate": 1e-05, "loss": 0.264, "step": 1747 }, { "epoch": 0.5037463976945245, "grad_norm": 2.9027343527283103, "learning_rate": 1e-05, "loss": 0.4356, "step": 1748 }, { "epoch": 0.5040345821325648, "grad_norm": 3.034671484976285, "learning_rate": 1e-05, "loss": 0.4664, "step": 1749 }, { "epoch": 0.5043227665706052, "grad_norm": 2.7764727211021434, "learning_rate": 1e-05, "loss": 0.2722, "step": 1750 }, { "epoch": 0.5043227665706052, "eval_websight_new_IoU": 0.019636476412415504, "eval_websight_new_MAE_x": 286.760009765625, "eval_websight_new_MAE_y": 594.97998046875, "eval_websight_new_coord_probability": 0.7459741830825806, "eval_websight_new_correct_token_probability": 0.0011280778562650084, "eval_websight_new_inside_bbox": 0.019999999552965164, "eval_websight_new_loss": 0.2900213897228241, "eval_websight_new_range_100_probability": 0.09146662056446075, "eval_websight_new_range_20_probability": 0.015594979748129845, "eval_websight_new_range_5_probability": 0.0030590335372835398, "eval_websight_new_runtime": 33.7773, "eval_websight_new_samples_per_second": 1.48, "eval_websight_new_steps_per_second": 0.207, "step": 1750 }, { "epoch": 0.5043227665706052, "eval_seeclick_IoU": 0.034775398671627045, "eval_seeclick_MAE_x": 473.1000061035156, "eval_seeclick_MAE_y": 505.5199890136719, "eval_seeclick_coord_probability": 0.7473167181015015, "eval_seeclick_correct_token_probability": 0.0006066637579351664, "eval_seeclick_inside_bbox": 0.05999999865889549, "eval_seeclick_loss": 0.38515812158584595, "eval_seeclick_range_100_probability": 0.13240179419517517, "eval_seeclick_range_20_probability": 0.05072963237762451, "eval_seeclick_range_5_probability": 0.007617821916937828, "eval_seeclick_runtime": 42.4492, "eval_seeclick_samples_per_second": 1.178, "eval_seeclick_steps_per_second": 0.165, "step": 1750 }, { "epoch": 0.5043227665706052, "eval_icons_IoU": 0.0025072256103157997, "eval_icons_MAE_x": 473.5400085449219, "eval_icons_MAE_y": 546.8800048828125, "eval_icons_coord_probability": 0.7476227283477783, "eval_icons_correct_token_probability": 0.00905266497284174, "eval_icons_inside_bbox": 0.0, "eval_icons_loss": 0.4429156482219696, "eval_icons_range_100_probability": 0.276961088180542, "eval_icons_range_20_probability": 0.09574487060308456, "eval_icons_range_5_probability": 0.03115815669298172, "eval_icons_runtime": 43.2189, "eval_icons_samples_per_second": 1.157, "eval_icons_steps_per_second": 0.162, "step": 1750 }, { "epoch": 0.5046109510086455, "grad_norm": 3.9831774379628264, "learning_rate": 1e-05, "loss": 0.4618, "step": 1751 }, { "epoch": 0.5048991354466859, "grad_norm": 1.4374301573776382, "learning_rate": 1e-05, "loss": 0.4568, "step": 1752 }, { "epoch": 0.5051873198847262, "grad_norm": 1.8921007186490801, "learning_rate": 1e-05, "loss": 0.2785, "step": 1753 }, { "epoch": 0.5054755043227666, "grad_norm": 2.615752425693824, "learning_rate": 1e-05, "loss": 0.477, "step": 1754 }, { "epoch": 0.5057636887608069, "grad_norm": 1.533452814309213, "learning_rate": 1e-05, "loss": 0.4545, "step": 1755 }, { "epoch": 0.5060518731988473, "grad_norm": 1.9957275008919058, "learning_rate": 1e-05, "loss": 0.2784, "step": 1756 }, { "epoch": 0.5063400576368876, "grad_norm": 3.1653940246653987, "learning_rate": 1e-05, "loss": 0.4532, "step": 1757 }, { "epoch": 0.506628242074928, "grad_norm": 1.407081682029088, "learning_rate": 1e-05, "loss": 0.4559, "step": 1758 }, { "epoch": 0.5069164265129683, "grad_norm": 1.5261723664185132, "learning_rate": 1e-05, "loss": 0.2775, "step": 1759 }, { "epoch": 0.5072046109510087, "grad_norm": 2.8459742883006585, "learning_rate": 1e-05, "loss": 0.4497, "step": 1760 }, { "epoch": 0.507492795389049, "grad_norm": 1.52626873226206, "learning_rate": 1e-05, "loss": 0.4485, "step": 1761 }, { "epoch": 0.5077809798270894, "grad_norm": 2.120607391919743, "learning_rate": 1e-05, "loss": 0.2781, "step": 1762 }, { "epoch": 0.5080691642651297, "grad_norm": 2.4977431110147483, "learning_rate": 1e-05, "loss": 0.4563, "step": 1763 }, { "epoch": 0.5083573487031701, "grad_norm": 2.0923075056563833, "learning_rate": 1e-05, "loss": 0.4692, "step": 1764 }, { "epoch": 0.5086455331412104, "grad_norm": 1.6577734390908576, "learning_rate": 1e-05, "loss": 0.2685, "step": 1765 }, { "epoch": 0.5089337175792508, "grad_norm": 3.4894389660396223, "learning_rate": 1e-05, "loss": 0.4617, "step": 1766 }, { "epoch": 0.5092219020172911, "grad_norm": 1.7791014605602602, "learning_rate": 1e-05, "loss": 0.4646, "step": 1767 }, { "epoch": 0.5095100864553314, "grad_norm": 1.7600775080832534, "learning_rate": 1e-05, "loss": 0.2607, "step": 1768 }, { "epoch": 0.5097982708933717, "grad_norm": 3.635526592272463, "learning_rate": 1e-05, "loss": 0.4646, "step": 1769 }, { "epoch": 0.5100864553314121, "grad_norm": 1.4535745053954856, "learning_rate": 1e-05, "loss": 0.4504, "step": 1770 }, { "epoch": 0.5103746397694524, "grad_norm": 1.9015723045727277, "learning_rate": 1e-05, "loss": 0.2665, "step": 1771 }, { "epoch": 0.5106628242074928, "grad_norm": 2.7503315131178994, "learning_rate": 1e-05, "loss": 0.445, "step": 1772 }, { "epoch": 0.5109510086455331, "grad_norm": 1.7947410749054482, "learning_rate": 1e-05, "loss": 0.4552, "step": 1773 }, { "epoch": 0.5112391930835735, "grad_norm": 2.0285048593527675, "learning_rate": 1e-05, "loss": 0.2817, "step": 1774 }, { "epoch": 0.5115273775216138, "grad_norm": 4.085655591407628, "learning_rate": 1e-05, "loss": 0.4605, "step": 1775 }, { "epoch": 0.5118155619596542, "grad_norm": 1.409100641755321, "learning_rate": 1e-05, "loss": 0.4568, "step": 1776 }, { "epoch": 0.5121037463976945, "grad_norm": 2.450601632694455, "learning_rate": 1e-05, "loss": 0.2464, "step": 1777 }, { "epoch": 0.5123919308357349, "grad_norm": 2.6694118572635075, "learning_rate": 1e-05, "loss": 0.4843, "step": 1778 }, { "epoch": 0.5126801152737752, "grad_norm": 2.68944159707934, "learning_rate": 1e-05, "loss": 0.4542, "step": 1779 }, { "epoch": 0.5129682997118156, "grad_norm": 1.4155636999897319, "learning_rate": 1e-05, "loss": 0.2684, "step": 1780 }, { "epoch": 0.5132564841498559, "grad_norm": 3.014353805737626, "learning_rate": 1e-05, "loss": 0.4892, "step": 1781 }, { "epoch": 0.5135446685878963, "grad_norm": 1.4594450185730636, "learning_rate": 1e-05, "loss": 0.4457, "step": 1782 }, { "epoch": 0.5138328530259366, "grad_norm": 2.008615649097747, "learning_rate": 1e-05, "loss": 0.2421, "step": 1783 }, { "epoch": 0.514121037463977, "grad_norm": 3.3669893461047993, "learning_rate": 1e-05, "loss": 0.4537, "step": 1784 }, { "epoch": 0.5144092219020173, "grad_norm": 2.8345861052522427, "learning_rate": 1e-05, "loss": 0.4673, "step": 1785 }, { "epoch": 0.5146974063400577, "grad_norm": 1.323601890023609, "learning_rate": 1e-05, "loss": 0.2613, "step": 1786 }, { "epoch": 0.514985590778098, "grad_norm": 2.74391568395978, "learning_rate": 1e-05, "loss": 0.464, "step": 1787 }, { "epoch": 0.5152737752161384, "grad_norm": 1.5384663730957742, "learning_rate": 1e-05, "loss": 0.4502, "step": 1788 }, { "epoch": 0.5155619596541787, "grad_norm": 2.276530745494601, "learning_rate": 1e-05, "loss": 0.2797, "step": 1789 }, { "epoch": 0.515850144092219, "grad_norm": 3.2809096257420007, "learning_rate": 1e-05, "loss": 0.4668, "step": 1790 }, { "epoch": 0.5161383285302593, "grad_norm": 2.230621188973937, "learning_rate": 1e-05, "loss": 0.4692, "step": 1791 }, { "epoch": 0.5164265129682997, "grad_norm": 1.3677062826287025, "learning_rate": 1e-05, "loss": 0.267, "step": 1792 }, { "epoch": 0.51671469740634, "grad_norm": 2.6172921544831285, "learning_rate": 1e-05, "loss": 0.4505, "step": 1793 }, { "epoch": 0.5170028818443804, "grad_norm": 1.4649113503607931, "learning_rate": 1e-05, "loss": 0.4589, "step": 1794 }, { "epoch": 0.5172910662824207, "grad_norm": 2.683917428362708, "learning_rate": 1e-05, "loss": 0.2774, "step": 1795 }, { "epoch": 0.5175792507204611, "grad_norm": 3.110624395145058, "learning_rate": 1e-05, "loss": 0.4531, "step": 1796 }, { "epoch": 0.5178674351585014, "grad_norm": 1.8149679816302997, "learning_rate": 1e-05, "loss": 0.4653, "step": 1797 }, { "epoch": 0.5181556195965418, "grad_norm": 1.5170431570832181, "learning_rate": 1e-05, "loss": 0.258, "step": 1798 }, { "epoch": 0.5184438040345821, "grad_norm": 3.2083316155370487, "learning_rate": 1e-05, "loss": 0.4573, "step": 1799 }, { "epoch": 0.5187319884726225, "grad_norm": 2.014435392735327, "learning_rate": 1e-05, "loss": 0.4509, "step": 1800 }, { "epoch": 0.5190201729106628, "grad_norm": 2.700097817195559, "learning_rate": 1e-05, "loss": 0.2678, "step": 1801 }, { "epoch": 0.5193083573487032, "grad_norm": 3.5033374580324566, "learning_rate": 1e-05, "loss": 0.4434, "step": 1802 }, { "epoch": 0.5195965417867435, "grad_norm": 1.4833570072957702, "learning_rate": 1e-05, "loss": 0.4629, "step": 1803 }, { "epoch": 0.5198847262247839, "grad_norm": 2.2591788072257843, "learning_rate": 1e-05, "loss": 0.2993, "step": 1804 }, { "epoch": 0.5201729106628242, "grad_norm": 3.3331176425439524, "learning_rate": 1e-05, "loss": 0.4366, "step": 1805 }, { "epoch": 0.5204610951008646, "grad_norm": 1.8628877406825821, "learning_rate": 1e-05, "loss": 0.4607, "step": 1806 }, { "epoch": 0.5207492795389049, "grad_norm": 1.8380156721674201, "learning_rate": 1e-05, "loss": 0.2554, "step": 1807 }, { "epoch": 0.5210374639769453, "grad_norm": 3.0873032711346275, "learning_rate": 1e-05, "loss": 0.4712, "step": 1808 }, { "epoch": 0.5213256484149856, "grad_norm": 1.6313365237093287, "learning_rate": 1e-05, "loss": 0.4621, "step": 1809 }, { "epoch": 0.521613832853026, "grad_norm": 2.4406031488855664, "learning_rate": 1e-05, "loss": 0.2919, "step": 1810 }, { "epoch": 0.5219020172910663, "grad_norm": 2.741731552872433, "learning_rate": 1e-05, "loss": 0.4344, "step": 1811 }, { "epoch": 0.5221902017291066, "grad_norm": 1.5711327264546315, "learning_rate": 1e-05, "loss": 0.4532, "step": 1812 }, { "epoch": 0.5224783861671469, "grad_norm": 1.6069010910310144, "learning_rate": 1e-05, "loss": 0.2954, "step": 1813 }, { "epoch": 0.5227665706051873, "grad_norm": 2.448133217933075, "learning_rate": 1e-05, "loss": 0.4851, "step": 1814 }, { "epoch": 0.5230547550432276, "grad_norm": 1.9430375921496603, "learning_rate": 1e-05, "loss": 0.459, "step": 1815 }, { "epoch": 0.523342939481268, "grad_norm": 2.8756318716673293, "learning_rate": 1e-05, "loss": 0.2794, "step": 1816 }, { "epoch": 0.5236311239193083, "grad_norm": 2.933961020242428, "learning_rate": 1e-05, "loss": 0.4325, "step": 1817 }, { "epoch": 0.5239193083573487, "grad_norm": 1.5475470865668899, "learning_rate": 1e-05, "loss": 0.4658, "step": 1818 }, { "epoch": 0.524207492795389, "grad_norm": 2.0619442989106505, "learning_rate": 1e-05, "loss": 0.2823, "step": 1819 }, { "epoch": 0.5244956772334294, "grad_norm": 2.7536751220796605, "learning_rate": 1e-05, "loss": 0.4881, "step": 1820 }, { "epoch": 0.5247838616714697, "grad_norm": 1.8208384751595372, "learning_rate": 1e-05, "loss": 0.4595, "step": 1821 }, { "epoch": 0.5250720461095101, "grad_norm": 2.217366030697359, "learning_rate": 1e-05, "loss": 0.27, "step": 1822 }, { "epoch": 0.5253602305475504, "grad_norm": 3.216314333500646, "learning_rate": 1e-05, "loss": 0.4161, "step": 1823 }, { "epoch": 0.5256484149855908, "grad_norm": 2.680423152848142, "learning_rate": 1e-05, "loss": 0.4627, "step": 1824 }, { "epoch": 0.5259365994236311, "grad_norm": 1.8212022253223754, "learning_rate": 1e-05, "loss": 0.2813, "step": 1825 }, { "epoch": 0.5262247838616715, "grad_norm": 3.6904326994867915, "learning_rate": 1e-05, "loss": 0.46, "step": 1826 }, { "epoch": 0.5265129682997118, "grad_norm": 1.874899690129343, "learning_rate": 1e-05, "loss": 0.4453, "step": 1827 }, { "epoch": 0.5268011527377522, "grad_norm": 1.5384880091568887, "learning_rate": 1e-05, "loss": 0.258, "step": 1828 }, { "epoch": 0.5270893371757925, "grad_norm": 2.8060985419907145, "learning_rate": 1e-05, "loss": 0.483, "step": 1829 }, { "epoch": 0.5273775216138329, "grad_norm": 1.9478776801876367, "learning_rate": 1e-05, "loss": 0.4598, "step": 1830 }, { "epoch": 0.5276657060518732, "grad_norm": 1.780793099901807, "learning_rate": 1e-05, "loss": 0.2834, "step": 1831 }, { "epoch": 0.5279538904899136, "grad_norm": 3.0492293138847093, "learning_rate": 1e-05, "loss": 0.4545, "step": 1832 }, { "epoch": 0.5282420749279539, "grad_norm": 1.4573449451227571, "learning_rate": 1e-05, "loss": 0.4484, "step": 1833 }, { "epoch": 0.5285302593659942, "grad_norm": 2.213340764443074, "learning_rate": 1e-05, "loss": 0.2808, "step": 1834 }, { "epoch": 0.5288184438040345, "grad_norm": 2.877377727558519, "learning_rate": 1e-05, "loss": 0.4732, "step": 1835 }, { "epoch": 0.5291066282420749, "grad_norm": 1.6714668937969217, "learning_rate": 1e-05, "loss": 0.4585, "step": 1836 }, { "epoch": 0.5293948126801152, "grad_norm": 1.9132058027926757, "learning_rate": 1e-05, "loss": 0.2797, "step": 1837 }, { "epoch": 0.5296829971181556, "grad_norm": 3.1423187466760276, "learning_rate": 1e-05, "loss": 0.4523, "step": 1838 }, { "epoch": 0.5299711815561959, "grad_norm": 1.6421478176952211, "learning_rate": 1e-05, "loss": 0.4609, "step": 1839 }, { "epoch": 0.5302593659942363, "grad_norm": 2.285409906363161, "learning_rate": 1e-05, "loss": 0.2752, "step": 1840 }, { "epoch": 0.5305475504322766, "grad_norm": 3.089328294282285, "learning_rate": 1e-05, "loss": 0.4611, "step": 1841 }, { "epoch": 0.530835734870317, "grad_norm": 2.53730982001956, "learning_rate": 1e-05, "loss": 0.4734, "step": 1842 }, { "epoch": 0.5311239193083573, "grad_norm": 1.893790387928652, "learning_rate": 1e-05, "loss": 0.2447, "step": 1843 }, { "epoch": 0.5314121037463977, "grad_norm": 3.015291369506835, "learning_rate": 1e-05, "loss": 0.4632, "step": 1844 }, { "epoch": 0.531700288184438, "grad_norm": 1.8123259306752908, "learning_rate": 1e-05, "loss": 0.4598, "step": 1845 }, { "epoch": 0.5319884726224784, "grad_norm": 1.4412994338870306, "learning_rate": 1e-05, "loss": 0.2632, "step": 1846 }, { "epoch": 0.5322766570605187, "grad_norm": 3.1533731896340558, "learning_rate": 1e-05, "loss": 0.4111, "step": 1847 }, { "epoch": 0.5325648414985591, "grad_norm": 1.7342875541722724, "learning_rate": 1e-05, "loss": 0.4592, "step": 1848 }, { "epoch": 0.5328530259365994, "grad_norm": 1.8922708143846898, "learning_rate": 1e-05, "loss": 0.2664, "step": 1849 }, { "epoch": 0.5331412103746398, "grad_norm": 2.4403878437728093, "learning_rate": 1e-05, "loss": 0.431, "step": 1850 }, { "epoch": 0.5334293948126801, "grad_norm": 1.5409715979370118, "learning_rate": 1e-05, "loss": 0.4519, "step": 1851 }, { "epoch": 0.5337175792507205, "grad_norm": 1.819497807365805, "learning_rate": 1e-05, "loss": 0.2755, "step": 1852 }, { "epoch": 0.5340057636887608, "grad_norm": 3.2243638315771785, "learning_rate": 1e-05, "loss": 0.4747, "step": 1853 }, { "epoch": 0.5342939481268012, "grad_norm": 1.6823526981268717, "learning_rate": 1e-05, "loss": 0.4483, "step": 1854 }, { "epoch": 0.5345821325648416, "grad_norm": 1.4946564867384682, "learning_rate": 1e-05, "loss": 0.2604, "step": 1855 }, { "epoch": 0.5348703170028818, "grad_norm": 2.9848690934637467, "learning_rate": 1e-05, "loss": 0.4461, "step": 1856 }, { "epoch": 0.5351585014409221, "grad_norm": 2.1098009369876443, "learning_rate": 1e-05, "loss": 0.4643, "step": 1857 }, { "epoch": 0.5354466858789625, "grad_norm": 1.2031374446979346, "learning_rate": 1e-05, "loss": 0.2603, "step": 1858 }, { "epoch": 0.5357348703170028, "grad_norm": 2.6112357037271328, "learning_rate": 1e-05, "loss": 0.4528, "step": 1859 }, { "epoch": 0.5360230547550432, "grad_norm": 1.584410884861249, "learning_rate": 1e-05, "loss": 0.4434, "step": 1860 }, { "epoch": 0.5363112391930835, "grad_norm": 1.9740859563312552, "learning_rate": 1e-05, "loss": 0.2508, "step": 1861 }, { "epoch": 0.5365994236311239, "grad_norm": 2.9887653882763687, "learning_rate": 1e-05, "loss": 0.4589, "step": 1862 }, { "epoch": 0.5368876080691642, "grad_norm": 1.8961156323859274, "learning_rate": 1e-05, "loss": 0.4487, "step": 1863 }, { "epoch": 0.5371757925072046, "grad_norm": 1.459533941523443, "learning_rate": 1e-05, "loss": 0.2757, "step": 1864 }, { "epoch": 0.537463976945245, "grad_norm": 2.908063744858589, "learning_rate": 1e-05, "loss": 0.5206, "step": 1865 }, { "epoch": 0.5377521613832853, "grad_norm": 1.9124436660439734, "learning_rate": 1e-05, "loss": 0.4649, "step": 1866 }, { "epoch": 0.5380403458213256, "grad_norm": 2.0064850758455224, "learning_rate": 1e-05, "loss": 0.2519, "step": 1867 }, { "epoch": 0.538328530259366, "grad_norm": 3.333105967232054, "learning_rate": 1e-05, "loss": 0.4499, "step": 1868 }, { "epoch": 0.5386167146974064, "grad_norm": 2.009065187974275, "learning_rate": 1e-05, "loss": 0.4622, "step": 1869 }, { "epoch": 0.5389048991354467, "grad_norm": 1.8532428290322112, "learning_rate": 1e-05, "loss": 0.2713, "step": 1870 }, { "epoch": 0.539193083573487, "grad_norm": 3.02397352823615, "learning_rate": 1e-05, "loss": 0.4841, "step": 1871 }, { "epoch": 0.5394812680115274, "grad_norm": 1.5324680567007694, "learning_rate": 1e-05, "loss": 0.4499, "step": 1872 }, { "epoch": 0.5397694524495678, "grad_norm": 2.0931095142787752, "learning_rate": 1e-05, "loss": 0.2645, "step": 1873 }, { "epoch": 0.5400576368876081, "grad_norm": 3.376271359266738, "learning_rate": 1e-05, "loss": 0.46, "step": 1874 }, { "epoch": 0.5403458213256485, "grad_norm": 1.8229770106171024, "learning_rate": 1e-05, "loss": 0.4527, "step": 1875 }, { "epoch": 0.5406340057636888, "grad_norm": 1.8368714849623498, "learning_rate": 1e-05, "loss": 0.2622, "step": 1876 }, { "epoch": 0.5409221902017292, "grad_norm": 2.933593558121684, "learning_rate": 1e-05, "loss": 0.444, "step": 1877 }, { "epoch": 0.5412103746397694, "grad_norm": 1.5578430127564136, "learning_rate": 1e-05, "loss": 0.4569, "step": 1878 }, { "epoch": 0.5414985590778097, "grad_norm": 2.089309179173947, "learning_rate": 1e-05, "loss": 0.2708, "step": 1879 }, { "epoch": 0.5417867435158501, "grad_norm": 2.512466219808063, "learning_rate": 1e-05, "loss": 0.4623, "step": 1880 }, { "epoch": 0.5420749279538905, "grad_norm": 2.0389629009086105, "learning_rate": 1e-05, "loss": 0.4448, "step": 1881 }, { "epoch": 0.5423631123919308, "grad_norm": 1.9030719199770536, "learning_rate": 1e-05, "loss": 0.2624, "step": 1882 }, { "epoch": 0.5426512968299712, "grad_norm": 2.307530192752208, "learning_rate": 1e-05, "loss": 0.4589, "step": 1883 }, { "epoch": 0.5429394812680115, "grad_norm": 1.629736150698527, "learning_rate": 1e-05, "loss": 0.4514, "step": 1884 }, { "epoch": 0.5432276657060519, "grad_norm": 2.4088248450727927, "learning_rate": 1e-05, "loss": 0.2347, "step": 1885 }, { "epoch": 0.5435158501440922, "grad_norm": 3.1431509643435582, "learning_rate": 1e-05, "loss": 0.4812, "step": 1886 }, { "epoch": 0.5438040345821326, "grad_norm": 1.8826733706424852, "learning_rate": 1e-05, "loss": 0.4583, "step": 1887 }, { "epoch": 0.5440922190201729, "grad_norm": 2.5953362787037273, "learning_rate": 1e-05, "loss": 0.278, "step": 1888 }, { "epoch": 0.5443804034582133, "grad_norm": 2.687749652434169, "learning_rate": 1e-05, "loss": 0.4513, "step": 1889 }, { "epoch": 0.5446685878962536, "grad_norm": 1.8653911290409602, "learning_rate": 1e-05, "loss": 0.4465, "step": 1890 }, { "epoch": 0.544956772334294, "grad_norm": 2.2593739986321064, "learning_rate": 1e-05, "loss": 0.2572, "step": 1891 }, { "epoch": 0.5452449567723343, "grad_norm": 2.5782587798696697, "learning_rate": 1e-05, "loss": 0.453, "step": 1892 }, { "epoch": 0.5455331412103747, "grad_norm": 2.8581536633741833, "learning_rate": 1e-05, "loss": 0.4571, "step": 1893 }, { "epoch": 0.545821325648415, "grad_norm": 3.2311140930376188, "learning_rate": 1e-05, "loss": 0.2625, "step": 1894 }, { "epoch": 0.5461095100864554, "grad_norm": 3.776262194209603, "learning_rate": 1e-05, "loss": 0.4864, "step": 1895 }, { "epoch": 0.5463976945244957, "grad_norm": 1.913305405328957, "learning_rate": 1e-05, "loss": 0.4489, "step": 1896 }, { "epoch": 0.5466858789625361, "grad_norm": 2.293688537709464, "learning_rate": 1e-05, "loss": 0.2809, "step": 1897 }, { "epoch": 0.5469740634005764, "grad_norm": 2.850231559640113, "learning_rate": 1e-05, "loss": 0.4485, "step": 1898 }, { "epoch": 0.5472622478386168, "grad_norm": 1.7905016688720539, "learning_rate": 1e-05, "loss": 0.4526, "step": 1899 }, { "epoch": 0.547550432276657, "grad_norm": 2.081835389558327, "learning_rate": 1e-05, "loss": 0.2573, "step": 1900 }, { "epoch": 0.5478386167146974, "grad_norm": 3.4872844874424436, "learning_rate": 1e-05, "loss": 0.451, "step": 1901 }, { "epoch": 0.5481268011527377, "grad_norm": 2.1065416879642966, "learning_rate": 1e-05, "loss": 0.456, "step": 1902 }, { "epoch": 0.5484149855907781, "grad_norm": 2.0926646314436694, "learning_rate": 1e-05, "loss": 0.2693, "step": 1903 }, { "epoch": 0.5487031700288184, "grad_norm": 2.245442164994045, "learning_rate": 1e-05, "loss": 0.4263, "step": 1904 }, { "epoch": 0.5489913544668588, "grad_norm": 2.2731655613094084, "learning_rate": 1e-05, "loss": 0.4583, "step": 1905 }, { "epoch": 0.5492795389048991, "grad_norm": 2.285044456313453, "learning_rate": 1e-05, "loss": 0.2473, "step": 1906 }, { "epoch": 0.5495677233429395, "grad_norm": 3.9068601259022016, "learning_rate": 1e-05, "loss": 0.429, "step": 1907 }, { "epoch": 0.5498559077809798, "grad_norm": 1.82427426832822, "learning_rate": 1e-05, "loss": 0.4535, "step": 1908 }, { "epoch": 0.5501440922190202, "grad_norm": 1.7584421286956202, "learning_rate": 1e-05, "loss": 0.2741, "step": 1909 }, { "epoch": 0.5504322766570605, "grad_norm": 2.6504193638075457, "learning_rate": 1e-05, "loss": 0.441, "step": 1910 }, { "epoch": 0.5507204610951009, "grad_norm": 1.9922239506499178, "learning_rate": 1e-05, "loss": 0.4509, "step": 1911 }, { "epoch": 0.5510086455331412, "grad_norm": 2.6730502047285776, "learning_rate": 1e-05, "loss": 0.2758, "step": 1912 }, { "epoch": 0.5512968299711816, "grad_norm": 3.401251289453041, "learning_rate": 1e-05, "loss": 0.475, "step": 1913 }, { "epoch": 0.5515850144092219, "grad_norm": 2.1860798214734856, "learning_rate": 1e-05, "loss": 0.4443, "step": 1914 }, { "epoch": 0.5518731988472623, "grad_norm": 1.4306611831603044, "learning_rate": 1e-05, "loss": 0.2711, "step": 1915 }, { "epoch": 0.5521613832853026, "grad_norm": 2.882775587691361, "learning_rate": 1e-05, "loss": 0.4758, "step": 1916 }, { "epoch": 0.552449567723343, "grad_norm": 2.1634947836778182, "learning_rate": 1e-05, "loss": 0.4501, "step": 1917 }, { "epoch": 0.5527377521613833, "grad_norm": 1.8622904481100178, "learning_rate": 1e-05, "loss": 0.2574, "step": 1918 }, { "epoch": 0.5530259365994237, "grad_norm": 2.9059816826291085, "learning_rate": 1e-05, "loss": 0.4656, "step": 1919 }, { "epoch": 0.553314121037464, "grad_norm": 1.6339918829837667, "learning_rate": 1e-05, "loss": 0.4526, "step": 1920 }, { "epoch": 0.5536023054755044, "grad_norm": 1.8584706483869857, "learning_rate": 1e-05, "loss": 0.2711, "step": 1921 }, { "epoch": 0.5538904899135446, "grad_norm": 3.372951866244728, "learning_rate": 1e-05, "loss": 0.4183, "step": 1922 }, { "epoch": 0.554178674351585, "grad_norm": 1.8625918927995937, "learning_rate": 1e-05, "loss": 0.4487, "step": 1923 }, { "epoch": 0.5544668587896253, "grad_norm": 2.0076964461445876, "learning_rate": 1e-05, "loss": 0.2611, "step": 1924 }, { "epoch": 0.5547550432276657, "grad_norm": 3.659663255000959, "learning_rate": 1e-05, "loss": 0.4282, "step": 1925 }, { "epoch": 0.555043227665706, "grad_norm": 3.1758154607924904, "learning_rate": 1e-05, "loss": 0.4535, "step": 1926 }, { "epoch": 0.5553314121037464, "grad_norm": 1.8110667318905647, "learning_rate": 1e-05, "loss": 0.2486, "step": 1927 }, { "epoch": 0.5556195965417867, "grad_norm": 3.191348287919045, "learning_rate": 1e-05, "loss": 0.4748, "step": 1928 }, { "epoch": 0.5559077809798271, "grad_norm": 2.223725809958116, "learning_rate": 1e-05, "loss": 0.4491, "step": 1929 }, { "epoch": 0.5561959654178674, "grad_norm": 1.4495361720657725, "learning_rate": 1e-05, "loss": 0.2756, "step": 1930 }, { "epoch": 0.5564841498559078, "grad_norm": 2.5107807968663756, "learning_rate": 1e-05, "loss": 0.4407, "step": 1931 }, { "epoch": 0.5567723342939481, "grad_norm": 2.3669842152948815, "learning_rate": 1e-05, "loss": 0.4479, "step": 1932 }, { "epoch": 0.5570605187319885, "grad_norm": 1.3904215328448146, "learning_rate": 1e-05, "loss": 0.2666, "step": 1933 }, { "epoch": 0.5573487031700288, "grad_norm": 3.4674459239613507, "learning_rate": 1e-05, "loss": 0.4352, "step": 1934 }, { "epoch": 0.5576368876080692, "grad_norm": 2.7767132905402256, "learning_rate": 1e-05, "loss": 0.463, "step": 1935 }, { "epoch": 0.5579250720461095, "grad_norm": 1.9492353474204853, "learning_rate": 1e-05, "loss": 0.2803, "step": 1936 }, { "epoch": 0.5582132564841499, "grad_norm": 2.691510021933372, "learning_rate": 1e-05, "loss": 0.4226, "step": 1937 }, { "epoch": 0.5585014409221902, "grad_norm": 1.5621947677382966, "learning_rate": 1e-05, "loss": 0.4541, "step": 1938 }, { "epoch": 0.5587896253602306, "grad_norm": 1.5869357621000888, "learning_rate": 1e-05, "loss": 0.2691, "step": 1939 }, { "epoch": 0.5590778097982709, "grad_norm": 3.205006071970074, "learning_rate": 1e-05, "loss": 0.4332, "step": 1940 }, { "epoch": 0.5593659942363113, "grad_norm": 2.3376904211592966, "learning_rate": 1e-05, "loss": 0.4633, "step": 1941 }, { "epoch": 0.5596541786743516, "grad_norm": 1.9904667136804608, "learning_rate": 1e-05, "loss": 0.2459, "step": 1942 }, { "epoch": 0.559942363112392, "grad_norm": 3.4591905583054086, "learning_rate": 1e-05, "loss": 0.4183, "step": 1943 }, { "epoch": 0.5602305475504322, "grad_norm": 1.753562242041457, "learning_rate": 1e-05, "loss": 0.4409, "step": 1944 }, { "epoch": 0.5605187319884726, "grad_norm": 1.902508296796601, "learning_rate": 1e-05, "loss": 0.2568, "step": 1945 }, { "epoch": 0.5608069164265129, "grad_norm": 2.423376203861388, "learning_rate": 1e-05, "loss": 0.4601, "step": 1946 }, { "epoch": 0.5610951008645533, "grad_norm": 1.754324898109713, "learning_rate": 1e-05, "loss": 0.4634, "step": 1947 }, { "epoch": 0.5613832853025936, "grad_norm": 1.8735754639317097, "learning_rate": 1e-05, "loss": 0.2709, "step": 1948 }, { "epoch": 0.561671469740634, "grad_norm": 3.452931451866348, "learning_rate": 1e-05, "loss": 0.4295, "step": 1949 }, { "epoch": 0.5619596541786743, "grad_norm": 1.5226875454891986, "learning_rate": 1e-05, "loss": 0.4518, "step": 1950 }, { "epoch": 0.5622478386167147, "grad_norm": 2.211168511151061, "learning_rate": 1e-05, "loss": 0.2622, "step": 1951 }, { "epoch": 0.562536023054755, "grad_norm": 2.1656830793762434, "learning_rate": 1e-05, "loss": 0.4609, "step": 1952 }, { "epoch": 0.5628242074927954, "grad_norm": 1.6655781931884497, "learning_rate": 1e-05, "loss": 0.4571, "step": 1953 }, { "epoch": 0.5631123919308357, "grad_norm": 2.141959730144056, "learning_rate": 1e-05, "loss": 0.2444, "step": 1954 }, { "epoch": 0.5634005763688761, "grad_norm": 2.9061920590169312, "learning_rate": 1e-05, "loss": 0.4446, "step": 1955 }, { "epoch": 0.5636887608069164, "grad_norm": 1.790019799247271, "learning_rate": 1e-05, "loss": 0.4592, "step": 1956 }, { "epoch": 0.5639769452449568, "grad_norm": 1.6906293802811507, "learning_rate": 1e-05, "loss": 0.2459, "step": 1957 }, { "epoch": 0.5642651296829971, "grad_norm": 2.65820233033142, "learning_rate": 1e-05, "loss": 0.4584, "step": 1958 }, { "epoch": 0.5645533141210375, "grad_norm": 1.6689396683044961, "learning_rate": 1e-05, "loss": 0.443, "step": 1959 }, { "epoch": 0.5648414985590778, "grad_norm": 1.7996097107784526, "learning_rate": 1e-05, "loss": 0.276, "step": 1960 }, { "epoch": 0.5651296829971182, "grad_norm": 2.7420215876567333, "learning_rate": 1e-05, "loss": 0.4471, "step": 1961 }, { "epoch": 0.5654178674351585, "grad_norm": 2.0964142455103643, "learning_rate": 1e-05, "loss": 0.4463, "step": 1962 }, { "epoch": 0.5657060518731989, "grad_norm": 2.3527628641881058, "learning_rate": 1e-05, "loss": 0.2624, "step": 1963 }, { "epoch": 0.5659942363112392, "grad_norm": 3.6000461194743734, "learning_rate": 1e-05, "loss": 0.4486, "step": 1964 }, { "epoch": 0.5662824207492796, "grad_norm": 1.734018448165514, "learning_rate": 1e-05, "loss": 0.4459, "step": 1965 }, { "epoch": 0.5665706051873198, "grad_norm": 1.736113764684417, "learning_rate": 1e-05, "loss": 0.2535, "step": 1966 }, { "epoch": 0.5668587896253602, "grad_norm": 2.820831180274689, "learning_rate": 1e-05, "loss": 0.4533, "step": 1967 }, { "epoch": 0.5671469740634005, "grad_norm": 1.7774150858091302, "learning_rate": 1e-05, "loss": 0.4648, "step": 1968 }, { "epoch": 0.5674351585014409, "grad_norm": 2.2558021269642152, "learning_rate": 1e-05, "loss": 0.2666, "step": 1969 }, { "epoch": 0.5677233429394812, "grad_norm": 2.877564985648358, "learning_rate": 1e-05, "loss": 0.4544, "step": 1970 }, { "epoch": 0.5680115273775216, "grad_norm": 3.172130672807336, "learning_rate": 1e-05, "loss": 0.4458, "step": 1971 }, { "epoch": 0.5682997118155619, "grad_norm": 1.7251175232041882, "learning_rate": 1e-05, "loss": 0.2623, "step": 1972 }, { "epoch": 0.5685878962536023, "grad_norm": 3.306136752859454, "learning_rate": 1e-05, "loss": 0.4066, "step": 1973 }, { "epoch": 0.5688760806916426, "grad_norm": 1.9264481567426284, "learning_rate": 1e-05, "loss": 0.4513, "step": 1974 }, { "epoch": 0.569164265129683, "grad_norm": 2.4078500074421965, "learning_rate": 1e-05, "loss": 0.25, "step": 1975 }, { "epoch": 0.5694524495677233, "grad_norm": 2.897428927921229, "learning_rate": 1e-05, "loss": 0.424, "step": 1976 }, { "epoch": 0.5697406340057637, "grad_norm": 2.3378278523124263, "learning_rate": 1e-05, "loss": 0.459, "step": 1977 }, { "epoch": 0.570028818443804, "grad_norm": 2.210653772895215, "learning_rate": 1e-05, "loss": 0.2728, "step": 1978 }, { "epoch": 0.5703170028818444, "grad_norm": 3.3872854527593295, "learning_rate": 1e-05, "loss": 0.4635, "step": 1979 }, { "epoch": 0.5706051873198847, "grad_norm": 1.8882172250271712, "learning_rate": 1e-05, "loss": 0.4606, "step": 1980 }, { "epoch": 0.5708933717579251, "grad_norm": 2.9454060428828086, "learning_rate": 1e-05, "loss": 0.2662, "step": 1981 }, { "epoch": 0.5711815561959654, "grad_norm": 3.275216721117935, "learning_rate": 1e-05, "loss": 0.4199, "step": 1982 }, { "epoch": 0.5714697406340058, "grad_norm": 1.4034515029833858, "learning_rate": 1e-05, "loss": 0.4517, "step": 1983 }, { "epoch": 0.5717579250720461, "grad_norm": 1.459181585993267, "learning_rate": 1e-05, "loss": 0.2429, "step": 1984 }, { "epoch": 0.5720461095100865, "grad_norm": 3.191726955890059, "learning_rate": 1e-05, "loss": 0.4206, "step": 1985 }, { "epoch": 0.5723342939481268, "grad_norm": 1.5915879246958469, "learning_rate": 1e-05, "loss": 0.454, "step": 1986 }, { "epoch": 0.5726224783861672, "grad_norm": 2.2028803814240443, "learning_rate": 1e-05, "loss": 0.2551, "step": 1987 }, { "epoch": 0.5729106628242074, "grad_norm": 2.789248722855279, "learning_rate": 1e-05, "loss": 0.4277, "step": 1988 }, { "epoch": 0.5731988472622478, "grad_norm": 1.9533881732362772, "learning_rate": 1e-05, "loss": 0.4409, "step": 1989 }, { "epoch": 0.5734870317002881, "grad_norm": 1.911761756965588, "learning_rate": 1e-05, "loss": 0.2404, "step": 1990 }, { "epoch": 0.5737752161383285, "grad_norm": 2.5606081431461725, "learning_rate": 1e-05, "loss": 0.4301, "step": 1991 }, { "epoch": 0.5740634005763688, "grad_norm": 1.88672893655734, "learning_rate": 1e-05, "loss": 0.4517, "step": 1992 }, { "epoch": 0.5743515850144092, "grad_norm": 1.8707287091071405, "learning_rate": 1e-05, "loss": 0.2709, "step": 1993 }, { "epoch": 0.5746397694524495, "grad_norm": 2.67820686651302, "learning_rate": 1e-05, "loss": 0.4452, "step": 1994 }, { "epoch": 0.5749279538904899, "grad_norm": 1.5825442843867061, "learning_rate": 1e-05, "loss": 0.4494, "step": 1995 }, { "epoch": 0.5752161383285302, "grad_norm": 1.721512530766781, "learning_rate": 1e-05, "loss": 0.2467, "step": 1996 }, { "epoch": 0.5755043227665706, "grad_norm": 2.2895279745774886, "learning_rate": 1e-05, "loss": 0.4192, "step": 1997 }, { "epoch": 0.575792507204611, "grad_norm": 2.2188312061584012, "learning_rate": 1e-05, "loss": 0.4421, "step": 1998 }, { "epoch": 0.5760806916426513, "grad_norm": 1.9452511989807846, "learning_rate": 1e-05, "loss": 0.2404, "step": 1999 }, { "epoch": 0.5763688760806917, "grad_norm": 3.3970071542160714, "learning_rate": 1e-05, "loss": 0.4299, "step": 2000 }, { "epoch": 0.5763688760806917, "eval_websight_new_IoU": 0.019274206832051277, "eval_websight_new_MAE_x": 286.760009765625, "eval_websight_new_MAE_y": 587.2000122070312, "eval_websight_new_coord_probability": 0.7459678649902344, "eval_websight_new_correct_token_probability": 0.0011421897215768695, "eval_websight_new_inside_bbox": 0.019999999552965164, "eval_websight_new_loss": 0.2701660394668579, "eval_websight_new_range_100_probability": 0.08858620375394821, "eval_websight_new_range_20_probability": 0.015967553481459618, "eval_websight_new_range_5_probability": 0.002914814744144678, "eval_websight_new_runtime": 34.5561, "eval_websight_new_samples_per_second": 1.447, "eval_websight_new_steps_per_second": 0.203, "step": 2000 }, { "epoch": 0.5763688760806917, "eval_seeclick_IoU": 0.03321577608585358, "eval_seeclick_MAE_x": 473.1000061035156, "eval_seeclick_MAE_y": 516.47998046875, "eval_seeclick_coord_probability": 0.7474353313446045, "eval_seeclick_correct_token_probability": 0.0005565558094531298, "eval_seeclick_inside_bbox": 0.05999999865889549, "eval_seeclick_loss": 0.36988624930381775, "eval_seeclick_range_100_probability": 0.12603196501731873, "eval_seeclick_range_20_probability": 0.04917369410395622, "eval_seeclick_range_5_probability": 0.007311419118195772, "eval_seeclick_runtime": 42.4256, "eval_seeclick_samples_per_second": 1.179, "eval_seeclick_steps_per_second": 0.165, "step": 2000 }, { "epoch": 0.5763688760806917, "eval_icons_IoU": 0.0019011390395462513, "eval_icons_MAE_x": 473.5400085449219, "eval_icons_MAE_y": 567.260009765625, "eval_icons_coord_probability": 0.7475740313529968, "eval_icons_correct_token_probability": 0.00864581111818552, "eval_icons_inside_bbox": 0.0, "eval_icons_loss": 0.42921721935272217, "eval_icons_range_100_probability": 0.27580708265304565, "eval_icons_range_20_probability": 0.09803719818592072, "eval_icons_range_5_probability": 0.03172297775745392, "eval_icons_runtime": 43.3526, "eval_icons_samples_per_second": 1.153, "eval_icons_steps_per_second": 0.161, "step": 2000 }, { "epoch": 0.576657060518732, "grad_norm": 1.9550860468198583, "learning_rate": 1e-05, "loss": 0.4442, "step": 2001 }, { "epoch": 0.5769452449567724, "grad_norm": 1.577494146650957, "learning_rate": 1e-05, "loss": 0.2566, "step": 2002 }, { "epoch": 0.5772334293948127, "grad_norm": 2.767049720277851, "learning_rate": 1e-05, "loss": 0.4566, "step": 2003 }, { "epoch": 0.577521613832853, "grad_norm": 2.5465025491254725, "learning_rate": 1e-05, "loss": 0.4424, "step": 2004 }, { "epoch": 0.5778097982708934, "grad_norm": 1.845007603334073, "learning_rate": 1e-05, "loss": 0.2723, "step": 2005 }, { "epoch": 0.5780979827089338, "grad_norm": 3.64718050719754, "learning_rate": 1e-05, "loss": 0.4472, "step": 2006 }, { "epoch": 0.5783861671469741, "grad_norm": 1.8344972295935609, "learning_rate": 1e-05, "loss": 0.4499, "step": 2007 }, { "epoch": 0.5786743515850145, "grad_norm": 1.6739145034281844, "learning_rate": 1e-05, "loss": 0.2392, "step": 2008 }, { "epoch": 0.5789625360230548, "grad_norm": 3.358983656918239, "learning_rate": 1e-05, "loss": 0.4054, "step": 2009 }, { "epoch": 0.579250720461095, "grad_norm": 1.5959066660117969, "learning_rate": 1e-05, "loss": 0.4551, "step": 2010 }, { "epoch": 0.5795389048991354, "grad_norm": 1.5841033665933912, "learning_rate": 1e-05, "loss": 0.2738, "step": 2011 }, { "epoch": 0.5798270893371757, "grad_norm": 3.4534268362892773, "learning_rate": 1e-05, "loss": 0.4443, "step": 2012 }, { "epoch": 0.5801152737752161, "grad_norm": 2.382349696943559, "learning_rate": 1e-05, "loss": 0.4492, "step": 2013 }, { "epoch": 0.5804034582132565, "grad_norm": 2.0286916261947745, "learning_rate": 1e-05, "loss": 0.2592, "step": 2014 }, { "epoch": 0.5806916426512968, "grad_norm": 2.6258538343031015, "learning_rate": 1e-05, "loss": 0.4514, "step": 2015 }, { "epoch": 0.5809798270893372, "grad_norm": 1.754907531010363, "learning_rate": 1e-05, "loss": 0.4545, "step": 2016 }, { "epoch": 0.5812680115273775, "grad_norm": 1.6571951399680935, "learning_rate": 1e-05, "loss": 0.2638, "step": 2017 }, { "epoch": 0.5815561959654179, "grad_norm": 3.0299171142987222, "learning_rate": 1e-05, "loss": 0.441, "step": 2018 }, { "epoch": 0.5818443804034582, "grad_norm": 2.144761348268048, "learning_rate": 1e-05, "loss": 0.4493, "step": 2019 }, { "epoch": 0.5821325648414986, "grad_norm": 2.070044806614756, "learning_rate": 1e-05, "loss": 0.2497, "step": 2020 }, { "epoch": 0.5824207492795389, "grad_norm": 3.099683898606559, "learning_rate": 1e-05, "loss": 0.4386, "step": 2021 }, { "epoch": 0.5827089337175793, "grad_norm": 2.2102813923737847, "learning_rate": 1e-05, "loss": 0.453, "step": 2022 }, { "epoch": 0.5829971181556196, "grad_norm": 1.5718659771633983, "learning_rate": 1e-05, "loss": 0.264, "step": 2023 }, { "epoch": 0.58328530259366, "grad_norm": 2.8147473350664454, "learning_rate": 1e-05, "loss": 0.4552, "step": 2024 }, { "epoch": 0.5835734870317003, "grad_norm": 1.8180551058536898, "learning_rate": 1e-05, "loss": 0.4588, "step": 2025 }, { "epoch": 0.5838616714697407, "grad_norm": 1.474364620837879, "learning_rate": 1e-05, "loss": 0.2634, "step": 2026 }, { "epoch": 0.584149855907781, "grad_norm": 3.029159607505013, "learning_rate": 1e-05, "loss": 0.4647, "step": 2027 }, { "epoch": 0.5844380403458214, "grad_norm": 2.759542058969036, "learning_rate": 1e-05, "loss": 0.4514, "step": 2028 }, { "epoch": 0.5847262247838617, "grad_norm": 1.8850118790012194, "learning_rate": 1e-05, "loss": 0.286, "step": 2029 }, { "epoch": 0.5850144092219021, "grad_norm": 3.26299391417422, "learning_rate": 1e-05, "loss": 0.4338, "step": 2030 }, { "epoch": 0.5853025936599424, "grad_norm": 2.6609988806535823, "learning_rate": 1e-05, "loss": 0.4535, "step": 2031 }, { "epoch": 0.5855907780979827, "grad_norm": 1.5752606261152426, "learning_rate": 1e-05, "loss": 0.2454, "step": 2032 }, { "epoch": 0.585878962536023, "grad_norm": 2.824232903081195, "learning_rate": 1e-05, "loss": 0.4713, "step": 2033 }, { "epoch": 0.5861671469740634, "grad_norm": 1.417072975639568, "learning_rate": 1e-05, "loss": 0.4496, "step": 2034 }, { "epoch": 0.5864553314121037, "grad_norm": 2.1161175785011914, "learning_rate": 1e-05, "loss": 0.2347, "step": 2035 }, { "epoch": 0.5867435158501441, "grad_norm": 2.6492547309497536, "learning_rate": 1e-05, "loss": 0.4195, "step": 2036 }, { "epoch": 0.5870317002881844, "grad_norm": 2.258183658195679, "learning_rate": 1e-05, "loss": 0.4461, "step": 2037 }, { "epoch": 0.5873198847262248, "grad_norm": 1.6103495381978243, "learning_rate": 1e-05, "loss": 0.2462, "step": 2038 }, { "epoch": 0.5876080691642651, "grad_norm": 3.044882260748048, "learning_rate": 1e-05, "loss": 0.4571, "step": 2039 }, { "epoch": 0.5878962536023055, "grad_norm": 1.6161419156818677, "learning_rate": 1e-05, "loss": 0.4451, "step": 2040 }, { "epoch": 0.5881844380403458, "grad_norm": 2.041516391278625, "learning_rate": 1e-05, "loss": 0.2591, "step": 2041 }, { "epoch": 0.5884726224783862, "grad_norm": 3.297007260653783, "learning_rate": 1e-05, "loss": 0.4452, "step": 2042 }, { "epoch": 0.5887608069164265, "grad_norm": 2.8069238351408665, "learning_rate": 1e-05, "loss": 0.4651, "step": 2043 }, { "epoch": 0.5890489913544669, "grad_norm": 2.5115994883017785, "learning_rate": 1e-05, "loss": 0.24, "step": 2044 }, { "epoch": 0.5893371757925072, "grad_norm": 2.9379329035629276, "learning_rate": 1e-05, "loss": 0.4284, "step": 2045 }, { "epoch": 0.5896253602305476, "grad_norm": 1.441645710217809, "learning_rate": 1e-05, "loss": 0.4577, "step": 2046 }, { "epoch": 0.5899135446685879, "grad_norm": 2.1597409075201117, "learning_rate": 1e-05, "loss": 0.257, "step": 2047 }, { "epoch": 0.5902017291066283, "grad_norm": 2.5934115332175978, "learning_rate": 1e-05, "loss": 0.4384, "step": 2048 }, { "epoch": 0.5904899135446686, "grad_norm": 1.5548112423461338, "learning_rate": 1e-05, "loss": 0.4497, "step": 2049 }, { "epoch": 0.590778097982709, "grad_norm": 1.6076087527300826, "learning_rate": 1e-05, "loss": 0.27, "step": 2050 }, { "epoch": 0.5910662824207493, "grad_norm": 2.666318432601144, "learning_rate": 1e-05, "loss": 0.4444, "step": 2051 }, { "epoch": 0.5913544668587897, "grad_norm": 1.7010832910359102, "learning_rate": 1e-05, "loss": 0.4525, "step": 2052 }, { "epoch": 0.59164265129683, "grad_norm": 2.0574392333952356, "learning_rate": 1e-05, "loss": 0.2574, "step": 2053 }, { "epoch": 0.5919308357348703, "grad_norm": 2.6422015683283426, "learning_rate": 1e-05, "loss": 0.4758, "step": 2054 }, { "epoch": 0.5922190201729106, "grad_norm": 1.7135159638115012, "learning_rate": 1e-05, "loss": 0.4483, "step": 2055 }, { "epoch": 0.592507204610951, "grad_norm": 1.9732068399032463, "learning_rate": 1e-05, "loss": 0.2607, "step": 2056 }, { "epoch": 0.5927953890489913, "grad_norm": 2.5308717917953647, "learning_rate": 1e-05, "loss": 0.4478, "step": 2057 }, { "epoch": 0.5930835734870317, "grad_norm": 1.5856993531847656, "learning_rate": 1e-05, "loss": 0.4423, "step": 2058 }, { "epoch": 0.593371757925072, "grad_norm": 2.085203981257147, "learning_rate": 1e-05, "loss": 0.2627, "step": 2059 }, { "epoch": 0.5936599423631124, "grad_norm": 3.137261404606973, "learning_rate": 1e-05, "loss": 0.4271, "step": 2060 }, { "epoch": 0.5939481268011527, "grad_norm": 1.8869257815420204, "learning_rate": 1e-05, "loss": 0.4401, "step": 2061 }, { "epoch": 0.5942363112391931, "grad_norm": 1.7277056983931665, "learning_rate": 1e-05, "loss": 0.253, "step": 2062 }, { "epoch": 0.5945244956772334, "grad_norm": 2.396605176150871, "learning_rate": 1e-05, "loss": 0.4594, "step": 2063 }, { "epoch": 0.5948126801152738, "grad_norm": 1.9784710057565758, "learning_rate": 1e-05, "loss": 0.4399, "step": 2064 }, { "epoch": 0.5951008645533141, "grad_norm": 1.581618264593078, "learning_rate": 1e-05, "loss": 0.251, "step": 2065 }, { "epoch": 0.5953890489913545, "grad_norm": 3.2845507258138884, "learning_rate": 1e-05, "loss": 0.4369, "step": 2066 }, { "epoch": 0.5956772334293948, "grad_norm": 1.8470353326205402, "learning_rate": 1e-05, "loss": 0.4397, "step": 2067 }, { "epoch": 0.5959654178674352, "grad_norm": 1.7224709385549548, "learning_rate": 1e-05, "loss": 0.2424, "step": 2068 }, { "epoch": 0.5962536023054755, "grad_norm": 4.22654189852175, "learning_rate": 1e-05, "loss": 0.4908, "step": 2069 }, { "epoch": 0.5965417867435159, "grad_norm": 2.028066531551479, "learning_rate": 1e-05, "loss": 0.4393, "step": 2070 }, { "epoch": 0.5968299711815562, "grad_norm": 2.061785829491306, "learning_rate": 1e-05, "loss": 0.2436, "step": 2071 }, { "epoch": 0.5971181556195966, "grad_norm": 2.751133923717285, "learning_rate": 1e-05, "loss": 0.4098, "step": 2072 }, { "epoch": 0.5974063400576369, "grad_norm": 2.53651075542974, "learning_rate": 1e-05, "loss": 0.4598, "step": 2073 }, { "epoch": 0.5976945244956773, "grad_norm": 2.1147054121136093, "learning_rate": 1e-05, "loss": 0.2832, "step": 2074 }, { "epoch": 0.5979827089337176, "grad_norm": 2.8304387907203155, "learning_rate": 1e-05, "loss": 0.4446, "step": 2075 }, { "epoch": 0.5982708933717579, "grad_norm": 1.7704325905974798, "learning_rate": 1e-05, "loss": 0.4484, "step": 2076 }, { "epoch": 0.5985590778097982, "grad_norm": 1.8008024416461477, "learning_rate": 1e-05, "loss": 0.2708, "step": 2077 }, { "epoch": 0.5988472622478386, "grad_norm": 3.6878907328456374, "learning_rate": 1e-05, "loss": 0.415, "step": 2078 }, { "epoch": 0.5991354466858789, "grad_norm": 1.8713840610503134, "learning_rate": 1e-05, "loss": 0.456, "step": 2079 }, { "epoch": 0.5994236311239193, "grad_norm": 1.8219286489434328, "learning_rate": 1e-05, "loss": 0.2338, "step": 2080 }, { "epoch": 0.5997118155619596, "grad_norm": 2.3137385073111174, "learning_rate": 1e-05, "loss": 0.4041, "step": 2081 }, { "epoch": 0.6, "grad_norm": 1.544285393211998, "learning_rate": 1e-05, "loss": 0.4518, "step": 2082 }, { "epoch": 0.6002881844380403, "grad_norm": 2.411733248292235, "learning_rate": 1e-05, "loss": 0.23, "step": 2083 }, { "epoch": 0.6005763688760807, "grad_norm": 3.112627136276708, "learning_rate": 1e-05, "loss": 0.4607, "step": 2084 }, { "epoch": 0.600864553314121, "grad_norm": 1.9719538611650116, "learning_rate": 1e-05, "loss": 0.45, "step": 2085 }, { "epoch": 0.6011527377521614, "grad_norm": 1.7376767223745377, "learning_rate": 1e-05, "loss": 0.2479, "step": 2086 }, { "epoch": 0.6014409221902017, "grad_norm": 2.8881007892073938, "learning_rate": 1e-05, "loss": 0.4135, "step": 2087 }, { "epoch": 0.6017291066282421, "grad_norm": 1.6446162108491584, "learning_rate": 1e-05, "loss": 0.4399, "step": 2088 }, { "epoch": 0.6020172910662824, "grad_norm": 2.7271189379748937, "learning_rate": 1e-05, "loss": 0.2484, "step": 2089 }, { "epoch": 0.6023054755043228, "grad_norm": 2.4137280688630565, "learning_rate": 1e-05, "loss": 0.4634, "step": 2090 }, { "epoch": 0.6025936599423631, "grad_norm": 1.332354211507635, "learning_rate": 1e-05, "loss": 0.443, "step": 2091 }, { "epoch": 0.6028818443804035, "grad_norm": 1.8085084227820358, "learning_rate": 1e-05, "loss": 0.2551, "step": 2092 }, { "epoch": 0.6031700288184438, "grad_norm": 3.4780717621047175, "learning_rate": 1e-05, "loss": 0.4513, "step": 2093 }, { "epoch": 0.6034582132564842, "grad_norm": 1.8600354139642854, "learning_rate": 1e-05, "loss": 0.452, "step": 2094 }, { "epoch": 0.6037463976945245, "grad_norm": 1.3794793343273797, "learning_rate": 1e-05, "loss": 0.2327, "step": 2095 }, { "epoch": 0.6040345821325649, "grad_norm": 2.9968348958854776, "learning_rate": 1e-05, "loss": 0.4055, "step": 2096 }, { "epoch": 0.6043227665706052, "grad_norm": 1.8585527895953826, "learning_rate": 1e-05, "loss": 0.4457, "step": 2097 }, { "epoch": 0.6046109510086455, "grad_norm": 1.6809499829679715, "learning_rate": 1e-05, "loss": 0.2397, "step": 2098 }, { "epoch": 0.6048991354466858, "grad_norm": 3.1547568437589546, "learning_rate": 1e-05, "loss": 0.466, "step": 2099 }, { "epoch": 0.6051873198847262, "grad_norm": 1.5434664194751035, "learning_rate": 1e-05, "loss": 0.4355, "step": 2100 }, { "epoch": 0.6054755043227665, "grad_norm": 1.332929757256648, "learning_rate": 1e-05, "loss": 0.2457, "step": 2101 }, { "epoch": 0.6057636887608069, "grad_norm": 3.2479715797322, "learning_rate": 1e-05, "loss": 0.4392, "step": 2102 }, { "epoch": 0.6060518731988472, "grad_norm": 1.564221211658833, "learning_rate": 1e-05, "loss": 0.4557, "step": 2103 }, { "epoch": 0.6063400576368876, "grad_norm": 1.5854904076813274, "learning_rate": 1e-05, "loss": 0.2338, "step": 2104 }, { "epoch": 0.6066282420749279, "grad_norm": 2.546217553642099, "learning_rate": 1e-05, "loss": 0.4393, "step": 2105 }, { "epoch": 0.6069164265129683, "grad_norm": 2.4010235327100222, "learning_rate": 1e-05, "loss": 0.4532, "step": 2106 }, { "epoch": 0.6072046109510086, "grad_norm": 2.564789725880665, "learning_rate": 1e-05, "loss": 0.2498, "step": 2107 }, { "epoch": 0.607492795389049, "grad_norm": 2.919474498363441, "learning_rate": 1e-05, "loss": 0.4448, "step": 2108 }, { "epoch": 0.6077809798270893, "grad_norm": 1.687511716719379, "learning_rate": 1e-05, "loss": 0.4555, "step": 2109 }, { "epoch": 0.6080691642651297, "grad_norm": 1.8155151009626451, "learning_rate": 1e-05, "loss": 0.2766, "step": 2110 }, { "epoch": 0.60835734870317, "grad_norm": 3.3289963509643115, "learning_rate": 1e-05, "loss": 0.4555, "step": 2111 }, { "epoch": 0.6086455331412104, "grad_norm": 1.7171666192458641, "learning_rate": 1e-05, "loss": 0.4466, "step": 2112 }, { "epoch": 0.6089337175792507, "grad_norm": 1.4719082693173857, "learning_rate": 1e-05, "loss": 0.2462, "step": 2113 }, { "epoch": 0.6092219020172911, "grad_norm": 2.549437257988022, "learning_rate": 1e-05, "loss": 0.4185, "step": 2114 }, { "epoch": 0.6095100864553314, "grad_norm": 1.5133571866347024, "learning_rate": 1e-05, "loss": 0.4434, "step": 2115 }, { "epoch": 0.6097982708933718, "grad_norm": 1.6460959768234367, "learning_rate": 1e-05, "loss": 0.2327, "step": 2116 }, { "epoch": 0.6100864553314121, "grad_norm": 3.265837251000058, "learning_rate": 1e-05, "loss": 0.4489, "step": 2117 }, { "epoch": 0.6103746397694525, "grad_norm": 1.9242162867891692, "learning_rate": 1e-05, "loss": 0.4426, "step": 2118 }, { "epoch": 0.6106628242074928, "grad_norm": 1.6707231066794073, "learning_rate": 1e-05, "loss": 0.2424, "step": 2119 }, { "epoch": 0.6109510086455331, "grad_norm": 3.040514531953241, "learning_rate": 1e-05, "loss": 0.4243, "step": 2120 }, { "epoch": 0.6112391930835734, "grad_norm": 2.0241337559111474, "learning_rate": 1e-05, "loss": 0.4487, "step": 2121 }, { "epoch": 0.6115273775216138, "grad_norm": 2.0137596032860605, "learning_rate": 1e-05, "loss": 0.2659, "step": 2122 }, { "epoch": 0.6118155619596541, "grad_norm": 2.748840221276641, "learning_rate": 1e-05, "loss": 0.4226, "step": 2123 }, { "epoch": 0.6121037463976945, "grad_norm": 1.7530949492316532, "learning_rate": 1e-05, "loss": 0.4391, "step": 2124 }, { "epoch": 0.6123919308357348, "grad_norm": 1.9134842833787362, "learning_rate": 1e-05, "loss": 0.2124, "step": 2125 }, { "epoch": 0.6126801152737752, "grad_norm": 2.6493052156796493, "learning_rate": 1e-05, "loss": 0.4462, "step": 2126 }, { "epoch": 0.6129682997118155, "grad_norm": 1.5546265100802588, "learning_rate": 1e-05, "loss": 0.4402, "step": 2127 }, { "epoch": 0.6132564841498559, "grad_norm": 1.5070519370111153, "learning_rate": 1e-05, "loss": 0.2406, "step": 2128 }, { "epoch": 0.6135446685878962, "grad_norm": 3.4153867999695797, "learning_rate": 1e-05, "loss": 0.4442, "step": 2129 }, { "epoch": 0.6138328530259366, "grad_norm": 1.8782783252038626, "learning_rate": 1e-05, "loss": 0.4412, "step": 2130 }, { "epoch": 0.614121037463977, "grad_norm": 1.4955906768963096, "learning_rate": 1e-05, "loss": 0.2502, "step": 2131 }, { "epoch": 0.6144092219020173, "grad_norm": 2.656539831948737, "learning_rate": 1e-05, "loss": 0.4265, "step": 2132 }, { "epoch": 0.6146974063400577, "grad_norm": 1.4498378146223199, "learning_rate": 1e-05, "loss": 0.4532, "step": 2133 }, { "epoch": 0.614985590778098, "grad_norm": 2.139351593612628, "learning_rate": 1e-05, "loss": 0.2513, "step": 2134 }, { "epoch": 0.6152737752161384, "grad_norm": 3.958012147042486, "learning_rate": 1e-05, "loss": 0.4616, "step": 2135 }, { "epoch": 0.6155619596541787, "grad_norm": 1.7572465145136853, "learning_rate": 1e-05, "loss": 0.4482, "step": 2136 }, { "epoch": 0.615850144092219, "grad_norm": 1.710388138782609, "learning_rate": 1e-05, "loss": 0.2638, "step": 2137 }, { "epoch": 0.6161383285302594, "grad_norm": 2.8416428396616182, "learning_rate": 1e-05, "loss": 0.443, "step": 2138 }, { "epoch": 0.6164265129682998, "grad_norm": 1.9421895457064535, "learning_rate": 1e-05, "loss": 0.4569, "step": 2139 }, { "epoch": 0.6167146974063401, "grad_norm": 1.5315925362235654, "learning_rate": 1e-05, "loss": 0.2386, "step": 2140 }, { "epoch": 0.6170028818443805, "grad_norm": 2.5412597660103557, "learning_rate": 1e-05, "loss": 0.4376, "step": 2141 }, { "epoch": 0.6172910662824207, "grad_norm": 1.5915989198552307, "learning_rate": 1e-05, "loss": 0.4401, "step": 2142 }, { "epoch": 0.617579250720461, "grad_norm": 2.1522199351416056, "learning_rate": 1e-05, "loss": 0.2544, "step": 2143 }, { "epoch": 0.6178674351585014, "grad_norm": 2.5485317240260077, "learning_rate": 1e-05, "loss": 0.4407, "step": 2144 }, { "epoch": 0.6181556195965417, "grad_norm": 1.5313382595483551, "learning_rate": 1e-05, "loss": 0.4557, "step": 2145 }, { "epoch": 0.6184438040345821, "grad_norm": 2.054486267832255, "learning_rate": 1e-05, "loss": 0.2723, "step": 2146 }, { "epoch": 0.6187319884726225, "grad_norm": 3.4090998875109837, "learning_rate": 1e-05, "loss": 0.4408, "step": 2147 }, { "epoch": 0.6190201729106628, "grad_norm": 1.9044220054973855, "learning_rate": 1e-05, "loss": 0.4521, "step": 2148 }, { "epoch": 0.6193083573487032, "grad_norm": 2.7121438865991148, "learning_rate": 1e-05, "loss": 0.2245, "step": 2149 }, { "epoch": 0.6195965417867435, "grad_norm": 2.5732597325520836, "learning_rate": 1e-05, "loss": 0.445, "step": 2150 }, { "epoch": 0.6198847262247839, "grad_norm": 2.1823612337363514, "learning_rate": 1e-05, "loss": 0.4467, "step": 2151 }, { "epoch": 0.6201729106628242, "grad_norm": 1.6829309697016506, "learning_rate": 1e-05, "loss": 0.2499, "step": 2152 }, { "epoch": 0.6204610951008646, "grad_norm": 2.637279853976246, "learning_rate": 1e-05, "loss": 0.4658, "step": 2153 }, { "epoch": 0.6207492795389049, "grad_norm": 1.5904686030479704, "learning_rate": 1e-05, "loss": 0.4341, "step": 2154 }, { "epoch": 0.6210374639769453, "grad_norm": 1.7998781923573408, "learning_rate": 1e-05, "loss": 0.2492, "step": 2155 }, { "epoch": 0.6213256484149856, "grad_norm": 2.8768727807697743, "learning_rate": 1e-05, "loss": 0.4269, "step": 2156 }, { "epoch": 0.621613832853026, "grad_norm": 1.6298772350844697, "learning_rate": 1e-05, "loss": 0.4438, "step": 2157 }, { "epoch": 0.6219020172910663, "grad_norm": 1.9792746786563291, "learning_rate": 1e-05, "loss": 0.2437, "step": 2158 }, { "epoch": 0.6221902017291067, "grad_norm": 2.6610043866659434, "learning_rate": 1e-05, "loss": 0.4365, "step": 2159 }, { "epoch": 0.622478386167147, "grad_norm": 2.1428413522782495, "learning_rate": 1e-05, "loss": 0.4459, "step": 2160 }, { "epoch": 0.6227665706051874, "grad_norm": 2.668678412115329, "learning_rate": 1e-05, "loss": 0.2337, "step": 2161 }, { "epoch": 0.6230547550432277, "grad_norm": 2.8447343605724194, "learning_rate": 1e-05, "loss": 0.4771, "step": 2162 }, { "epoch": 0.6233429394812681, "grad_norm": 1.8125656931211491, "learning_rate": 1e-05, "loss": 0.4355, "step": 2163 }, { "epoch": 0.6236311239193083, "grad_norm": 1.947844304822238, "learning_rate": 1e-05, "loss": 0.266, "step": 2164 }, { "epoch": 0.6239193083573487, "grad_norm": 3.6323691529340185, "learning_rate": 1e-05, "loss": 0.4631, "step": 2165 }, { "epoch": 0.624207492795389, "grad_norm": 2.2842934584255774, "learning_rate": 1e-05, "loss": 0.4512, "step": 2166 }, { "epoch": 0.6244956772334294, "grad_norm": 1.930752931023711, "learning_rate": 1e-05, "loss": 0.2246, "step": 2167 }, { "epoch": 0.6247838616714697, "grad_norm": 2.560422305241705, "learning_rate": 1e-05, "loss": 0.4496, "step": 2168 }, { "epoch": 0.6250720461095101, "grad_norm": 1.4329863627493236, "learning_rate": 1e-05, "loss": 0.4413, "step": 2169 }, { "epoch": 0.6253602305475504, "grad_norm": 1.945434478132978, "learning_rate": 1e-05, "loss": 0.2304, "step": 2170 }, { "epoch": 0.6256484149855908, "grad_norm": 2.9717639049275038, "learning_rate": 1e-05, "loss": 0.3918, "step": 2171 }, { "epoch": 0.6259365994236311, "grad_norm": 1.9612900231556427, "learning_rate": 1e-05, "loss": 0.4452, "step": 2172 }, { "epoch": 0.6262247838616715, "grad_norm": 2.065594340795662, "learning_rate": 1e-05, "loss": 0.2569, "step": 2173 }, { "epoch": 0.6265129682997118, "grad_norm": 2.4738696719914053, "learning_rate": 1e-05, "loss": 0.4151, "step": 2174 }, { "epoch": 0.6268011527377522, "grad_norm": 1.8902591724331854, "learning_rate": 1e-05, "loss": 0.4523, "step": 2175 }, { "epoch": 0.6270893371757925, "grad_norm": 1.6745493299984548, "learning_rate": 1e-05, "loss": 0.2395, "step": 2176 }, { "epoch": 0.6273775216138329, "grad_norm": 3.2443745361221126, "learning_rate": 1e-05, "loss": 0.4415, "step": 2177 }, { "epoch": 0.6276657060518732, "grad_norm": 2.276271708958747, "learning_rate": 1e-05, "loss": 0.4486, "step": 2178 }, { "epoch": 0.6279538904899136, "grad_norm": 1.7325942030950319, "learning_rate": 1e-05, "loss": 0.2308, "step": 2179 }, { "epoch": 0.6282420749279539, "grad_norm": 2.7488994586572955, "learning_rate": 1e-05, "loss": 0.4428, "step": 2180 }, { "epoch": 0.6285302593659943, "grad_norm": 1.6960998545372619, "learning_rate": 1e-05, "loss": 0.4439, "step": 2181 }, { "epoch": 0.6288184438040346, "grad_norm": 1.5366062651961152, "learning_rate": 1e-05, "loss": 0.2323, "step": 2182 }, { "epoch": 0.629106628242075, "grad_norm": 3.2775864590246266, "learning_rate": 1e-05, "loss": 0.4763, "step": 2183 }, { "epoch": 0.6293948126801153, "grad_norm": 1.398966470366019, "learning_rate": 1e-05, "loss": 0.4406, "step": 2184 }, { "epoch": 0.6296829971181557, "grad_norm": 1.6627867337782436, "learning_rate": 1e-05, "loss": 0.2534, "step": 2185 }, { "epoch": 0.6299711815561959, "grad_norm": 2.3357956265932067, "learning_rate": 1e-05, "loss": 0.4364, "step": 2186 }, { "epoch": 0.6302593659942363, "grad_norm": 2.4272619214150675, "learning_rate": 1e-05, "loss": 0.4433, "step": 2187 }, { "epoch": 0.6305475504322766, "grad_norm": 1.5897311251059187, "learning_rate": 1e-05, "loss": 0.2323, "step": 2188 }, { "epoch": 0.630835734870317, "grad_norm": 3.6255015074985177, "learning_rate": 1e-05, "loss": 0.47, "step": 2189 }, { "epoch": 0.6311239193083573, "grad_norm": 1.710826581051854, "learning_rate": 1e-05, "loss": 0.4348, "step": 2190 }, { "epoch": 0.6314121037463977, "grad_norm": 1.3214079312521063, "learning_rate": 1e-05, "loss": 0.2396, "step": 2191 }, { "epoch": 0.631700288184438, "grad_norm": 3.6581278309385974, "learning_rate": 1e-05, "loss": 0.4557, "step": 2192 }, { "epoch": 0.6319884726224784, "grad_norm": 1.9873933956261685, "learning_rate": 1e-05, "loss": 0.4396, "step": 2193 }, { "epoch": 0.6322766570605187, "grad_norm": 1.945389894464319, "learning_rate": 1e-05, "loss": 0.2469, "step": 2194 }, { "epoch": 0.6325648414985591, "grad_norm": 2.4274401087222532, "learning_rate": 1e-05, "loss": 0.4449, "step": 2195 }, { "epoch": 0.6328530259365994, "grad_norm": 1.7005894252433857, "learning_rate": 1e-05, "loss": 0.4442, "step": 2196 }, { "epoch": 0.6331412103746398, "grad_norm": 1.706655592221686, "learning_rate": 1e-05, "loss": 0.2731, "step": 2197 }, { "epoch": 0.6334293948126801, "grad_norm": 3.1349413931657444, "learning_rate": 1e-05, "loss": 0.433, "step": 2198 }, { "epoch": 0.6337175792507205, "grad_norm": 1.6117010170738182, "learning_rate": 1e-05, "loss": 0.4403, "step": 2199 }, { "epoch": 0.6340057636887608, "grad_norm": 2.029578739632731, "learning_rate": 1e-05, "loss": 0.2425, "step": 2200 }, { "epoch": 0.6342939481268012, "grad_norm": 2.5226996042723933, "learning_rate": 1e-05, "loss": 0.4478, "step": 2201 }, { "epoch": 0.6345821325648415, "grad_norm": 1.4749346665703282, "learning_rate": 1e-05, "loss": 0.4342, "step": 2202 }, { "epoch": 0.6348703170028819, "grad_norm": 1.7929226949785588, "learning_rate": 1e-05, "loss": 0.2469, "step": 2203 }, { "epoch": 0.6351585014409222, "grad_norm": 3.1932894460576, "learning_rate": 1e-05, "loss": 0.4478, "step": 2204 }, { "epoch": 0.6354466858789626, "grad_norm": 1.6411041485005728, "learning_rate": 1e-05, "loss": 0.4527, "step": 2205 }, { "epoch": 0.6357348703170029, "grad_norm": 1.782848833184517, "learning_rate": 1e-05, "loss": 0.2319, "step": 2206 }, { "epoch": 0.6360230547550433, "grad_norm": 2.813445973391478, "learning_rate": 1e-05, "loss": 0.4259, "step": 2207 }, { "epoch": 0.6363112391930835, "grad_norm": 1.570466548333388, "learning_rate": 1e-05, "loss": 0.4536, "step": 2208 }, { "epoch": 0.6365994236311239, "grad_norm": 1.9323020194051894, "learning_rate": 1e-05, "loss": 0.2475, "step": 2209 }, { "epoch": 0.6368876080691642, "grad_norm": 2.848547237007071, "learning_rate": 1e-05, "loss": 0.4506, "step": 2210 }, { "epoch": 0.6371757925072046, "grad_norm": 1.6034499628142453, "learning_rate": 1e-05, "loss": 0.4457, "step": 2211 }, { "epoch": 0.6374639769452449, "grad_norm": 1.5031872979236383, "learning_rate": 1e-05, "loss": 0.2531, "step": 2212 }, { "epoch": 0.6377521613832853, "grad_norm": 4.04524836779013, "learning_rate": 1e-05, "loss": 0.4232, "step": 2213 }, { "epoch": 0.6380403458213256, "grad_norm": 2.6051989171617915, "learning_rate": 1e-05, "loss": 0.4447, "step": 2214 }, { "epoch": 0.638328530259366, "grad_norm": 2.4253756509703863, "learning_rate": 1e-05, "loss": 0.2495, "step": 2215 }, { "epoch": 0.6386167146974063, "grad_norm": 3.992543219607941, "learning_rate": 1e-05, "loss": 0.4477, "step": 2216 }, { "epoch": 0.6389048991354467, "grad_norm": 2.746369999913768, "learning_rate": 1e-05, "loss": 0.4516, "step": 2217 }, { "epoch": 0.639193083573487, "grad_norm": 2.161832034593531, "learning_rate": 1e-05, "loss": 0.2387, "step": 2218 }, { "epoch": 0.6394812680115274, "grad_norm": 2.8117143050302555, "learning_rate": 1e-05, "loss": 0.4242, "step": 2219 }, { "epoch": 0.6397694524495677, "grad_norm": 2.153834976919712, "learning_rate": 1e-05, "loss": 0.4366, "step": 2220 }, { "epoch": 0.6400576368876081, "grad_norm": 2.005668586029055, "learning_rate": 1e-05, "loss": 0.2281, "step": 2221 }, { "epoch": 0.6403458213256484, "grad_norm": 3.7929156336115066, "learning_rate": 1e-05, "loss": 0.4786, "step": 2222 }, { "epoch": 0.6406340057636888, "grad_norm": 2.4655314378548514, "learning_rate": 1e-05, "loss": 0.4458, "step": 2223 }, { "epoch": 0.6409221902017291, "grad_norm": 1.9935463871528907, "learning_rate": 1e-05, "loss": 0.243, "step": 2224 }, { "epoch": 0.6412103746397695, "grad_norm": 2.970243036083692, "learning_rate": 1e-05, "loss": 0.4445, "step": 2225 }, { "epoch": 0.6414985590778098, "grad_norm": 1.4983249879619065, "learning_rate": 1e-05, "loss": 0.4363, "step": 2226 }, { "epoch": 0.6417867435158502, "grad_norm": 1.7566778247567378, "learning_rate": 1e-05, "loss": 0.2574, "step": 2227 }, { "epoch": 0.6420749279538905, "grad_norm": 2.755327867084357, "learning_rate": 1e-05, "loss": 0.4414, "step": 2228 }, { "epoch": 0.6423631123919309, "grad_norm": 1.5241609644119447, "learning_rate": 1e-05, "loss": 0.4541, "step": 2229 }, { "epoch": 0.6426512968299711, "grad_norm": 1.6525976260023942, "learning_rate": 1e-05, "loss": 0.2416, "step": 2230 }, { "epoch": 0.6429394812680115, "grad_norm": 2.6696724083824166, "learning_rate": 1e-05, "loss": 0.4304, "step": 2231 }, { "epoch": 0.6432276657060518, "grad_norm": 1.6959295340722362, "learning_rate": 1e-05, "loss": 0.4355, "step": 2232 }, { "epoch": 0.6435158501440922, "grad_norm": 1.5511738255136893, "learning_rate": 1e-05, "loss": 0.2395, "step": 2233 }, { "epoch": 0.6438040345821325, "grad_norm": 2.8396846352629117, "learning_rate": 1e-05, "loss": 0.3925, "step": 2234 }, { "epoch": 0.6440922190201729, "grad_norm": 1.6811320261958476, "learning_rate": 1e-05, "loss": 0.4409, "step": 2235 }, { "epoch": 0.6443804034582132, "grad_norm": 1.5161129898204209, "learning_rate": 1e-05, "loss": 0.2473, "step": 2236 }, { "epoch": 0.6446685878962536, "grad_norm": 2.276729482868985, "learning_rate": 1e-05, "loss": 0.4029, "step": 2237 }, { "epoch": 0.6449567723342939, "grad_norm": 2.0327197148905523, "learning_rate": 1e-05, "loss": 0.4414, "step": 2238 }, { "epoch": 0.6452449567723343, "grad_norm": 1.5297950175355017, "learning_rate": 1e-05, "loss": 0.2331, "step": 2239 }, { "epoch": 0.6455331412103746, "grad_norm": 2.494499277973723, "learning_rate": 1e-05, "loss": 0.4435, "step": 2240 }, { "epoch": 0.645821325648415, "grad_norm": 1.6988591974997773, "learning_rate": 1e-05, "loss": 0.4421, "step": 2241 }, { "epoch": 0.6461095100864553, "grad_norm": 1.6490346279860597, "learning_rate": 1e-05, "loss": 0.2377, "step": 2242 }, { "epoch": 0.6463976945244957, "grad_norm": 3.2965946421832437, "learning_rate": 1e-05, "loss": 0.4428, "step": 2243 }, { "epoch": 0.646685878962536, "grad_norm": 1.6432481617393475, "learning_rate": 1e-05, "loss": 0.4413, "step": 2244 }, { "epoch": 0.6469740634005764, "grad_norm": 2.005173590063975, "learning_rate": 1e-05, "loss": 0.2428, "step": 2245 }, { "epoch": 0.6472622478386167, "grad_norm": 3.170974833067571, "learning_rate": 1e-05, "loss": 0.4177, "step": 2246 }, { "epoch": 0.6475504322766571, "grad_norm": 2.075420646325355, "learning_rate": 1e-05, "loss": 0.4267, "step": 2247 }, { "epoch": 0.6478386167146974, "grad_norm": 1.6139530020843824, "learning_rate": 1e-05, "loss": 0.2549, "step": 2248 }, { "epoch": 0.6481268011527378, "grad_norm": 2.451013669431529, "learning_rate": 1e-05, "loss": 0.4126, "step": 2249 }, { "epoch": 0.6484149855907781, "grad_norm": 1.9988837079216608, "learning_rate": 1e-05, "loss": 0.4418, "step": 2250 }, { "epoch": 0.6484149855907781, "eval_websight_new_IoU": 0.01669546402990818, "eval_websight_new_MAE_x": 286.760009765625, "eval_websight_new_MAE_y": 588.2999877929688, "eval_websight_new_coord_probability": 0.7470889091491699, "eval_websight_new_correct_token_probability": 0.0008188483188860118, "eval_websight_new_inside_bbox": 0.019999999552965164, "eval_websight_new_loss": 0.2525531053543091, "eval_websight_new_range_100_probability": 0.0876244381070137, "eval_websight_new_range_20_probability": 0.014237750321626663, "eval_websight_new_range_5_probability": 0.0023667821660637856, "eval_websight_new_runtime": 33.5452, "eval_websight_new_samples_per_second": 1.491, "eval_websight_new_steps_per_second": 0.209, "step": 2250 }, { "epoch": 0.6484149855907781, "eval_seeclick_IoU": 0.032595857977867126, "eval_seeclick_MAE_x": 473.1000061035156, "eval_seeclick_MAE_y": 579.4199829101562, "eval_seeclick_coord_probability": 0.748126745223999, "eval_seeclick_correct_token_probability": 0.0005362248048186302, "eval_seeclick_inside_bbox": 0.05999999865889549, "eval_seeclick_loss": 0.36191099882125854, "eval_seeclick_range_100_probability": 0.1297806203365326, "eval_seeclick_range_20_probability": 0.0489816851913929, "eval_seeclick_range_5_probability": 0.0075534796342253685, "eval_seeclick_runtime": 43.4368, "eval_seeclick_samples_per_second": 1.151, "eval_seeclick_steps_per_second": 0.161, "step": 2250 }, { "epoch": 0.6484149855907781, "eval_icons_IoU": 0.0030106292106211185, "eval_icons_MAE_x": 473.5199890136719, "eval_icons_MAE_y": 569.219970703125, "eval_icons_coord_probability": 0.7479195594787598, "eval_icons_correct_token_probability": 0.009233924560248852, "eval_icons_inside_bbox": 0.0, "eval_icons_loss": 0.42071011662483215, "eval_icons_range_100_probability": 0.2812122106552124, "eval_icons_range_20_probability": 0.10173630714416504, "eval_icons_range_5_probability": 0.035193752497434616, "eval_icons_runtime": 43.6615, "eval_icons_samples_per_second": 1.145, "eval_icons_steps_per_second": 0.16, "step": 2250 } ], "logging_steps": 1.0, "max_steps": 20000, "num_input_tokens_seen": 0, "num_train_epochs": 6, "save_steps": 250, "total_flos": 1070009904922624.0, "train_batch_size": 16, "trial_name": null, "trial_params": null }