diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,14615 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 2.0, + "eval_steps": 200.0, + "global_step": 10410, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00019212295869356388, + "grad_norm": 0.08390819281339645, + "learning_rate": 5.758157389635316e-07, + "loss": 3.274601936340332, + "step": 1 + }, + { + "epoch": 0.0009606147934678194, + "grad_norm": 0.08173709362745285, + "learning_rate": 2.879078694817658e-06, + "loss": 3.2788190841674805, + "step": 5 + }, + { + "epoch": 0.0019212295869356388, + "grad_norm": 0.09356535971164703, + "learning_rate": 5.758157389635316e-06, + "loss": 3.2803062438964843, + "step": 10 + }, + { + "epoch": 0.002881844380403458, + "grad_norm": 0.11097322404384613, + "learning_rate": 8.637236084452974e-06, + "loss": 3.277487564086914, + "step": 15 + }, + { + "epoch": 0.0038424591738712775, + "grad_norm": 0.15006589889526367, + "learning_rate": 1.1516314779270632e-05, + "loss": 3.281480407714844, + "step": 20 + }, + { + "epoch": 0.004803073967339097, + "grad_norm": 0.14452151954174042, + "learning_rate": 1.439539347408829e-05, + "loss": 3.2803359985351563, + "step": 25 + }, + { + "epoch": 0.005763688760806916, + "grad_norm": 0.16946451365947723, + "learning_rate": 1.7274472168905948e-05, + "loss": 3.287946319580078, + "step": 30 + }, + { + "epoch": 0.0067243035542747355, + "grad_norm": 0.14511597156524658, + "learning_rate": 2.015355086372361e-05, + "loss": 3.278676986694336, + "step": 35 + }, + { + "epoch": 0.007684918347742555, + "grad_norm": 0.21068508923053741, + "learning_rate": 2.3032629558541264e-05, + "loss": 3.2816986083984374, + "step": 40 + }, + { + "epoch": 0.008645533141210375, + "grad_norm": 0.2152123898267746, + "learning_rate": 2.591170825335892e-05, + "loss": 3.2830833435058593, + "step": 45 + }, + { + "epoch": 0.009606147934678195, + "grad_norm": 0.2621569037437439, + "learning_rate": 2.879078694817658e-05, + "loss": 3.279366302490234, + "step": 50 + }, + { + "epoch": 0.010566762728146013, + "grad_norm": 0.29736942052841187, + "learning_rate": 3.166986564299424e-05, + "loss": 3.2836788177490233, + "step": 55 + }, + { + "epoch": 0.011527377521613832, + "grad_norm": 0.31358596682548523, + "learning_rate": 3.4548944337811895e-05, + "loss": 3.2797149658203124, + "step": 60 + }, + { + "epoch": 0.012487992315081652, + "grad_norm": 0.42882370948791504, + "learning_rate": 3.742802303262955e-05, + "loss": 3.2872390747070312, + "step": 65 + }, + { + "epoch": 0.013448607108549471, + "grad_norm": 0.3247455060482025, + "learning_rate": 4.030710172744722e-05, + "loss": 3.2825721740722655, + "step": 70 + }, + { + "epoch": 0.01440922190201729, + "grad_norm": 0.40256041288375854, + "learning_rate": 4.318618042226487e-05, + "loss": 3.279859924316406, + "step": 75 + }, + { + "epoch": 0.01536983669548511, + "grad_norm": 0.5199999809265137, + "learning_rate": 4.606525911708253e-05, + "loss": 3.284808349609375, + "step": 80 + }, + { + "epoch": 0.01633045148895293, + "grad_norm": 0.6197025179862976, + "learning_rate": 4.894433781190019e-05, + "loss": 3.2813262939453125, + "step": 85 + }, + { + "epoch": 0.01729106628242075, + "grad_norm": 0.4358411133289337, + "learning_rate": 5.182341650671784e-05, + "loss": 3.2836116790771483, + "step": 90 + }, + { + "epoch": 0.01825168107588857, + "grad_norm": 0.5619639158248901, + "learning_rate": 5.4702495201535504e-05, + "loss": 3.281964874267578, + "step": 95 + }, + { + "epoch": 0.01921229586935639, + "grad_norm": 0.42559099197387695, + "learning_rate": 5.758157389635316e-05, + "loss": 3.284071350097656, + "step": 100 + }, + { + "epoch": 0.020172910662824207, + "grad_norm": 0.5353518128395081, + "learning_rate": 6.0460652591170814e-05, + "loss": 3.286463165283203, + "step": 105 + }, + { + "epoch": 0.021133525456292025, + "grad_norm": 0.8242971897125244, + "learning_rate": 6.333973128598848e-05, + "loss": 3.28314094543457, + "step": 110 + }, + { + "epoch": 0.022094140249759846, + "grad_norm": 0.5128299593925476, + "learning_rate": 6.621880998080614e-05, + "loss": 3.2824722290039063, + "step": 115 + }, + { + "epoch": 0.023054755043227664, + "grad_norm": 0.7521133422851562, + "learning_rate": 6.909788867562379e-05, + "loss": 3.2825164794921875, + "step": 120 + }, + { + "epoch": 0.024015369836695485, + "grad_norm": 0.7099725008010864, + "learning_rate": 7.197696737044145e-05, + "loss": 3.2856216430664062, + "step": 125 + }, + { + "epoch": 0.024975984630163303, + "grad_norm": 0.673387885093689, + "learning_rate": 7.48560460652591e-05, + "loss": 3.2845367431640624, + "step": 130 + }, + { + "epoch": 0.025936599423631124, + "grad_norm": 0.47837865352630615, + "learning_rate": 7.773512476007677e-05, + "loss": 3.2803184509277346, + "step": 135 + }, + { + "epoch": 0.026897214217098942, + "grad_norm": 0.5746335983276367, + "learning_rate": 8.061420345489444e-05, + "loss": 3.2853370666503907, + "step": 140 + }, + { + "epoch": 0.027857829010566763, + "grad_norm": 0.9516947269439697, + "learning_rate": 8.349328214971209e-05, + "loss": 3.2811763763427733, + "step": 145 + }, + { + "epoch": 0.02881844380403458, + "grad_norm": 0.7206563949584961, + "learning_rate": 8.637236084452975e-05, + "loss": 3.2858177185058595, + "step": 150 + }, + { + "epoch": 0.029779058597502402, + "grad_norm": 0.9468950033187866, + "learning_rate": 8.92514395393474e-05, + "loss": 3.2817359924316407, + "step": 155 + }, + { + "epoch": 0.03073967339097022, + "grad_norm": 0.9247293472290039, + "learning_rate": 9.213051823416505e-05, + "loss": 3.283827209472656, + "step": 160 + }, + { + "epoch": 0.03170028818443804, + "grad_norm": 0.7639815807342529, + "learning_rate": 9.500959692898272e-05, + "loss": 3.2805789947509765, + "step": 165 + }, + { + "epoch": 0.03266090297790586, + "grad_norm": 1.0864217281341553, + "learning_rate": 9.788867562380038e-05, + "loss": 3.282984161376953, + "step": 170 + }, + { + "epoch": 0.03362151777137368, + "grad_norm": 1.0888683795928955, + "learning_rate": 0.00010076775431861803, + "loss": 3.2868572235107423, + "step": 175 + }, + { + "epoch": 0.0345821325648415, + "grad_norm": 0.7901690602302551, + "learning_rate": 0.00010364683301343569, + "loss": 3.2854034423828127, + "step": 180 + }, + { + "epoch": 0.03554274735830932, + "grad_norm": 1.132866621017456, + "learning_rate": 0.00010652591170825334, + "loss": 3.283061218261719, + "step": 185 + }, + { + "epoch": 0.03650336215177714, + "grad_norm": 1.1301084756851196, + "learning_rate": 0.00010940499040307101, + "loss": 3.2865623474121093, + "step": 190 + }, + { + "epoch": 0.037463976945244955, + "grad_norm": 0.9956747889518738, + "learning_rate": 0.00011228406909788866, + "loss": 3.2860641479492188, + "step": 195 + }, + { + "epoch": 0.03842459173871278, + "grad_norm": 1.1139202117919922, + "learning_rate": 0.00011516314779270632, + "loss": 3.2860763549804686, + "step": 200 + }, + { + "epoch": 0.0393852065321806, + "grad_norm": 1.1494112014770508, + "learning_rate": 0.00011804222648752397, + "loss": 3.2874046325683595, + "step": 205 + }, + { + "epoch": 0.040345821325648415, + "grad_norm": 1.1791850328445435, + "learning_rate": 0.00012092130518234163, + "loss": 3.28511962890625, + "step": 210 + }, + { + "epoch": 0.04130643611911623, + "grad_norm": 0.8412664532661438, + "learning_rate": 0.0001238003838771593, + "loss": 3.2826396942138674, + "step": 215 + }, + { + "epoch": 0.04226705091258405, + "grad_norm": 1.0537320375442505, + "learning_rate": 0.00012667946257197696, + "loss": 3.2859474182128907, + "step": 220 + }, + { + "epoch": 0.043227665706051875, + "grad_norm": 0.9535348415374756, + "learning_rate": 0.00012955854126679462, + "loss": 3.2872699737548827, + "step": 225 + }, + { + "epoch": 0.04418828049951969, + "grad_norm": 1.2287579774856567, + "learning_rate": 0.00013243761996161227, + "loss": 3.2854686737060548, + "step": 230 + }, + { + "epoch": 0.04514889529298751, + "grad_norm": 0.8720296621322632, + "learning_rate": 0.00013531669865642993, + "loss": 3.2873931884765626, + "step": 235 + }, + { + "epoch": 0.04610951008645533, + "grad_norm": 0.9984288811683655, + "learning_rate": 0.00013819577735124758, + "loss": 3.28588752746582, + "step": 240 + }, + { + "epoch": 0.04707012487992315, + "grad_norm": 1.3232122659683228, + "learning_rate": 0.00014107485604606524, + "loss": 3.2899200439453127, + "step": 245 + }, + { + "epoch": 0.04803073967339097, + "grad_norm": 1.0983723402023315, + "learning_rate": 0.0001439539347408829, + "loss": 3.287739562988281, + "step": 250 + }, + { + "epoch": 0.04899135446685879, + "grad_norm": 0.9880014657974243, + "learning_rate": 0.00014683301343570055, + "loss": 3.2914413452148437, + "step": 255 + }, + { + "epoch": 0.049951969260326606, + "grad_norm": 1.0988938808441162, + "learning_rate": 0.0001497120921305182, + "loss": 3.2892181396484377, + "step": 260 + }, + { + "epoch": 0.05091258405379443, + "grad_norm": 0.9497994780540466, + "learning_rate": 0.00015259117082533588, + "loss": 3.289907455444336, + "step": 265 + }, + { + "epoch": 0.05187319884726225, + "grad_norm": 1.4786131381988525, + "learning_rate": 0.00015547024952015354, + "loss": 3.2885669708251952, + "step": 270 + }, + { + "epoch": 0.052833813640730067, + "grad_norm": 1.1913130283355713, + "learning_rate": 0.0001583493282149712, + "loss": 3.2930023193359377, + "step": 275 + }, + { + "epoch": 0.053794428434197884, + "grad_norm": 1.0576707124710083, + "learning_rate": 0.00016122840690978887, + "loss": 3.285129165649414, + "step": 280 + }, + { + "epoch": 0.05475504322766571, + "grad_norm": 1.3840175867080688, + "learning_rate": 0.0001641074856046065, + "loss": 3.2881790161132813, + "step": 285 + }, + { + "epoch": 0.05571565802113353, + "grad_norm": 1.1912461519241333, + "learning_rate": 0.00016698656429942418, + "loss": 3.2907310485839845, + "step": 290 + }, + { + "epoch": 0.056676272814601344, + "grad_norm": 0.8021464347839355, + "learning_rate": 0.0001698656429942418, + "loss": 3.2897079467773436, + "step": 295 + }, + { + "epoch": 0.05763688760806916, + "grad_norm": 1.2219209671020508, + "learning_rate": 0.0001727447216890595, + "loss": 3.288962554931641, + "step": 300 + }, + { + "epoch": 0.05859750240153699, + "grad_norm": 1.13152015209198, + "learning_rate": 0.00017562380038387714, + "loss": 3.2893699645996093, + "step": 305 + }, + { + "epoch": 0.059558117195004805, + "grad_norm": 1.5439852476119995, + "learning_rate": 0.0001785028790786948, + "loss": 3.293457794189453, + "step": 310 + }, + { + "epoch": 0.06051873198847262, + "grad_norm": 1.4060869216918945, + "learning_rate": 0.00018138195777351245, + "loss": 3.2903827667236327, + "step": 315 + }, + { + "epoch": 0.06147934678194044, + "grad_norm": 1.0909712314605713, + "learning_rate": 0.0001842610364683301, + "loss": 3.293254852294922, + "step": 320 + }, + { + "epoch": 0.06243996157540826, + "grad_norm": 1.4758312702178955, + "learning_rate": 0.00018714011516314776, + "loss": 3.291452407836914, + "step": 325 + }, + { + "epoch": 0.06340057636887608, + "grad_norm": 1.3612134456634521, + "learning_rate": 0.00019001919385796544, + "loss": 3.2947055816650392, + "step": 330 + }, + { + "epoch": 0.0643611911623439, + "grad_norm": 1.017069935798645, + "learning_rate": 0.00019289827255278307, + "loss": 3.2914665222167967, + "step": 335 + }, + { + "epoch": 0.06532180595581172, + "grad_norm": 1.114537239074707, + "learning_rate": 0.00019577735124760075, + "loss": 3.2918880462646483, + "step": 340 + }, + { + "epoch": 0.06628242074927954, + "grad_norm": 1.359885334968567, + "learning_rate": 0.00019865642994241838, + "loss": 3.291158676147461, + "step": 345 + }, + { + "epoch": 0.06724303554274735, + "grad_norm": 1.5216479301452637, + "learning_rate": 0.00020153550863723606, + "loss": 3.2951950073242187, + "step": 350 + }, + { + "epoch": 0.06820365033621517, + "grad_norm": 1.0338709354400635, + "learning_rate": 0.00020441458733205374, + "loss": 3.2910877227783204, + "step": 355 + }, + { + "epoch": 0.069164265129683, + "grad_norm": 1.3676224946975708, + "learning_rate": 0.00020729366602687137, + "loss": 3.2979248046875, + "step": 360 + }, + { + "epoch": 0.07012487992315082, + "grad_norm": 1.508606195449829, + "learning_rate": 0.00021017274472168905, + "loss": 3.2947929382324217, + "step": 365 + }, + { + "epoch": 0.07108549471661864, + "grad_norm": 1.4694204330444336, + "learning_rate": 0.00021305182341650668, + "loss": 3.290261077880859, + "step": 370 + }, + { + "epoch": 0.07204610951008646, + "grad_norm": 1.629552960395813, + "learning_rate": 0.00021593090211132436, + "loss": 3.292864990234375, + "step": 375 + }, + { + "epoch": 0.07300672430355427, + "grad_norm": 1.3059035539627075, + "learning_rate": 0.00021880998080614202, + "loss": 3.297530746459961, + "step": 380 + }, + { + "epoch": 0.07396733909702209, + "grad_norm": 1.6734189987182617, + "learning_rate": 0.00022168905950095967, + "loss": 3.2937080383300783, + "step": 385 + }, + { + "epoch": 0.07492795389048991, + "grad_norm": 1.9466580152511597, + "learning_rate": 0.00022456813819577733, + "loss": 3.2989707946777345, + "step": 390 + }, + { + "epoch": 0.07588856868395773, + "grad_norm": 1.4589389562606812, + "learning_rate": 0.00022744721689059498, + "loss": 3.2969600677490236, + "step": 395 + }, + { + "epoch": 0.07684918347742556, + "grad_norm": 1.273814082145691, + "learning_rate": 0.00023032629558541264, + "loss": 3.2937450408935547, + "step": 400 + }, + { + "epoch": 0.07780979827089338, + "grad_norm": 1.6920170783996582, + "learning_rate": 0.00023320537428023032, + "loss": 3.2936920166015624, + "step": 405 + }, + { + "epoch": 0.0787704130643612, + "grad_norm": 1.7424761056900024, + "learning_rate": 0.00023608445297504794, + "loss": 3.29752197265625, + "step": 410 + }, + { + "epoch": 0.07973102785782901, + "grad_norm": 1.4278359413146973, + "learning_rate": 0.00023896353166986563, + "loss": 3.2931358337402346, + "step": 415 + }, + { + "epoch": 0.08069164265129683, + "grad_norm": 1.3596338033676147, + "learning_rate": 0.00024184261036468325, + "loss": 3.2951515197753904, + "step": 420 + }, + { + "epoch": 0.08165225744476465, + "grad_norm": 1.4509109258651733, + "learning_rate": 0.00024472168905950096, + "loss": 3.2955284118652344, + "step": 425 + }, + { + "epoch": 0.08261287223823247, + "grad_norm": 2.0564208030700684, + "learning_rate": 0.0002476007677543186, + "loss": 3.2958652496337892, + "step": 430 + }, + { + "epoch": 0.08357348703170028, + "grad_norm": 1.80084228515625, + "learning_rate": 0.00025047984644913627, + "loss": 3.299582672119141, + "step": 435 + }, + { + "epoch": 0.0845341018251681, + "grad_norm": 1.2458295822143555, + "learning_rate": 0.0002533589251439539, + "loss": 3.2957221984863283, + "step": 440 + }, + { + "epoch": 0.08549471661863593, + "grad_norm": 1.5943655967712402, + "learning_rate": 0.0002562380038387716, + "loss": 3.296315002441406, + "step": 445 + }, + { + "epoch": 0.08645533141210375, + "grad_norm": 1.4959216117858887, + "learning_rate": 0.00025911708253358924, + "loss": 3.2936065673828123, + "step": 450 + }, + { + "epoch": 0.08741594620557157, + "grad_norm": 0.9593575596809387, + "learning_rate": 0.0002619961612284069, + "loss": 3.2923255920410157, + "step": 455 + }, + { + "epoch": 0.08837656099903939, + "grad_norm": 1.7525019645690918, + "learning_rate": 0.00026487523992322454, + "loss": 3.297338104248047, + "step": 460 + }, + { + "epoch": 0.0893371757925072, + "grad_norm": 1.5253691673278809, + "learning_rate": 0.0002677543186180422, + "loss": 3.29698486328125, + "step": 465 + }, + { + "epoch": 0.09029779058597502, + "grad_norm": 1.7394542694091797, + "learning_rate": 0.00027063339731285985, + "loss": 3.2963401794433596, + "step": 470 + }, + { + "epoch": 0.09125840537944284, + "grad_norm": 1.3029968738555908, + "learning_rate": 0.0002735124760076775, + "loss": 3.2992328643798827, + "step": 475 + }, + { + "epoch": 0.09221902017291066, + "grad_norm": 1.3961071968078613, + "learning_rate": 0.00027639155470249516, + "loss": 3.3040702819824217, + "step": 480 + }, + { + "epoch": 0.09317963496637849, + "grad_norm": 0.9128854274749756, + "learning_rate": 0.0002792706333973128, + "loss": 3.2994720458984377, + "step": 485 + }, + { + "epoch": 0.0941402497598463, + "grad_norm": 1.39821195602417, + "learning_rate": 0.00028214971209213047, + "loss": 3.3009670257568358, + "step": 490 + }, + { + "epoch": 0.09510086455331412, + "grad_norm": 1.3767975568771362, + "learning_rate": 0.0002850287907869481, + "loss": 3.298174285888672, + "step": 495 + }, + { + "epoch": 0.09606147934678194, + "grad_norm": 1.5919218063354492, + "learning_rate": 0.0002879078694817658, + "loss": 3.2930580139160157, + "step": 500 + }, + { + "epoch": 0.09702209414024976, + "grad_norm": 2.060764789581299, + "learning_rate": 0.0002907869481765835, + "loss": 3.3007606506347655, + "step": 505 + }, + { + "epoch": 0.09798270893371758, + "grad_norm": 1.864414095878601, + "learning_rate": 0.0002936660268714011, + "loss": 3.302916717529297, + "step": 510 + }, + { + "epoch": 0.0989433237271854, + "grad_norm": 1.410025954246521, + "learning_rate": 0.0002965451055662188, + "loss": 3.297088623046875, + "step": 515 + }, + { + "epoch": 0.09990393852065321, + "grad_norm": 1.4724565744400024, + "learning_rate": 0.0002994241842610364, + "loss": 3.2969940185546873, + "step": 520 + }, + { + "epoch": 0.10086455331412104, + "grad_norm": 1.1308706998825073, + "learning_rate": 0.0002999998788910666, + "loss": 3.295259475708008, + "step": 525 + }, + { + "epoch": 0.10182516810758886, + "grad_norm": 2.0092132091522217, + "learning_rate": 0.00029999938688636, + "loss": 3.3064605712890627, + "step": 530 + }, + { + "epoch": 0.10278578290105668, + "grad_norm": 1.5756242275238037, + "learning_rate": 0.0002999985164178122, + "loss": 3.302138900756836, + "step": 535 + }, + { + "epoch": 0.1037463976945245, + "grad_norm": 1.6854816675186157, + "learning_rate": 0.00029999726748761955, + "loss": 3.29990234375, + "step": 540 + }, + { + "epoch": 0.10470701248799232, + "grad_norm": 1.5925079584121704, + "learning_rate": 0.0002999956400989331, + "loss": 3.3017444610595703, + "step": 545 + }, + { + "epoch": 0.10566762728146013, + "grad_norm": 0.9805465340614319, + "learning_rate": 0.00029999363425585907, + "loss": 3.2947647094726564, + "step": 550 + }, + { + "epoch": 0.10662824207492795, + "grad_norm": 1.143500804901123, + "learning_rate": 0.0002999912499634584, + "loss": 3.3032691955566404, + "step": 555 + }, + { + "epoch": 0.10758885686839577, + "grad_norm": 1.57884681224823, + "learning_rate": 0.00029998848722774676, + "loss": 3.3033897399902346, + "step": 560 + }, + { + "epoch": 0.10854947166186359, + "grad_norm": 1.9447771310806274, + "learning_rate": 0.00029998534605569496, + "loss": 3.2998809814453125, + "step": 565 + }, + { + "epoch": 0.10951008645533142, + "grad_norm": 1.4967366456985474, + "learning_rate": 0.0002999818264552284, + "loss": 3.2998199462890625, + "step": 570 + }, + { + "epoch": 0.11047070124879924, + "grad_norm": 1.5965176820755005, + "learning_rate": 0.0002999779284352275, + "loss": 3.2984855651855467, + "step": 575 + }, + { + "epoch": 0.11143131604226705, + "grad_norm": 1.3109067678451538, + "learning_rate": 0.0002999736520055273, + "loss": 3.293305206298828, + "step": 580 + }, + { + "epoch": 0.11239193083573487, + "grad_norm": 1.374250054359436, + "learning_rate": 0.00029996899717691766, + "loss": 3.2990135192871093, + "step": 585 + }, + { + "epoch": 0.11335254562920269, + "grad_norm": 1.8241665363311768, + "learning_rate": 0.0002999639639611432, + "loss": 3.2949737548828124, + "step": 590 + }, + { + "epoch": 0.1143131604226705, + "grad_norm": 1.1245280504226685, + "learning_rate": 0.00029995855237090327, + "loss": 3.3007225036621093, + "step": 595 + }, + { + "epoch": 0.11527377521613832, + "grad_norm": 1.290839433670044, + "learning_rate": 0.0002999527624198518, + "loss": 3.2996570587158205, + "step": 600 + }, + { + "epoch": 0.11623439000960614, + "grad_norm": 1.100143551826477, + "learning_rate": 0.00029994659412259746, + "loss": 3.2947521209716797, + "step": 605 + }, + { + "epoch": 0.11719500480307397, + "grad_norm": 1.6502522230148315, + "learning_rate": 0.0002999400474947036, + "loss": 3.295918273925781, + "step": 610 + }, + { + "epoch": 0.11815561959654179, + "grad_norm": 1.4078381061553955, + "learning_rate": 0.0002999331225526879, + "loss": 3.299440383911133, + "step": 615 + }, + { + "epoch": 0.11911623439000961, + "grad_norm": 1.4280340671539307, + "learning_rate": 0.0002999258193140227, + "loss": 3.296113967895508, + "step": 620 + }, + { + "epoch": 0.12007684918347743, + "grad_norm": 1.210864543914795, + "learning_rate": 0.00029991813779713485, + "loss": 3.292288970947266, + "step": 625 + }, + { + "epoch": 0.12103746397694524, + "grad_norm": 1.7756609916687012, + "learning_rate": 0.0002999100780214057, + "loss": 3.291936492919922, + "step": 630 + }, + { + "epoch": 0.12199807877041306, + "grad_norm": 2.3280141353607178, + "learning_rate": 0.0002999016400071708, + "loss": 3.2992881774902343, + "step": 635 + }, + { + "epoch": 0.12295869356388088, + "grad_norm": 1.2652477025985718, + "learning_rate": 0.00029989282377572006, + "loss": 3.302785110473633, + "step": 640 + }, + { + "epoch": 0.1239193083573487, + "grad_norm": 1.1368358135223389, + "learning_rate": 0.00029988362934929786, + "loss": 3.298853302001953, + "step": 645 + }, + { + "epoch": 0.12487992315081652, + "grad_norm": 1.5556303262710571, + "learning_rate": 0.0002998740567511026, + "loss": 3.2962833404541017, + "step": 650 + }, + { + "epoch": 0.12584053794428435, + "grad_norm": 1.2470626831054688, + "learning_rate": 0.000299864106005287, + "loss": 3.296977996826172, + "step": 655 + }, + { + "epoch": 0.12680115273775217, + "grad_norm": 1.8531115055084229, + "learning_rate": 0.00029985377713695775, + "loss": 3.294302749633789, + "step": 660 + }, + { + "epoch": 0.12776176753121998, + "grad_norm": 1.524565577507019, + "learning_rate": 0.0002998430701721757, + "loss": 3.2964263916015626, + "step": 665 + }, + { + "epoch": 0.1287223823246878, + "grad_norm": 1.3688217401504517, + "learning_rate": 0.00029983198513795556, + "loss": 3.2987823486328125, + "step": 670 + }, + { + "epoch": 0.12968299711815562, + "grad_norm": 1.3249455690383911, + "learning_rate": 0.000299820522062266, + "loss": 3.293718719482422, + "step": 675 + }, + { + "epoch": 0.13064361191162344, + "grad_norm": 1.0875604152679443, + "learning_rate": 0.0002998086809740296, + "loss": 3.294192886352539, + "step": 680 + }, + { + "epoch": 0.13160422670509125, + "grad_norm": 1.4566972255706787, + "learning_rate": 0.00029979646190312264, + "loss": 3.2981857299804687, + "step": 685 + }, + { + "epoch": 0.13256484149855907, + "grad_norm": 1.506569743156433, + "learning_rate": 0.0002997838648803751, + "loss": 3.2970245361328123, + "step": 690 + }, + { + "epoch": 0.1335254562920269, + "grad_norm": 1.2062523365020752, + "learning_rate": 0.00029977088993757045, + "loss": 3.294635772705078, + "step": 695 + }, + { + "epoch": 0.1344860710854947, + "grad_norm": 1.4316554069519043, + "learning_rate": 0.0002997575371074458, + "loss": 3.294428253173828, + "step": 700 + }, + { + "epoch": 0.13544668587896252, + "grad_norm": 1.0954806804656982, + "learning_rate": 0.0002997438064236919, + "loss": 3.2926597595214844, + "step": 705 + }, + { + "epoch": 0.13640730067243034, + "grad_norm": 1.2398245334625244, + "learning_rate": 0.0002997296979209526, + "loss": 3.293048858642578, + "step": 710 + }, + { + "epoch": 0.1373679154658982, + "grad_norm": 1.7894244194030762, + "learning_rate": 0.00029971521163482494, + "loss": 3.2915740966796876, + "step": 715 + }, + { + "epoch": 0.138328530259366, + "grad_norm": 1.1798651218414307, + "learning_rate": 0.0002997003476018595, + "loss": 3.2928890228271483, + "step": 720 + }, + { + "epoch": 0.13928914505283382, + "grad_norm": 1.4055883884429932, + "learning_rate": 0.0002996851058595596, + "loss": 3.295462417602539, + "step": 725 + }, + { + "epoch": 0.14024975984630164, + "grad_norm": 1.3189594745635986, + "learning_rate": 0.00029966948644638184, + "loss": 3.2915382385253906, + "step": 730 + }, + { + "epoch": 0.14121037463976946, + "grad_norm": 2.0678322315216064, + "learning_rate": 0.0002996534894017356, + "loss": 3.2929306030273438, + "step": 735 + }, + { + "epoch": 0.14217098943323728, + "grad_norm": 1.0185853242874146, + "learning_rate": 0.000299637114765983, + "loss": 3.2973453521728517, + "step": 740 + }, + { + "epoch": 0.1431316042267051, + "grad_norm": 1.1556357145309448, + "learning_rate": 0.00029962036258043893, + "loss": 3.2935462951660157, + "step": 745 + }, + { + "epoch": 0.1440922190201729, + "grad_norm": 1.392240285873413, + "learning_rate": 0.00029960323288737097, + "loss": 3.2977447509765625, + "step": 750 + }, + { + "epoch": 0.14505283381364073, + "grad_norm": 1.6778427362442017, + "learning_rate": 0.000299585725729999, + "loss": 3.2953746795654295, + "step": 755 + }, + { + "epoch": 0.14601344860710855, + "grad_norm": 0.8692209720611572, + "learning_rate": 0.00029956784115249535, + "loss": 3.2932708740234373, + "step": 760 + }, + { + "epoch": 0.14697406340057637, + "grad_norm": 1.0066865682601929, + "learning_rate": 0.0002995495791999847, + "loss": 3.291333770751953, + "step": 765 + }, + { + "epoch": 0.14793467819404418, + "grad_norm": 1.1920307874679565, + "learning_rate": 0.0002995309399185439, + "loss": 3.2948654174804686, + "step": 770 + }, + { + "epoch": 0.148895292987512, + "grad_norm": 0.9732314348220825, + "learning_rate": 0.0002995119233552016, + "loss": 3.2924430847167967, + "step": 775 + }, + { + "epoch": 0.14985590778097982, + "grad_norm": 1.306164264678955, + "learning_rate": 0.0002994925295579386, + "loss": 3.2954177856445312, + "step": 780 + }, + { + "epoch": 0.15081652257444764, + "grad_norm": 1.2705974578857422, + "learning_rate": 0.0002994727585756875, + "loss": 3.291736602783203, + "step": 785 + }, + { + "epoch": 0.15177713736791545, + "grad_norm": 1.4319415092468262, + "learning_rate": 0.00029945261045833256, + "loss": 3.2872940063476563, + "step": 790 + }, + { + "epoch": 0.15273775216138327, + "grad_norm": 1.5740435123443604, + "learning_rate": 0.0002994320852567094, + "loss": 3.2921607971191404, + "step": 795 + }, + { + "epoch": 0.15369836695485112, + "grad_norm": 1.2682417631149292, + "learning_rate": 0.0002994111830226053, + "loss": 3.2847610473632813, + "step": 800 + }, + { + "epoch": 0.15465898174831894, + "grad_norm": 1.3516907691955566, + "learning_rate": 0.0002993899038087587, + "loss": 3.2877117156982423, + "step": 805 + }, + { + "epoch": 0.15561959654178675, + "grad_norm": 1.1541163921356201, + "learning_rate": 0.00029936824766885927, + "loss": 3.28739013671875, + "step": 810 + }, + { + "epoch": 0.15658021133525457, + "grad_norm": 1.4818735122680664, + "learning_rate": 0.0002993462146575477, + "loss": 3.2846580505371095, + "step": 815 + }, + { + "epoch": 0.1575408261287224, + "grad_norm": 1.0636875629425049, + "learning_rate": 0.00029932380483041547, + "loss": 3.2877876281738283, + "step": 820 + }, + { + "epoch": 0.1585014409221902, + "grad_norm": 1.1833419799804688, + "learning_rate": 0.00029930101824400486, + "loss": 3.289693069458008, + "step": 825 + }, + { + "epoch": 0.15946205571565802, + "grad_norm": 1.19170343875885, + "learning_rate": 0.00029927785495580887, + "loss": 3.283536911010742, + "step": 830 + }, + { + "epoch": 0.16042267050912584, + "grad_norm": 1.074803352355957, + "learning_rate": 0.00029925431502427077, + "loss": 3.2888442993164064, + "step": 835 + }, + { + "epoch": 0.16138328530259366, + "grad_norm": 1.205644130706787, + "learning_rate": 0.00029923039850878423, + "loss": 3.2893348693847657, + "step": 840 + }, + { + "epoch": 0.16234390009606148, + "grad_norm": 1.6655492782592773, + "learning_rate": 0.00029920610546969314, + "loss": 3.2857383728027343, + "step": 845 + }, + { + "epoch": 0.1633045148895293, + "grad_norm": 1.3908863067626953, + "learning_rate": 0.0002991814359682912, + "loss": 3.2931854248046877, + "step": 850 + }, + { + "epoch": 0.1642651296829971, + "grad_norm": 1.3854031562805176, + "learning_rate": 0.00029915639006682223, + "loss": 3.291329193115234, + "step": 855 + }, + { + "epoch": 0.16522574447646493, + "grad_norm": 1.3702648878097534, + "learning_rate": 0.0002991309678284795, + "loss": 3.283129119873047, + "step": 860 + }, + { + "epoch": 0.16618635926993275, + "grad_norm": 1.3857735395431519, + "learning_rate": 0.000299105169317406, + "loss": 3.285200500488281, + "step": 865 + }, + { + "epoch": 0.16714697406340057, + "grad_norm": 1.0670974254608154, + "learning_rate": 0.000299078994598694, + "loss": 3.287001037597656, + "step": 870 + }, + { + "epoch": 0.16810758885686838, + "grad_norm": 1.2670496702194214, + "learning_rate": 0.000299052443738385, + "loss": 3.28797607421875, + "step": 875 + }, + { + "epoch": 0.1690682036503362, + "grad_norm": 1.4439146518707275, + "learning_rate": 0.00029902551680346956, + "loss": 3.2858448028564453, + "step": 880 + }, + { + "epoch": 0.17002881844380405, + "grad_norm": 1.4315037727355957, + "learning_rate": 0.000298998213861887, + "loss": 3.28741340637207, + "step": 885 + }, + { + "epoch": 0.17098943323727187, + "grad_norm": 1.227054238319397, + "learning_rate": 0.0002989705349825256, + "loss": 3.285395050048828, + "step": 890 + }, + { + "epoch": 0.17195004803073968, + "grad_norm": 1.9081028699874878, + "learning_rate": 0.0002989424802352218, + "loss": 3.2881591796875, + "step": 895 + }, + { + "epoch": 0.1729106628242075, + "grad_norm": 1.5766661167144775, + "learning_rate": 0.0002989140496907609, + "loss": 3.282776641845703, + "step": 900 + }, + { + "epoch": 0.17387127761767532, + "grad_norm": 1.3651994466781616, + "learning_rate": 0.00029888524342087583, + "loss": 3.288460540771484, + "step": 905 + }, + { + "epoch": 0.17483189241114314, + "grad_norm": 1.1733394861221313, + "learning_rate": 0.00029885606149824786, + "loss": 3.2853572845458983, + "step": 910 + }, + { + "epoch": 0.17579250720461095, + "grad_norm": 1.798338770866394, + "learning_rate": 0.00029882650399650596, + "loss": 3.2855796813964844, + "step": 915 + }, + { + "epoch": 0.17675312199807877, + "grad_norm": 1.5101557970046997, + "learning_rate": 0.00029879657099022677, + "loss": 3.2913845062255858, + "step": 920 + }, + { + "epoch": 0.1777137367915466, + "grad_norm": 1.1141709089279175, + "learning_rate": 0.0002987662625549343, + "loss": 3.2880714416503904, + "step": 925 + }, + { + "epoch": 0.1786743515850144, + "grad_norm": 1.4672577381134033, + "learning_rate": 0.00029873557876709984, + "loss": 3.2870162963867187, + "step": 930 + }, + { + "epoch": 0.17963496637848222, + "grad_norm": 1.2008938789367676, + "learning_rate": 0.00029870451970414176, + "loss": 3.281757354736328, + "step": 935 + }, + { + "epoch": 0.18059558117195004, + "grad_norm": 1.4659417867660522, + "learning_rate": 0.00029867308544442525, + "loss": 3.2835357666015623, + "step": 940 + }, + { + "epoch": 0.18155619596541786, + "grad_norm": 1.200221300125122, + "learning_rate": 0.0002986412760672622, + "loss": 3.2824310302734374, + "step": 945 + }, + { + "epoch": 0.18251681075888568, + "grad_norm": 1.1019620895385742, + "learning_rate": 0.0002986090916529109, + "loss": 3.2842033386230467, + "step": 950 + }, + { + "epoch": 0.1834774255523535, + "grad_norm": 1.3469750881195068, + "learning_rate": 0.00029857653228257587, + "loss": 3.285472106933594, + "step": 955 + }, + { + "epoch": 0.1844380403458213, + "grad_norm": 1.3108643293380737, + "learning_rate": 0.0002985435980384078, + "loss": 3.285649871826172, + "step": 960 + }, + { + "epoch": 0.18539865513928913, + "grad_norm": 1.452244758605957, + "learning_rate": 0.0002985102890035032, + "loss": 3.2803878784179688, + "step": 965 + }, + { + "epoch": 0.18635926993275698, + "grad_norm": 1.229927897453308, + "learning_rate": 0.0002984766052619041, + "loss": 3.283854675292969, + "step": 970 + }, + { + "epoch": 0.1873198847262248, + "grad_norm": 1.1009323596954346, + "learning_rate": 0.0002984425468985982, + "loss": 3.2806129455566406, + "step": 975 + }, + { + "epoch": 0.1882804995196926, + "grad_norm": 1.5917284488677979, + "learning_rate": 0.00029840811399951806, + "loss": 3.282550048828125, + "step": 980 + }, + { + "epoch": 0.18924111431316043, + "grad_norm": 1.1452780961990356, + "learning_rate": 0.00029837330665154156, + "loss": 3.2815521240234373, + "step": 985 + }, + { + "epoch": 0.19020172910662825, + "grad_norm": 0.9764254093170166, + "learning_rate": 0.000298338124942491, + "loss": 3.284914016723633, + "step": 990 + }, + { + "epoch": 0.19116234390009607, + "grad_norm": 1.045482873916626, + "learning_rate": 0.0002983025689611337, + "loss": 3.2865413665771483, + "step": 995 + }, + { + "epoch": 0.19212295869356388, + "grad_norm": 1.3126083612442017, + "learning_rate": 0.0002982666387971809, + "loss": 3.281140899658203, + "step": 1000 + }, + { + "epoch": 0.1930835734870317, + "grad_norm": 1.3957022428512573, + "learning_rate": 0.0002982303345412881, + "loss": 3.2785770416259767, + "step": 1005 + }, + { + "epoch": 0.19404418828049952, + "grad_norm": 1.2035447359085083, + "learning_rate": 0.0002981936562850547, + "loss": 3.279229736328125, + "step": 1010 + }, + { + "epoch": 0.19500480307396734, + "grad_norm": 0.8585781455039978, + "learning_rate": 0.00029815660412102366, + "loss": 3.2802391052246094, + "step": 1015 + }, + { + "epoch": 0.19596541786743515, + "grad_norm": 1.203058123588562, + "learning_rate": 0.00029811917814268144, + "loss": 3.2790596008300783, + "step": 1020 + }, + { + "epoch": 0.19692603266090297, + "grad_norm": 1.4647698402404785, + "learning_rate": 0.0002980813784444576, + "loss": 3.2819541931152343, + "step": 1025 + }, + { + "epoch": 0.1978866474543708, + "grad_norm": 1.4942032098770142, + "learning_rate": 0.0002980432051217246, + "loss": 3.2802181243896484, + "step": 1030 + }, + { + "epoch": 0.1988472622478386, + "grad_norm": 1.1856545209884644, + "learning_rate": 0.0002980046582707978, + "loss": 3.290009307861328, + "step": 1035 + }, + { + "epoch": 0.19980787704130643, + "grad_norm": 1.166818380355835, + "learning_rate": 0.00029796573798893464, + "loss": 3.28170166015625, + "step": 1040 + }, + { + "epoch": 0.20076849183477424, + "grad_norm": 1.310017466545105, + "learning_rate": 0.0002979264443743352, + "loss": 3.282215118408203, + "step": 1045 + }, + { + "epoch": 0.2017291066282421, + "grad_norm": 1.7260593175888062, + "learning_rate": 0.00029788677752614116, + "loss": 3.2847213745117188, + "step": 1050 + }, + { + "epoch": 0.2026897214217099, + "grad_norm": 1.083449125289917, + "learning_rate": 0.0002978467375444361, + "loss": 3.278899002075195, + "step": 1055 + }, + { + "epoch": 0.20365033621517772, + "grad_norm": 1.2556031942367554, + "learning_rate": 0.000297806324530245, + "loss": 3.280004119873047, + "step": 1060 + }, + { + "epoch": 0.20461095100864554, + "grad_norm": 0.8801208138465881, + "learning_rate": 0.00029776553858553405, + "loss": 3.2757003784179686, + "step": 1065 + }, + { + "epoch": 0.20557156580211336, + "grad_norm": 1.419739007949829, + "learning_rate": 0.0002977243798132103, + "loss": 3.2807285308837892, + "step": 1070 + }, + { + "epoch": 0.20653218059558118, + "grad_norm": 1.3385144472122192, + "learning_rate": 0.00029768284831712175, + "loss": 3.280821990966797, + "step": 1075 + }, + { + "epoch": 0.207492795389049, + "grad_norm": 1.0410194396972656, + "learning_rate": 0.0002976409442020564, + "loss": 3.279698944091797, + "step": 1080 + }, + { + "epoch": 0.2084534101825168, + "grad_norm": 1.2109322547912598, + "learning_rate": 0.0002975986675737427, + "loss": 3.2776737213134766, + "step": 1085 + }, + { + "epoch": 0.20941402497598463, + "grad_norm": 1.598090648651123, + "learning_rate": 0.00029755601853884893, + "loss": 3.277222442626953, + "step": 1090 + }, + { + "epoch": 0.21037463976945245, + "grad_norm": 1.0197858810424805, + "learning_rate": 0.000297512997204983, + "loss": 3.2803329467773437, + "step": 1095 + }, + { + "epoch": 0.21133525456292027, + "grad_norm": 0.9741032123565674, + "learning_rate": 0.0002974696036806922, + "loss": 3.2853607177734374, + "step": 1100 + }, + { + "epoch": 0.21229586935638808, + "grad_norm": 1.3451876640319824, + "learning_rate": 0.00029742583807546274, + "loss": 3.276333236694336, + "step": 1105 + }, + { + "epoch": 0.2132564841498559, + "grad_norm": 1.2260080575942993, + "learning_rate": 0.00029738170049971974, + "loss": 3.2778556823730467, + "step": 1110 + }, + { + "epoch": 0.21421709894332372, + "grad_norm": 1.4238629341125488, + "learning_rate": 0.00029733719106482694, + "loss": 3.27607421875, + "step": 1115 + }, + { + "epoch": 0.21517771373679154, + "grad_norm": 1.0617904663085938, + "learning_rate": 0.0002972923098830861, + "loss": 3.278319549560547, + "step": 1120 + }, + { + "epoch": 0.21613832853025935, + "grad_norm": 1.3632889986038208, + "learning_rate": 0.0002972470570677371, + "loss": 3.2795860290527346, + "step": 1125 + }, + { + "epoch": 0.21709894332372717, + "grad_norm": 1.2691421508789062, + "learning_rate": 0.00029720143273295735, + "loss": 3.277021026611328, + "step": 1130 + }, + { + "epoch": 0.21805955811719502, + "grad_norm": 1.2249642610549927, + "learning_rate": 0.0002971554369938619, + "loss": 3.274774169921875, + "step": 1135 + }, + { + "epoch": 0.21902017291066284, + "grad_norm": 1.6009410619735718, + "learning_rate": 0.0002971090699665025, + "loss": 3.2758060455322267, + "step": 1140 + }, + { + "epoch": 0.21998078770413065, + "grad_norm": 1.1862974166870117, + "learning_rate": 0.0002970623317678681, + "loss": 3.278969955444336, + "step": 1145 + }, + { + "epoch": 0.22094140249759847, + "grad_norm": 1.1896966695785522, + "learning_rate": 0.00029701522251588395, + "loss": 3.277644729614258, + "step": 1150 + }, + { + "epoch": 0.2219020172910663, + "grad_norm": 1.2047041654586792, + "learning_rate": 0.00029696774232941135, + "loss": 3.276856231689453, + "step": 1155 + }, + { + "epoch": 0.2228626320845341, + "grad_norm": 1.4511812925338745, + "learning_rate": 0.0002969198913282479, + "loss": 3.2762985229492188, + "step": 1160 + }, + { + "epoch": 0.22382324687800192, + "grad_norm": 1.0419248342514038, + "learning_rate": 0.0002968716696331265, + "loss": 3.2781463623046876, + "step": 1165 + }, + { + "epoch": 0.22478386167146974, + "grad_norm": 1.0537165403366089, + "learning_rate": 0.00029682307736571533, + "loss": 3.280584716796875, + "step": 1170 + }, + { + "epoch": 0.22574447646493756, + "grad_norm": 0.9234645366668701, + "learning_rate": 0.00029677411464861783, + "loss": 3.2776565551757812, + "step": 1175 + }, + { + "epoch": 0.22670509125840538, + "grad_norm": 1.0644463300704956, + "learning_rate": 0.00029672478160537184, + "loss": 3.2758762359619142, + "step": 1180 + }, + { + "epoch": 0.2276657060518732, + "grad_norm": 0.9237310290336609, + "learning_rate": 0.00029667507836044984, + "loss": 3.2722846984863283, + "step": 1185 + }, + { + "epoch": 0.228626320845341, + "grad_norm": 1.5153244733810425, + "learning_rate": 0.00029662500503925803, + "loss": 3.2741416931152343, + "step": 1190 + }, + { + "epoch": 0.22958693563880883, + "grad_norm": 1.3842270374298096, + "learning_rate": 0.00029657456176813664, + "loss": 3.2764862060546873, + "step": 1195 + }, + { + "epoch": 0.23054755043227665, + "grad_norm": 1.1134364604949951, + "learning_rate": 0.0002965237486743592, + "loss": 3.275590515136719, + "step": 1200 + }, + { + "epoch": 0.23150816522574447, + "grad_norm": 1.4811969995498657, + "learning_rate": 0.0002964725658861323, + "loss": 3.2718353271484375, + "step": 1205 + }, + { + "epoch": 0.23246878001921228, + "grad_norm": 1.2888314723968506, + "learning_rate": 0.0002964210135325955, + "loss": 3.2731605529785157, + "step": 1210 + }, + { + "epoch": 0.2334293948126801, + "grad_norm": 1.530621886253357, + "learning_rate": 0.0002963690917438206, + "loss": 3.2751121520996094, + "step": 1215 + }, + { + "epoch": 0.23439000960614795, + "grad_norm": 1.0951229333877563, + "learning_rate": 0.0002963168006508116, + "loss": 3.2797927856445312, + "step": 1220 + }, + { + "epoch": 0.23535062439961577, + "grad_norm": 1.0309784412384033, + "learning_rate": 0.00029626414038550424, + "loss": 3.279781723022461, + "step": 1225 + }, + { + "epoch": 0.23631123919308358, + "grad_norm": 1.211702823638916, + "learning_rate": 0.00029621111108076584, + "loss": 3.276560592651367, + "step": 1230 + }, + { + "epoch": 0.2372718539865514, + "grad_norm": 1.0704128742218018, + "learning_rate": 0.0002961577128703948, + "loss": 3.2736587524414062, + "step": 1235 + }, + { + "epoch": 0.23823246878001922, + "grad_norm": 1.4519094228744507, + "learning_rate": 0.0002961039458891202, + "loss": 3.274006652832031, + "step": 1240 + }, + { + "epoch": 0.23919308357348704, + "grad_norm": 1.2808977365493774, + "learning_rate": 0.00029604981027260167, + "loss": 3.27347412109375, + "step": 1245 + }, + { + "epoch": 0.24015369836695485, + "grad_norm": 0.8584306836128235, + "learning_rate": 0.00029599530615742896, + "loss": 3.2714912414550783, + "step": 1250 + }, + { + "epoch": 0.24111431316042267, + "grad_norm": 1.1275231838226318, + "learning_rate": 0.0002959404336811215, + "loss": 3.2730926513671874, + "step": 1255 + }, + { + "epoch": 0.2420749279538905, + "grad_norm": 1.422745943069458, + "learning_rate": 0.0002958851929821283, + "loss": 3.27635498046875, + "step": 1260 + }, + { + "epoch": 0.2430355427473583, + "grad_norm": 1.0575655698776245, + "learning_rate": 0.00029582958419982713, + "loss": 3.272149658203125, + "step": 1265 + }, + { + "epoch": 0.24399615754082613, + "grad_norm": 1.2576707601547241, + "learning_rate": 0.0002957736074745248, + "loss": 3.275560760498047, + "step": 1270 + }, + { + "epoch": 0.24495677233429394, + "grad_norm": 1.328376054763794, + "learning_rate": 0.0002957172629474562, + "loss": 3.272699737548828, + "step": 1275 + }, + { + "epoch": 0.24591738712776176, + "grad_norm": 1.0527936220169067, + "learning_rate": 0.00029566055076078447, + "loss": 3.269004058837891, + "step": 1280 + }, + { + "epoch": 0.24687800192122958, + "grad_norm": 1.1557375192642212, + "learning_rate": 0.00029560347105760023, + "loss": 3.2709724426269533, + "step": 1285 + }, + { + "epoch": 0.2478386167146974, + "grad_norm": 1.1958216428756714, + "learning_rate": 0.00029554602398192137, + "loss": 3.276508331298828, + "step": 1290 + }, + { + "epoch": 0.24879923150816521, + "grad_norm": 1.0567305088043213, + "learning_rate": 0.00029548820967869287, + "loss": 3.270316314697266, + "step": 1295 + }, + { + "epoch": 0.24975984630163303, + "grad_norm": 1.1860175132751465, + "learning_rate": 0.00029543002829378605, + "loss": 3.274994659423828, + "step": 1300 + }, + { + "epoch": 0.2507204610951009, + "grad_norm": 1.2902470827102661, + "learning_rate": 0.0002953714799739986, + "loss": 3.2740650177001953, + "step": 1305 + }, + { + "epoch": 0.2516810758885687, + "grad_norm": 1.0249444246292114, + "learning_rate": 0.0002953125648670538, + "loss": 3.273143768310547, + "step": 1310 + }, + { + "epoch": 0.2526416906820365, + "grad_norm": 1.051010251045227, + "learning_rate": 0.00029525328312160064, + "loss": 3.270255279541016, + "step": 1315 + }, + { + "epoch": 0.25360230547550433, + "grad_norm": 1.5545085668563843, + "learning_rate": 0.00029519363488721297, + "loss": 3.270277404785156, + "step": 1320 + }, + { + "epoch": 0.25456292026897215, + "grad_norm": 1.096474289894104, + "learning_rate": 0.0002951336203143895, + "loss": 3.2695423126220704, + "step": 1325 + }, + { + "epoch": 0.25552353506243997, + "grad_norm": 1.3266576528549194, + "learning_rate": 0.0002950732395545531, + "loss": 3.268444061279297, + "step": 1330 + }, + { + "epoch": 0.2564841498559078, + "grad_norm": 1.3308000564575195, + "learning_rate": 0.0002950124927600505, + "loss": 3.270121765136719, + "step": 1335 + }, + { + "epoch": 0.2574447646493756, + "grad_norm": 0.9661091566085815, + "learning_rate": 0.0002949513800841523, + "loss": 3.2689208984375, + "step": 1340 + }, + { + "epoch": 0.2584053794428434, + "grad_norm": 1.0783677101135254, + "learning_rate": 0.0002948899016810519, + "loss": 3.2726470947265627, + "step": 1345 + }, + { + "epoch": 0.25936599423631124, + "grad_norm": 1.44166898727417, + "learning_rate": 0.00029482805770586566, + "loss": 3.2676158905029298, + "step": 1350 + }, + { + "epoch": 0.26032660902977905, + "grad_norm": 0.9048498868942261, + "learning_rate": 0.0002947658483146323, + "loss": 3.268976593017578, + "step": 1355 + }, + { + "epoch": 0.2612872238232469, + "grad_norm": 0.9976827502250671, + "learning_rate": 0.00029470327366431257, + "loss": 3.2651641845703123, + "step": 1360 + }, + { + "epoch": 0.2622478386167147, + "grad_norm": 1.4668549299240112, + "learning_rate": 0.00029464033391278857, + "loss": 3.267919158935547, + "step": 1365 + }, + { + "epoch": 0.2632084534101825, + "grad_norm": 1.3420720100402832, + "learning_rate": 0.0002945770292188638, + "loss": 3.2697959899902345, + "step": 1370 + }, + { + "epoch": 0.2641690682036503, + "grad_norm": 1.3224592208862305, + "learning_rate": 0.00029451335974226255, + "loss": 3.2723613739013673, + "step": 1375 + }, + { + "epoch": 0.26512968299711814, + "grad_norm": 1.1924461126327515, + "learning_rate": 0.0002944493256436294, + "loss": 3.270499038696289, + "step": 1380 + }, + { + "epoch": 0.26609029779058596, + "grad_norm": 1.2803032398223877, + "learning_rate": 0.000294384927084529, + "loss": 3.269664001464844, + "step": 1385 + }, + { + "epoch": 0.2670509125840538, + "grad_norm": 1.0187225341796875, + "learning_rate": 0.0002943201642274455, + "loss": 3.2647491455078126, + "step": 1390 + }, + { + "epoch": 0.2680115273775216, + "grad_norm": 1.3384113311767578, + "learning_rate": 0.00029425503723578216, + "loss": 3.2721004486083984, + "step": 1395 + }, + { + "epoch": 0.2689721421709894, + "grad_norm": 1.2812883853912354, + "learning_rate": 0.00029418954627386115, + "loss": 3.269911193847656, + "step": 1400 + }, + { + "epoch": 0.26993275696445723, + "grad_norm": 1.1734299659729004, + "learning_rate": 0.0002941236915069228, + "loss": 3.2672080993652344, + "step": 1405 + }, + { + "epoch": 0.27089337175792505, + "grad_norm": 1.069024682044983, + "learning_rate": 0.00029405747310112557, + "loss": 3.2702381134033205, + "step": 1410 + }, + { + "epoch": 0.27185398655139287, + "grad_norm": 1.286638617515564, + "learning_rate": 0.0002939908912235452, + "loss": 3.2697906494140625, + "step": 1415 + }, + { + "epoch": 0.2728146013448607, + "grad_norm": 1.1734787225723267, + "learning_rate": 0.0002939239460421746, + "loss": 3.2664981842041017, + "step": 1420 + }, + { + "epoch": 0.2737752161383285, + "grad_norm": 1.150191307067871, + "learning_rate": 0.00029385663772592334, + "loss": 3.2677589416503907, + "step": 1425 + }, + { + "epoch": 0.2747358309317964, + "grad_norm": 1.3920680284500122, + "learning_rate": 0.00029378896644461716, + "loss": 3.2696159362792967, + "step": 1430 + }, + { + "epoch": 0.2756964457252642, + "grad_norm": 1.1501007080078125, + "learning_rate": 0.0002937209323689978, + "loss": 3.268462371826172, + "step": 1435 + }, + { + "epoch": 0.276657060518732, + "grad_norm": 1.2256152629852295, + "learning_rate": 0.00029365253567072207, + "loss": 3.269298553466797, + "step": 1440 + }, + { + "epoch": 0.27761767531219983, + "grad_norm": 1.2467095851898193, + "learning_rate": 0.00029358377652236186, + "loss": 3.2696155548095702, + "step": 1445 + }, + { + "epoch": 0.27857829010566765, + "grad_norm": 0.7786490321159363, + "learning_rate": 0.00029351465509740366, + "loss": 3.265594482421875, + "step": 1450 + }, + { + "epoch": 0.27953890489913547, + "grad_norm": 0.9412602186203003, + "learning_rate": 0.00029344517157024773, + "loss": 3.2679397583007814, + "step": 1455 + }, + { + "epoch": 0.2804995196926033, + "grad_norm": 1.1962471008300781, + "learning_rate": 0.0002933753261162084, + "loss": 3.276123046875, + "step": 1460 + }, + { + "epoch": 0.2814601344860711, + "grad_norm": 1.1592360734939575, + "learning_rate": 0.00029330511891151263, + "loss": 3.2704566955566405, + "step": 1465 + }, + { + "epoch": 0.2824207492795389, + "grad_norm": 0.8931114673614502, + "learning_rate": 0.00029323455013330064, + "loss": 3.2649574279785156, + "step": 1470 + }, + { + "epoch": 0.28338136407300674, + "grad_norm": 0.8980585932731628, + "learning_rate": 0.00029316361995962453, + "loss": 3.2683528900146483, + "step": 1475 + }, + { + "epoch": 0.28434197886647455, + "grad_norm": 1.330986499786377, + "learning_rate": 0.0002930923285694485, + "loss": 3.2716033935546873, + "step": 1480 + }, + { + "epoch": 0.28530259365994237, + "grad_norm": 1.0797970294952393, + "learning_rate": 0.000293020676142648, + "loss": 3.268834686279297, + "step": 1485 + }, + { + "epoch": 0.2862632084534102, + "grad_norm": 1.2156567573547363, + "learning_rate": 0.00029294866286000946, + "loss": 3.2620380401611326, + "step": 1490 + }, + { + "epoch": 0.287223823246878, + "grad_norm": 1.3845075368881226, + "learning_rate": 0.00029287628890322986, + "loss": 3.2627891540527343, + "step": 1495 + }, + { + "epoch": 0.2881844380403458, + "grad_norm": 1.2039449214935303, + "learning_rate": 0.00029280355445491596, + "loss": 3.2676162719726562, + "step": 1500 + }, + { + "epoch": 0.28914505283381364, + "grad_norm": 1.2778565883636475, + "learning_rate": 0.00029273045969858437, + "loss": 3.267996978759766, + "step": 1505 + }, + { + "epoch": 0.29010566762728146, + "grad_norm": 0.9892801642417908, + "learning_rate": 0.0002926570048186606, + "loss": 3.2660484313964844, + "step": 1510 + }, + { + "epoch": 0.2910662824207493, + "grad_norm": 1.0218427181243896, + "learning_rate": 0.00029258319000047885, + "loss": 3.269457244873047, + "step": 1515 + }, + { + "epoch": 0.2920268972142171, + "grad_norm": 1.0072425603866577, + "learning_rate": 0.00029250901543028144, + "loss": 3.267464828491211, + "step": 1520 + }, + { + "epoch": 0.2929875120076849, + "grad_norm": 1.4437980651855469, + "learning_rate": 0.00029243448129521847, + "loss": 3.2681037902832033, + "step": 1525 + }, + { + "epoch": 0.29394812680115273, + "grad_norm": 0.8555430769920349, + "learning_rate": 0.0002923595877833472, + "loss": 3.2668888092041017, + "step": 1530 + }, + { + "epoch": 0.29490874159462055, + "grad_norm": 1.2450693845748901, + "learning_rate": 0.0002922843350836317, + "loss": 3.2664356231689453, + "step": 1535 + }, + { + "epoch": 0.29586935638808837, + "grad_norm": 0.9553253650665283, + "learning_rate": 0.00029220872338594215, + "loss": 3.263041687011719, + "step": 1540 + }, + { + "epoch": 0.2968299711815562, + "grad_norm": 1.2076423168182373, + "learning_rate": 0.0002921327528810547, + "loss": 3.266950988769531, + "step": 1545 + }, + { + "epoch": 0.297790585975024, + "grad_norm": 2.1191205978393555, + "learning_rate": 0.00029205642376065066, + "loss": 3.2656558990478515, + "step": 1550 + }, + { + "epoch": 0.2987512007684918, + "grad_norm": 1.4229607582092285, + "learning_rate": 0.0002919797362173163, + "loss": 3.2715381622314452, + "step": 1555 + }, + { + "epoch": 0.29971181556195964, + "grad_norm": 1.2368606328964233, + "learning_rate": 0.00029190269044454206, + "loss": 3.2648460388183596, + "step": 1560 + }, + { + "epoch": 0.30067243035542746, + "grad_norm": 1.266961693763733, + "learning_rate": 0.00029182528663672245, + "loss": 3.2667465209960938, + "step": 1565 + }, + { + "epoch": 0.3016330451488953, + "grad_norm": 0.9238030314445496, + "learning_rate": 0.00029174752498915515, + "loss": 3.262348175048828, + "step": 1570 + }, + { + "epoch": 0.3025936599423631, + "grad_norm": 1.0845513343811035, + "learning_rate": 0.0002916694056980408, + "loss": 3.266324996948242, + "step": 1575 + }, + { + "epoch": 0.3035542747358309, + "grad_norm": 1.050512433052063, + "learning_rate": 0.0002915909289604823, + "loss": 3.2667007446289062, + "step": 1580 + }, + { + "epoch": 0.3045148895292987, + "grad_norm": 1.25586998462677, + "learning_rate": 0.00029151209497448467, + "loss": 3.262495422363281, + "step": 1585 + }, + { + "epoch": 0.30547550432276654, + "grad_norm": 1.1228755712509155, + "learning_rate": 0.000291432903938954, + "loss": 3.26649169921875, + "step": 1590 + }, + { + "epoch": 0.30643611911623436, + "grad_norm": 1.2967936992645264, + "learning_rate": 0.0002913533560536975, + "loss": 3.264292526245117, + "step": 1595 + }, + { + "epoch": 0.30739673390970224, + "grad_norm": 1.1702841520309448, + "learning_rate": 0.00029127345151942253, + "loss": 3.2614276885986326, + "step": 1600 + }, + { + "epoch": 0.30835734870317005, + "grad_norm": 1.438833236694336, + "learning_rate": 0.00029119319053773655, + "loss": 3.2649909973144533, + "step": 1605 + }, + { + "epoch": 0.30931796349663787, + "grad_norm": 1.1444823741912842, + "learning_rate": 0.00029111257331114617, + "loss": 3.2618698120117187, + "step": 1610 + }, + { + "epoch": 0.3102785782901057, + "grad_norm": 1.2421047687530518, + "learning_rate": 0.00029103160004305693, + "loss": 3.260860061645508, + "step": 1615 + }, + { + "epoch": 0.3112391930835735, + "grad_norm": 1.3885422945022583, + "learning_rate": 0.00029095027093777276, + "loss": 3.2636131286621093, + "step": 1620 + }, + { + "epoch": 0.3121998078770413, + "grad_norm": 1.403899073600769, + "learning_rate": 0.00029086858620049527, + "loss": 3.261060333251953, + "step": 1625 + }, + { + "epoch": 0.31316042267050914, + "grad_norm": 1.4193172454833984, + "learning_rate": 0.0002907865460373234, + "loss": 3.2587600708007813, + "step": 1630 + }, + { + "epoch": 0.31412103746397696, + "grad_norm": 0.9989815354347229, + "learning_rate": 0.00029070415065525295, + "loss": 3.2618736267089843, + "step": 1635 + }, + { + "epoch": 0.3150816522574448, + "grad_norm": 1.0307117700576782, + "learning_rate": 0.00029062140026217596, + "loss": 3.25679931640625, + "step": 1640 + }, + { + "epoch": 0.3160422670509126, + "grad_norm": 1.3937115669250488, + "learning_rate": 0.0002905382950668801, + "loss": 3.2610740661621094, + "step": 1645 + }, + { + "epoch": 0.3170028818443804, + "grad_norm": 1.0371273756027222, + "learning_rate": 0.0002904548352790483, + "loss": 3.264708709716797, + "step": 1650 + }, + { + "epoch": 0.31796349663784823, + "grad_norm": 1.202774167060852, + "learning_rate": 0.0002903710211092582, + "loss": 3.261525344848633, + "step": 1655 + }, + { + "epoch": 0.31892411143131605, + "grad_norm": 1.2215559482574463, + "learning_rate": 0.0002902868527689816, + "loss": 3.264019012451172, + "step": 1660 + }, + { + "epoch": 0.31988472622478387, + "grad_norm": 0.8910490274429321, + "learning_rate": 0.0002902023304705837, + "loss": 3.262919616699219, + "step": 1665 + }, + { + "epoch": 0.3208453410182517, + "grad_norm": 1.35845148563385, + "learning_rate": 0.00029011745442732303, + "loss": 3.2655372619628906, + "step": 1670 + }, + { + "epoch": 0.3218059558117195, + "grad_norm": 1.354686975479126, + "learning_rate": 0.0002900322248533506, + "loss": 3.26285400390625, + "step": 1675 + }, + { + "epoch": 0.3227665706051873, + "grad_norm": 1.0969462394714355, + "learning_rate": 0.0002899466419637093, + "loss": 3.257940673828125, + "step": 1680 + }, + { + "epoch": 0.32372718539865514, + "grad_norm": 1.291715145111084, + "learning_rate": 0.00028986070597433354, + "loss": 3.2633174896240233, + "step": 1685 + }, + { + "epoch": 0.32468780019212296, + "grad_norm": 1.4123157262802124, + "learning_rate": 0.00028977441710204867, + "loss": 3.2593170166015626, + "step": 1690 + }, + { + "epoch": 0.3256484149855908, + "grad_norm": 1.1881605386734009, + "learning_rate": 0.0002896877755645703, + "loss": 3.2590873718261717, + "step": 1695 + }, + { + "epoch": 0.3266090297790586, + "grad_norm": 1.043522596359253, + "learning_rate": 0.00028960078158050403, + "loss": 3.258655548095703, + "step": 1700 + }, + { + "epoch": 0.3275696445725264, + "grad_norm": 1.0990368127822876, + "learning_rate": 0.0002895134353693445, + "loss": 3.258600616455078, + "step": 1705 + }, + { + "epoch": 0.3285302593659942, + "grad_norm": 1.4628812074661255, + "learning_rate": 0.0002894257371514752, + "loss": 3.262346649169922, + "step": 1710 + }, + { + "epoch": 0.32949087415946204, + "grad_norm": 0.8175418376922607, + "learning_rate": 0.0002893376871481678, + "loss": 3.255454254150391, + "step": 1715 + }, + { + "epoch": 0.33045148895292986, + "grad_norm": 1.0393476486206055, + "learning_rate": 0.0002892492855815814, + "loss": 3.2613197326660157, + "step": 1720 + }, + { + "epoch": 0.3314121037463977, + "grad_norm": 1.250258445739746, + "learning_rate": 0.0002891605326747622, + "loss": 3.260877227783203, + "step": 1725 + }, + { + "epoch": 0.3323727185398655, + "grad_norm": 0.8695818781852722, + "learning_rate": 0.00028907142865164305, + "loss": 3.259082794189453, + "step": 1730 + }, + { + "epoch": 0.3333333333333333, + "grad_norm": 1.0892332792282104, + "learning_rate": 0.00028898197373704234, + "loss": 3.261730194091797, + "step": 1735 + }, + { + "epoch": 0.33429394812680113, + "grad_norm": 1.2874714136123657, + "learning_rate": 0.00028889216815666406, + "loss": 3.2563186645507813, + "step": 1740 + }, + { + "epoch": 0.33525456292026895, + "grad_norm": 1.2988804578781128, + "learning_rate": 0.000288802012137097, + "loss": 3.2569385528564454, + "step": 1745 + }, + { + "epoch": 0.33621517771373677, + "grad_norm": 1.2552379369735718, + "learning_rate": 0.0002887115059058139, + "loss": 3.2592254638671876, + "step": 1750 + }, + { + "epoch": 0.3371757925072046, + "grad_norm": 0.9785555601119995, + "learning_rate": 0.0002886206496911714, + "loss": 3.257631301879883, + "step": 1755 + }, + { + "epoch": 0.3381364073006724, + "grad_norm": 1.3783308267593384, + "learning_rate": 0.00028852944372240896, + "loss": 3.263458251953125, + "step": 1760 + }, + { + "epoch": 0.3390970220941403, + "grad_norm": 1.2814546823501587, + "learning_rate": 0.00028843788822964864, + "loss": 3.2578887939453125, + "step": 1765 + }, + { + "epoch": 0.3400576368876081, + "grad_norm": 1.2692376375198364, + "learning_rate": 0.00028834598344389437, + "loss": 3.2599555969238283, + "step": 1770 + }, + { + "epoch": 0.3410182516810759, + "grad_norm": 1.3274823427200317, + "learning_rate": 0.0002882537295970312, + "loss": 3.261073684692383, + "step": 1775 + }, + { + "epoch": 0.34197886647454373, + "grad_norm": 1.1028993129730225, + "learning_rate": 0.00028816112692182526, + "loss": 3.2600685119628907, + "step": 1780 + }, + { + "epoch": 0.34293948126801155, + "grad_norm": 1.0940430164337158, + "learning_rate": 0.0002880681756519224, + "loss": 3.2631156921386717, + "step": 1785 + }, + { + "epoch": 0.34390009606147937, + "grad_norm": 1.1741334199905396, + "learning_rate": 0.0002879748760218484, + "loss": 3.2562606811523436, + "step": 1790 + }, + { + "epoch": 0.3448607108549472, + "grad_norm": 1.4189056158065796, + "learning_rate": 0.0002878812282670077, + "loss": 3.2629486083984376, + "step": 1795 + }, + { + "epoch": 0.345821325648415, + "grad_norm": 0.9202475547790527, + "learning_rate": 0.00028778723262368313, + "loss": 3.258461761474609, + "step": 1800 + }, + { + "epoch": 0.3467819404418828, + "grad_norm": 1.0866518020629883, + "learning_rate": 0.00028769288932903547, + "loss": 3.2561798095703125, + "step": 1805 + }, + { + "epoch": 0.34774255523535064, + "grad_norm": 1.1435341835021973, + "learning_rate": 0.00028759819862110244, + "loss": 3.253205108642578, + "step": 1810 + }, + { + "epoch": 0.34870317002881845, + "grad_norm": 0.9997341632843018, + "learning_rate": 0.00028750316073879845, + "loss": 3.256957244873047, + "step": 1815 + }, + { + "epoch": 0.34966378482228627, + "grad_norm": 1.5304245948791504, + "learning_rate": 0.0002874077759219138, + "loss": 3.253832244873047, + "step": 1820 + }, + { + "epoch": 0.3506243996157541, + "grad_norm": 1.3517911434173584, + "learning_rate": 0.0002873120444111142, + "loss": 3.2599609375, + "step": 1825 + }, + { + "epoch": 0.3515850144092219, + "grad_norm": 1.1517263650894165, + "learning_rate": 0.0002872159664479401, + "loss": 3.253156280517578, + "step": 1830 + }, + { + "epoch": 0.3525456292026897, + "grad_norm": 1.2207057476043701, + "learning_rate": 0.00028711954227480596, + "loss": 3.255885696411133, + "step": 1835 + }, + { + "epoch": 0.35350624399615754, + "grad_norm": 1.4356719255447388, + "learning_rate": 0.00028702277213499993, + "loss": 3.2568824768066404, + "step": 1840 + }, + { + "epoch": 0.35446685878962536, + "grad_norm": 0.9663187265396118, + "learning_rate": 0.000286925656272683, + "loss": 3.2607173919677734, + "step": 1845 + }, + { + "epoch": 0.3554274735830932, + "grad_norm": 1.0548568964004517, + "learning_rate": 0.0002868281949328884, + "loss": 3.254198455810547, + "step": 1850 + }, + { + "epoch": 0.356388088376561, + "grad_norm": 1.216261386871338, + "learning_rate": 0.0002867303883615212, + "loss": 3.25797119140625, + "step": 1855 + }, + { + "epoch": 0.3573487031700288, + "grad_norm": 0.9959939122200012, + "learning_rate": 0.00028663223680535727, + "loss": 3.2549095153808594, + "step": 1860 + }, + { + "epoch": 0.35830931796349663, + "grad_norm": 1.0812616348266602, + "learning_rate": 0.0002865337405120432, + "loss": 3.2560733795166015, + "step": 1865 + }, + { + "epoch": 0.35926993275696445, + "grad_norm": 1.2685303688049316, + "learning_rate": 0.0002864348997300951, + "loss": 3.2555168151855467, + "step": 1870 + }, + { + "epoch": 0.36023054755043227, + "grad_norm": 1.0879360437393188, + "learning_rate": 0.0002863357147088985, + "loss": 3.255245590209961, + "step": 1875 + }, + { + "epoch": 0.3611911623439001, + "grad_norm": 1.0502004623413086, + "learning_rate": 0.0002862361856987073, + "loss": 3.260142517089844, + "step": 1880 + }, + { + "epoch": 0.3621517771373679, + "grad_norm": 1.2763992547988892, + "learning_rate": 0.0002861363129506435, + "loss": 3.2576175689697267, + "step": 1885 + }, + { + "epoch": 0.3631123919308357, + "grad_norm": 1.2673193216323853, + "learning_rate": 0.0002860360967166963, + "loss": 3.259668731689453, + "step": 1890 + }, + { + "epoch": 0.36407300672430354, + "grad_norm": 1.0005592107772827, + "learning_rate": 0.00028593553724972146, + "loss": 3.2603923797607424, + "step": 1895 + }, + { + "epoch": 0.36503362151777136, + "grad_norm": 1.2387863397598267, + "learning_rate": 0.0002858346348034408, + "loss": 3.2571121215820313, + "step": 1900 + }, + { + "epoch": 0.3659942363112392, + "grad_norm": 1.0775574445724487, + "learning_rate": 0.0002857333896324417, + "loss": 3.2519359588623047, + "step": 1905 + }, + { + "epoch": 0.366954851104707, + "grad_norm": 1.1767851114273071, + "learning_rate": 0.00028563180199217596, + "loss": 3.256037139892578, + "step": 1910 + }, + { + "epoch": 0.3679154658981748, + "grad_norm": 1.2488412857055664, + "learning_rate": 0.00028552987213895965, + "loss": 3.2543380737304686, + "step": 1915 + }, + { + "epoch": 0.3688760806916426, + "grad_norm": 1.2731399536132812, + "learning_rate": 0.0002854276003299723, + "loss": 3.251921844482422, + "step": 1920 + }, + { + "epoch": 0.36983669548511044, + "grad_norm": 1.2652751207351685, + "learning_rate": 0.00028532498682325613, + "loss": 3.254206085205078, + "step": 1925 + }, + { + "epoch": 0.37079731027857826, + "grad_norm": 1.3380461931228638, + "learning_rate": 0.0002852220318777156, + "loss": 3.2542865753173826, + "step": 1930 + }, + { + "epoch": 0.37175792507204614, + "grad_norm": 1.0234968662261963, + "learning_rate": 0.0002851187357531165, + "loss": 3.2551605224609377, + "step": 1935 + }, + { + "epoch": 0.37271853986551395, + "grad_norm": 0.7750507593154907, + "learning_rate": 0.00028501509871008563, + "loss": 3.258992004394531, + "step": 1940 + }, + { + "epoch": 0.37367915465898177, + "grad_norm": 0.9654103517532349, + "learning_rate": 0.00028491112101010986, + "loss": 3.2557037353515623, + "step": 1945 + }, + { + "epoch": 0.3746397694524496, + "grad_norm": 1.1453462839126587, + "learning_rate": 0.0002848068029155356, + "loss": 3.249644470214844, + "step": 1950 + }, + { + "epoch": 0.3756003842459174, + "grad_norm": 1.0247918367385864, + "learning_rate": 0.0002847021446895681, + "loss": 3.253766632080078, + "step": 1955 + }, + { + "epoch": 0.3765609990393852, + "grad_norm": 1.4280420541763306, + "learning_rate": 0.0002845971465962708, + "loss": 3.2554515838623046, + "step": 1960 + }, + { + "epoch": 0.37752161383285304, + "grad_norm": 1.534204125404358, + "learning_rate": 0.00028449180890056464, + "loss": 3.2532962799072265, + "step": 1965 + }, + { + "epoch": 0.37848222862632086, + "grad_norm": 1.2856396436691284, + "learning_rate": 0.00028438613186822746, + "loss": 3.251531219482422, + "step": 1970 + }, + { + "epoch": 0.3794428434197887, + "grad_norm": 1.0911295413970947, + "learning_rate": 0.00028428011576589327, + "loss": 3.250580978393555, + "step": 1975 + }, + { + "epoch": 0.3804034582132565, + "grad_norm": 1.142147421836853, + "learning_rate": 0.00028417376086105155, + "loss": 3.254068374633789, + "step": 1980 + }, + { + "epoch": 0.3813640730067243, + "grad_norm": 1.335033655166626, + "learning_rate": 0.0002840670674220466, + "loss": 3.2494667053222654, + "step": 1985 + }, + { + "epoch": 0.38232468780019213, + "grad_norm": 1.410131812095642, + "learning_rate": 0.00028396003571807697, + "loss": 3.2515232086181642, + "step": 1990 + }, + { + "epoch": 0.38328530259365995, + "grad_norm": 1.3424606323242188, + "learning_rate": 0.0002838526660191946, + "loss": 3.250862884521484, + "step": 1995 + }, + { + "epoch": 0.38424591738712777, + "grad_norm": 0.8202113509178162, + "learning_rate": 0.0002837449585963043, + "loss": 3.249342346191406, + "step": 2000 + }, + { + "epoch": 0.3852065321805956, + "grad_norm": 0.9713074564933777, + "learning_rate": 0.00028363691372116284, + "loss": 3.2535247802734375, + "step": 2005 + }, + { + "epoch": 0.3861671469740634, + "grad_norm": 0.7178729176521301, + "learning_rate": 0.0002835285316663786, + "loss": 3.2526882171630858, + "step": 2010 + }, + { + "epoch": 0.3871277617675312, + "grad_norm": 0.9730674624443054, + "learning_rate": 0.0002834198127054106, + "loss": 3.2526809692382814, + "step": 2015 + }, + { + "epoch": 0.38808837656099904, + "grad_norm": 1.3725463151931763, + "learning_rate": 0.000283310757112568, + "loss": 3.2512813568115235, + "step": 2020 + }, + { + "epoch": 0.38904899135446686, + "grad_norm": 1.5489643812179565, + "learning_rate": 0.0002832013651630091, + "loss": 3.2502918243408203, + "step": 2025 + }, + { + "epoch": 0.3900096061479347, + "grad_norm": 0.8844899535179138, + "learning_rate": 0.0002830916371327412, + "loss": 3.250738525390625, + "step": 2030 + }, + { + "epoch": 0.3909702209414025, + "grad_norm": 1.0500115156173706, + "learning_rate": 0.0002829815732986192, + "loss": 3.2540382385253905, + "step": 2035 + }, + { + "epoch": 0.3919308357348703, + "grad_norm": 1.0166023969650269, + "learning_rate": 0.00028287117393834563, + "loss": 3.2548370361328125, + "step": 2040 + }, + { + "epoch": 0.3928914505283381, + "grad_norm": 1.216377854347229, + "learning_rate": 0.00028276043933046926, + "loss": 3.248933792114258, + "step": 2045 + }, + { + "epoch": 0.39385206532180594, + "grad_norm": 0.9477090239524841, + "learning_rate": 0.0002826493697543849, + "loss": 3.2496387481689455, + "step": 2050 + }, + { + "epoch": 0.39481268011527376, + "grad_norm": 1.4546550512313843, + "learning_rate": 0.00028253796549033245, + "loss": 3.257366943359375, + "step": 2055 + }, + { + "epoch": 0.3957732949087416, + "grad_norm": 1.3477628231048584, + "learning_rate": 0.0002824262268193964, + "loss": 3.2546245574951174, + "step": 2060 + }, + { + "epoch": 0.3967339097022094, + "grad_norm": 1.1829005479812622, + "learning_rate": 0.00028231415402350476, + "loss": 3.2527755737304687, + "step": 2065 + }, + { + "epoch": 0.3976945244956772, + "grad_norm": 1.00674569606781, + "learning_rate": 0.00028220174738542876, + "loss": 3.249166488647461, + "step": 2070 + }, + { + "epoch": 0.39865513928914503, + "grad_norm": 1.143075942993164, + "learning_rate": 0.00028208900718878187, + "loss": 3.2508075714111326, + "step": 2075 + }, + { + "epoch": 0.39961575408261285, + "grad_norm": 0.9893442392349243, + "learning_rate": 0.00028197593371801915, + "loss": 3.256524658203125, + "step": 2080 + }, + { + "epoch": 0.40057636887608067, + "grad_norm": 0.9596649408340454, + "learning_rate": 0.00028186252725843664, + "loss": 3.2511486053466796, + "step": 2085 + }, + { + "epoch": 0.4015369836695485, + "grad_norm": 1.4209188222885132, + "learning_rate": 0.00028174878809617037, + "loss": 3.2548614501953126, + "step": 2090 + }, + { + "epoch": 0.4024975984630163, + "grad_norm": 1.0415889024734497, + "learning_rate": 0.000281634716518196, + "loss": 3.2491207122802734, + "step": 2095 + }, + { + "epoch": 0.4034582132564842, + "grad_norm": 1.0843944549560547, + "learning_rate": 0.00028152031281232775, + "loss": 3.2516613006591797, + "step": 2100 + }, + { + "epoch": 0.404418828049952, + "grad_norm": 1.5569794178009033, + "learning_rate": 0.00028140557726721795, + "loss": 3.2495628356933595, + "step": 2105 + }, + { + "epoch": 0.4053794428434198, + "grad_norm": 0.8837640881538391, + "learning_rate": 0.00028129051017235614, + "loss": 3.249878692626953, + "step": 2110 + }, + { + "epoch": 0.40634005763688763, + "grad_norm": 1.1749989986419678, + "learning_rate": 0.0002811751118180684, + "loss": 3.2475624084472656, + "step": 2115 + }, + { + "epoch": 0.40730067243035545, + "grad_norm": 1.2945916652679443, + "learning_rate": 0.00028105938249551666, + "loss": 3.2487873077392577, + "step": 2120 + }, + { + "epoch": 0.40826128722382327, + "grad_norm": 0.9443672299385071, + "learning_rate": 0.0002809433224966978, + "loss": 3.247914123535156, + "step": 2125 + }, + { + "epoch": 0.4092219020172911, + "grad_norm": 1.1064391136169434, + "learning_rate": 0.0002808269321144431, + "loss": 3.250864028930664, + "step": 2130 + }, + { + "epoch": 0.4101825168107589, + "grad_norm": 1.2150980234146118, + "learning_rate": 0.00028071021164241755, + "loss": 3.248881530761719, + "step": 2135 + }, + { + "epoch": 0.4111431316042267, + "grad_norm": 0.8171827793121338, + "learning_rate": 0.00028059316137511877, + "loss": 3.246689224243164, + "step": 2140 + }, + { + "epoch": 0.41210374639769454, + "grad_norm": 1.093009114265442, + "learning_rate": 0.0002804757816078766, + "loss": 3.2493213653564452, + "step": 2145 + }, + { + "epoch": 0.41306436119116235, + "grad_norm": 1.021937370300293, + "learning_rate": 0.00028035807263685224, + "loss": 3.2454933166503905, + "step": 2150 + }, + { + "epoch": 0.4140249759846302, + "grad_norm": 1.1384049654006958, + "learning_rate": 0.0002802400347590376, + "loss": 3.2516769409179687, + "step": 2155 + }, + { + "epoch": 0.414985590778098, + "grad_norm": 1.0551131963729858, + "learning_rate": 0.00028012166827225425, + "loss": 3.2476879119873048, + "step": 2160 + }, + { + "epoch": 0.4159462055715658, + "grad_norm": 0.9579499363899231, + "learning_rate": 0.0002800029734751529, + "loss": 3.247665023803711, + "step": 2165 + }, + { + "epoch": 0.4169068203650336, + "grad_norm": 1.290895938873291, + "learning_rate": 0.0002798839506672129, + "loss": 3.248125457763672, + "step": 2170 + }, + { + "epoch": 0.41786743515850144, + "grad_norm": 1.1702419519424438, + "learning_rate": 0.00027976460014874087, + "loss": 3.2500679016113283, + "step": 2175 + }, + { + "epoch": 0.41882804995196926, + "grad_norm": 1.2696362733840942, + "learning_rate": 0.0002796449222208704, + "loss": 3.2472507476806642, + "step": 2180 + }, + { + "epoch": 0.4197886647454371, + "grad_norm": 1.1748182773590088, + "learning_rate": 0.0002795249171855613, + "loss": 3.245960998535156, + "step": 2185 + }, + { + "epoch": 0.4207492795389049, + "grad_norm": 1.2495172023773193, + "learning_rate": 0.00027940458534559846, + "loss": 3.2419815063476562, + "step": 2190 + }, + { + "epoch": 0.4217098943323727, + "grad_norm": 1.2086557149887085, + "learning_rate": 0.0002792839270045916, + "loss": 3.2496551513671874, + "step": 2195 + }, + { + "epoch": 0.42267050912584053, + "grad_norm": 0.9928983449935913, + "learning_rate": 0.0002791629424669739, + "loss": 3.2451755523681642, + "step": 2200 + }, + { + "epoch": 0.42363112391930835, + "grad_norm": 1.264974594116211, + "learning_rate": 0.00027904163203800185, + "loss": 3.2482452392578125, + "step": 2205 + }, + { + "epoch": 0.42459173871277617, + "grad_norm": 1.2751965522766113, + "learning_rate": 0.0002789199960237542, + "loss": 3.251111602783203, + "step": 2210 + }, + { + "epoch": 0.425552353506244, + "grad_norm": 1.115212082862854, + "learning_rate": 0.0002787980347311309, + "loss": 3.2442550659179688, + "step": 2215 + }, + { + "epoch": 0.4265129682997118, + "grad_norm": 1.3036905527114868, + "learning_rate": 0.00027867574846785295, + "loss": 3.250843048095703, + "step": 2220 + }, + { + "epoch": 0.4274735830931796, + "grad_norm": 1.3169866800308228, + "learning_rate": 0.0002785531375424611, + "loss": 3.2423397064208985, + "step": 2225 + }, + { + "epoch": 0.42843419788664744, + "grad_norm": 1.1663439273834229, + "learning_rate": 0.00027843020226431535, + "loss": 3.2481651306152344, + "step": 2230 + }, + { + "epoch": 0.42939481268011526, + "grad_norm": 1.014664649963379, + "learning_rate": 0.0002783069429435939, + "loss": 3.245722198486328, + "step": 2235 + }, + { + "epoch": 0.4303554274735831, + "grad_norm": 1.1078412532806396, + "learning_rate": 0.00027818335989129275, + "loss": 3.2450366973876954, + "step": 2240 + }, + { + "epoch": 0.4313160422670509, + "grad_norm": 1.0631850957870483, + "learning_rate": 0.00027805945341922457, + "loss": 3.2482345581054686, + "step": 2245 + }, + { + "epoch": 0.4322766570605187, + "grad_norm": 1.182057499885559, + "learning_rate": 0.0002779352238400181, + "loss": 3.2490921020507812, + "step": 2250 + }, + { + "epoch": 0.4332372718539865, + "grad_norm": 1.021649718284607, + "learning_rate": 0.0002778106714671174, + "loss": 3.2460681915283205, + "step": 2255 + }, + { + "epoch": 0.43419788664745435, + "grad_norm": 1.0771024227142334, + "learning_rate": 0.00027768579661478077, + "loss": 3.2438682556152343, + "step": 2260 + }, + { + "epoch": 0.43515850144092216, + "grad_norm": 1.1866998672485352, + "learning_rate": 0.0002775605995980803, + "loss": 3.2406051635742186, + "step": 2265 + }, + { + "epoch": 0.43611911623439004, + "grad_norm": 1.2904285192489624, + "learning_rate": 0.00027743508073290097, + "loss": 3.247574234008789, + "step": 2270 + }, + { + "epoch": 0.43707973102785785, + "grad_norm": 0.990243136882782, + "learning_rate": 0.0002773092403359397, + "loss": 3.2439697265625, + "step": 2275 + }, + { + "epoch": 0.43804034582132567, + "grad_norm": 0.976154625415802, + "learning_rate": 0.0002771830787247047, + "loss": 3.251276397705078, + "step": 2280 + }, + { + "epoch": 0.4390009606147935, + "grad_norm": 1.2034424543380737, + "learning_rate": 0.0002770565962175148, + "loss": 3.2444232940673827, + "step": 2285 + }, + { + "epoch": 0.4399615754082613, + "grad_norm": 1.4228894710540771, + "learning_rate": 0.00027692979313349815, + "loss": 3.2472274780273436, + "step": 2290 + }, + { + "epoch": 0.4409221902017291, + "grad_norm": 1.3287744522094727, + "learning_rate": 0.000276802669792592, + "loss": 3.247750091552734, + "step": 2295 + }, + { + "epoch": 0.44188280499519694, + "grad_norm": 1.1008442640304565, + "learning_rate": 0.0002766752265155416, + "loss": 3.2413604736328123, + "step": 2300 + }, + { + "epoch": 0.44284341978866476, + "grad_norm": 1.3397494554519653, + "learning_rate": 0.0002765474636238994, + "loss": 3.243372344970703, + "step": 2305 + }, + { + "epoch": 0.4438040345821326, + "grad_norm": 1.2448689937591553, + "learning_rate": 0.00027641938144002434, + "loss": 3.24469108581543, + "step": 2310 + }, + { + "epoch": 0.4447646493756004, + "grad_norm": 1.2328872680664062, + "learning_rate": 0.0002762909802870809, + "loss": 3.2434837341308596, + "step": 2315 + }, + { + "epoch": 0.4457252641690682, + "grad_norm": 1.3019057512283325, + "learning_rate": 0.0002761622604890382, + "loss": 3.2452354431152344, + "step": 2320 + }, + { + "epoch": 0.44668587896253603, + "grad_norm": 1.0617128610610962, + "learning_rate": 0.0002760332223706697, + "loss": 3.243347930908203, + "step": 2325 + }, + { + "epoch": 0.44764649375600385, + "grad_norm": 1.2202798128128052, + "learning_rate": 0.0002759038662575518, + "loss": 3.2428775787353517, + "step": 2330 + }, + { + "epoch": 0.44860710854947167, + "grad_norm": 0.8098715543746948, + "learning_rate": 0.0002757741924760631, + "loss": 3.2433307647705076, + "step": 2335 + }, + { + "epoch": 0.4495677233429395, + "grad_norm": 1.2258954048156738, + "learning_rate": 0.0002756442013533839, + "loss": 3.2418006896972655, + "step": 2340 + }, + { + "epoch": 0.4505283381364073, + "grad_norm": 1.155310869216919, + "learning_rate": 0.0002755138932174952, + "loss": 3.2446517944335938, + "step": 2345 + }, + { + "epoch": 0.4514889529298751, + "grad_norm": 0.9408679604530334, + "learning_rate": 0.0002753832683971778, + "loss": 3.2486572265625, + "step": 2350 + }, + { + "epoch": 0.45244956772334294, + "grad_norm": 1.1427699327468872, + "learning_rate": 0.0002752523272220114, + "loss": 3.2479103088378904, + "step": 2355 + }, + { + "epoch": 0.45341018251681076, + "grad_norm": 0.9570598602294922, + "learning_rate": 0.00027512107002237415, + "loss": 3.247629165649414, + "step": 2360 + }, + { + "epoch": 0.4543707973102786, + "grad_norm": 0.8978891968727112, + "learning_rate": 0.0002749894971294414, + "loss": 3.242329406738281, + "step": 2365 + }, + { + "epoch": 0.4553314121037464, + "grad_norm": 1.0361249446868896, + "learning_rate": 0.000274857608875185, + "loss": 3.243325042724609, + "step": 2370 + }, + { + "epoch": 0.4562920268972142, + "grad_norm": 0.9221097230911255, + "learning_rate": 0.0002747254055923726, + "loss": 3.2444053649902345, + "step": 2375 + }, + { + "epoch": 0.457252641690682, + "grad_norm": 1.2986423969268799, + "learning_rate": 0.0002745928876145667, + "loss": 3.2442516326904296, + "step": 2380 + }, + { + "epoch": 0.45821325648414984, + "grad_norm": 0.9883876442909241, + "learning_rate": 0.0002744600552761236, + "loss": 3.2429885864257812, + "step": 2385 + }, + { + "epoch": 0.45917387127761766, + "grad_norm": 1.4296432733535767, + "learning_rate": 0.000274326908912193, + "loss": 3.2417251586914064, + "step": 2390 + }, + { + "epoch": 0.4601344860710855, + "grad_norm": 1.216537594795227, + "learning_rate": 0.00027419344885871685, + "loss": 3.2435916900634765, + "step": 2395 + }, + { + "epoch": 0.4610951008645533, + "grad_norm": 1.0442991256713867, + "learning_rate": 0.0002740596754524285, + "loss": 3.2400821685791015, + "step": 2400 + }, + { + "epoch": 0.4620557156580211, + "grad_norm": 1.4031963348388672, + "learning_rate": 0.0002739255890308521, + "loss": 3.2466384887695314, + "step": 2405 + }, + { + "epoch": 0.46301633045148893, + "grad_norm": 1.301132082939148, + "learning_rate": 0.0002737911899323013, + "loss": 3.241625213623047, + "step": 2410 + }, + { + "epoch": 0.46397694524495675, + "grad_norm": 1.3836039304733276, + "learning_rate": 0.00027365647849587893, + "loss": 3.2370471954345703, + "step": 2415 + }, + { + "epoch": 0.46493756003842457, + "grad_norm": 0.9351392984390259, + "learning_rate": 0.0002735214550614757, + "loss": 3.2448040008544923, + "step": 2420 + }, + { + "epoch": 0.4658981748318924, + "grad_norm": 0.9278304576873779, + "learning_rate": 0.00027338611996976955, + "loss": 3.239888000488281, + "step": 2425 + }, + { + "epoch": 0.4668587896253602, + "grad_norm": 0.8525119423866272, + "learning_rate": 0.0002732504735622248, + "loss": 3.2448410034179687, + "step": 2430 + }, + { + "epoch": 0.4678194044188281, + "grad_norm": 1.2899324893951416, + "learning_rate": 0.00027311451618109144, + "loss": 3.2423675537109373, + "step": 2435 + }, + { + "epoch": 0.4687800192122959, + "grad_norm": 0.8825019598007202, + "learning_rate": 0.00027297824816940365, + "loss": 3.240159606933594, + "step": 2440 + }, + { + "epoch": 0.4697406340057637, + "grad_norm": 1.301020860671997, + "learning_rate": 0.00027284166987097977, + "loss": 3.24395751953125, + "step": 2445 + }, + { + "epoch": 0.47070124879923153, + "grad_norm": 0.9350459575653076, + "learning_rate": 0.0002727047816304208, + "loss": 3.2402999877929686, + "step": 2450 + }, + { + "epoch": 0.47166186359269935, + "grad_norm": 1.1744569540023804, + "learning_rate": 0.00027256758379310975, + "loss": 3.239767837524414, + "step": 2455 + }, + { + "epoch": 0.47262247838616717, + "grad_norm": 0.9879583716392517, + "learning_rate": 0.00027243007670521086, + "loss": 3.242339324951172, + "step": 2460 + }, + { + "epoch": 0.473583093179635, + "grad_norm": 0.8559085726737976, + "learning_rate": 0.00027229226071366874, + "loss": 3.241751861572266, + "step": 2465 + }, + { + "epoch": 0.4745437079731028, + "grad_norm": 1.3251603841781616, + "learning_rate": 0.00027215413616620714, + "loss": 3.2432788848876952, + "step": 2470 + }, + { + "epoch": 0.4755043227665706, + "grad_norm": 1.1395015716552734, + "learning_rate": 0.00027201570341132855, + "loss": 3.239895248413086, + "step": 2475 + }, + { + "epoch": 0.47646493756003844, + "grad_norm": 1.1372860670089722, + "learning_rate": 0.0002718769627983129, + "loss": 3.2404502868652343, + "step": 2480 + }, + { + "epoch": 0.47742555235350626, + "grad_norm": 1.014133334159851, + "learning_rate": 0.00027173791467721714, + "loss": 3.238873291015625, + "step": 2485 + }, + { + "epoch": 0.4783861671469741, + "grad_norm": 1.2768558263778687, + "learning_rate": 0.00027159855939887386, + "loss": 3.2407760620117188, + "step": 2490 + }, + { + "epoch": 0.4793467819404419, + "grad_norm": 1.0559134483337402, + "learning_rate": 0.0002714588973148907, + "loss": 3.2391632080078123, + "step": 2495 + }, + { + "epoch": 0.4803073967339097, + "grad_norm": 1.212760090827942, + "learning_rate": 0.00027131892877764945, + "loss": 3.242381286621094, + "step": 2500 + }, + { + "epoch": 0.4812680115273775, + "grad_norm": 1.3705852031707764, + "learning_rate": 0.0002711786541403051, + "loss": 3.2453689575195312, + "step": 2505 + }, + { + "epoch": 0.48222862632084534, + "grad_norm": 1.1016364097595215, + "learning_rate": 0.00027103807375678494, + "loss": 3.240717315673828, + "step": 2510 + }, + { + "epoch": 0.48318924111431316, + "grad_norm": 1.0440564155578613, + "learning_rate": 0.0002708971879817877, + "loss": 3.2409019470214844, + "step": 2515 + }, + { + "epoch": 0.484149855907781, + "grad_norm": 1.175846815109253, + "learning_rate": 0.0002707559971707827, + "loss": 3.2366703033447264, + "step": 2520 + }, + { + "epoch": 0.4851104707012488, + "grad_norm": 1.2363582849502563, + "learning_rate": 0.00027061450168000875, + "loss": 3.235219192504883, + "step": 2525 + }, + { + "epoch": 0.4860710854947166, + "grad_norm": 1.2251267433166504, + "learning_rate": 0.00027047270186647353, + "loss": 3.2362335205078123, + "step": 2530 + }, + { + "epoch": 0.48703170028818443, + "grad_norm": 0.8210157752037048, + "learning_rate": 0.00027033059808795253, + "loss": 3.2397132873535157, + "step": 2535 + }, + { + "epoch": 0.48799231508165225, + "grad_norm": 1.229904055595398, + "learning_rate": 0.0002701881907029881, + "loss": 3.2385082244873047, + "step": 2540 + }, + { + "epoch": 0.48895292987512007, + "grad_norm": 1.1676335334777832, + "learning_rate": 0.00027004548007088876, + "loss": 3.2385066986083983, + "step": 2545 + }, + { + "epoch": 0.4899135446685879, + "grad_norm": 1.271478533744812, + "learning_rate": 0.000269902466551728, + "loss": 3.2415950775146483, + "step": 2550 + }, + { + "epoch": 0.4908741594620557, + "grad_norm": 1.395169973373413, + "learning_rate": 0.00026975915050634367, + "loss": 3.2399360656738283, + "step": 2555 + }, + { + "epoch": 0.4918347742555235, + "grad_norm": 1.3647792339324951, + "learning_rate": 0.00026961553229633685, + "loss": 3.243121337890625, + "step": 2560 + }, + { + "epoch": 0.49279538904899134, + "grad_norm": 1.0769857168197632, + "learning_rate": 0.000269471612284071, + "loss": 3.2393962860107424, + "step": 2565 + }, + { + "epoch": 0.49375600384245916, + "grad_norm": 0.9694387316703796, + "learning_rate": 0.00026932739083267105, + "loss": 3.2386848449707033, + "step": 2570 + }, + { + "epoch": 0.494716618635927, + "grad_norm": 0.9222965240478516, + "learning_rate": 0.00026918286830602254, + "loss": 3.238786315917969, + "step": 2575 + }, + { + "epoch": 0.4956772334293948, + "grad_norm": 1.1549302339553833, + "learning_rate": 0.00026903804506877064, + "loss": 3.239440155029297, + "step": 2580 + }, + { + "epoch": 0.4966378482228626, + "grad_norm": 1.0019875764846802, + "learning_rate": 0.0002688929214863192, + "loss": 3.237133026123047, + "step": 2585 + }, + { + "epoch": 0.49759846301633043, + "grad_norm": 1.1600396633148193, + "learning_rate": 0.0002687474979248299, + "loss": 3.2444324493408203, + "step": 2590 + }, + { + "epoch": 0.49855907780979825, + "grad_norm": 1.2407069206237793, + "learning_rate": 0.00026860177475122133, + "loss": 3.2405441284179686, + "step": 2595 + }, + { + "epoch": 0.49951969260326606, + "grad_norm": 1.313706874847412, + "learning_rate": 0.00026845575233316806, + "loss": 3.2400081634521483, + "step": 2600 + }, + { + "epoch": 0.5004803073967339, + "grad_norm": 1.102402925491333, + "learning_rate": 0.00026830943103909954, + "loss": 3.236441421508789, + "step": 2605 + }, + { + "epoch": 0.5014409221902018, + "grad_norm": 1.4062278270721436, + "learning_rate": 0.00026816281123819946, + "loss": 3.24122314453125, + "step": 2610 + }, + { + "epoch": 0.5024015369836695, + "grad_norm": 1.1859440803527832, + "learning_rate": 0.00026801589330040465, + "loss": 3.238700103759766, + "step": 2615 + }, + { + "epoch": 0.5033621517771374, + "grad_norm": 0.8828426599502563, + "learning_rate": 0.000267868677596404, + "loss": 3.2366287231445314, + "step": 2620 + }, + { + "epoch": 0.5043227665706052, + "grad_norm": 1.1908822059631348, + "learning_rate": 0.000267721164497638, + "loss": 3.2335018157958983, + "step": 2625 + }, + { + "epoch": 0.505283381364073, + "grad_norm": 1.3185487985610962, + "learning_rate": 0.00026757335437629725, + "loss": 3.2354408264160157, + "step": 2630 + }, + { + "epoch": 0.5062439961575408, + "grad_norm": 1.1835713386535645, + "learning_rate": 0.00026742524760532183, + "loss": 3.235929489135742, + "step": 2635 + }, + { + "epoch": 0.5072046109510087, + "grad_norm": 1.6619583368301392, + "learning_rate": 0.00026727684455840037, + "loss": 3.236317443847656, + "step": 2640 + }, + { + "epoch": 0.5081652257444764, + "grad_norm": 1.0703696012496948, + "learning_rate": 0.00026712814560996896, + "loss": 3.2386764526367187, + "step": 2645 + }, + { + "epoch": 0.5091258405379443, + "grad_norm": 1.149133324623108, + "learning_rate": 0.0002669791511352103, + "loss": 3.23681526184082, + "step": 2650 + }, + { + "epoch": 0.5100864553314121, + "grad_norm": 1.1290984153747559, + "learning_rate": 0.0002668298615100527, + "loss": 3.2398101806640627, + "step": 2655 + }, + { + "epoch": 0.5110470701248799, + "grad_norm": 1.0490314960479736, + "learning_rate": 0.0002666802771111693, + "loss": 3.238353729248047, + "step": 2660 + }, + { + "epoch": 0.5120076849183477, + "grad_norm": 0.849686861038208, + "learning_rate": 0.0002665303983159767, + "loss": 3.2377159118652346, + "step": 2665 + }, + { + "epoch": 0.5129682997118156, + "grad_norm": 1.13809072971344, + "learning_rate": 0.00026638022550263467, + "loss": 3.2373073577880858, + "step": 2670 + }, + { + "epoch": 0.5139289145052833, + "grad_norm": 1.1344239711761475, + "learning_rate": 0.0002662297590500445, + "loss": 3.233965301513672, + "step": 2675 + }, + { + "epoch": 0.5148895292987512, + "grad_norm": 1.3232028484344482, + "learning_rate": 0.0002660789993378486, + "loss": 3.236408233642578, + "step": 2680 + }, + { + "epoch": 0.515850144092219, + "grad_norm": 1.2224704027175903, + "learning_rate": 0.0002659279467464291, + "loss": 3.238209915161133, + "step": 2685 + }, + { + "epoch": 0.5168107588856868, + "grad_norm": 1.106103777885437, + "learning_rate": 0.0002657766016569072, + "loss": 3.2342296600341798, + "step": 2690 + }, + { + "epoch": 0.5177713736791547, + "grad_norm": 0.8532865643501282, + "learning_rate": 0.00026562496445114205, + "loss": 3.238364410400391, + "step": 2695 + }, + { + "epoch": 0.5187319884726225, + "grad_norm": 0.8444566130638123, + "learning_rate": 0.00026547303551172996, + "loss": 3.236472320556641, + "step": 2700 + }, + { + "epoch": 0.5196926032660903, + "grad_norm": 0.8221146464347839, + "learning_rate": 0.0002653208152220032, + "loss": 3.2367286682128906, + "step": 2705 + }, + { + "epoch": 0.5206532180595581, + "grad_norm": 1.1430490016937256, + "learning_rate": 0.00026516830396602914, + "loss": 3.2309627532958984, + "step": 2710 + }, + { + "epoch": 0.521613832853026, + "grad_norm": 1.3973969221115112, + "learning_rate": 0.00026501550212860937, + "loss": 3.2348331451416015, + "step": 2715 + }, + { + "epoch": 0.5225744476464937, + "grad_norm": 1.2749481201171875, + "learning_rate": 0.0002648624100952786, + "loss": 3.2363750457763674, + "step": 2720 + }, + { + "epoch": 0.5235350624399616, + "grad_norm": 1.2370373010635376, + "learning_rate": 0.00026470902825230367, + "loss": 3.2376075744628907, + "step": 2725 + }, + { + "epoch": 0.5244956772334294, + "grad_norm": 1.0846977233886719, + "learning_rate": 0.0002645553569866828, + "loss": 3.2337451934814454, + "step": 2730 + }, + { + "epoch": 0.5254562920268973, + "grad_norm": 1.026764154434204, + "learning_rate": 0.00026440139668614427, + "loss": 3.237920379638672, + "step": 2735 + }, + { + "epoch": 0.526416906820365, + "grad_norm": 1.045480489730835, + "learning_rate": 0.00026424714773914574, + "loss": 3.231882858276367, + "step": 2740 + }, + { + "epoch": 0.5273775216138329, + "grad_norm": 0.9559823274612427, + "learning_rate": 0.0002640926105348732, + "loss": 3.23189697265625, + "step": 2745 + }, + { + "epoch": 0.5283381364073007, + "grad_norm": 1.3470317125320435, + "learning_rate": 0.00026393778546323976, + "loss": 3.2387470245361327, + "step": 2750 + }, + { + "epoch": 0.5292987512007685, + "grad_norm": 1.1389282941818237, + "learning_rate": 0.00026378267291488506, + "loss": 3.2349002838134764, + "step": 2755 + }, + { + "epoch": 0.5302593659942363, + "grad_norm": 1.0445148944854736, + "learning_rate": 0.00026362727328117384, + "loss": 3.2324363708496096, + "step": 2760 + }, + { + "epoch": 0.5312199807877042, + "grad_norm": 1.0291074514389038, + "learning_rate": 0.00026347158695419546, + "loss": 3.2345924377441406, + "step": 2765 + }, + { + "epoch": 0.5321805955811719, + "grad_norm": 1.23056161403656, + "learning_rate": 0.00026331561432676244, + "loss": 3.2309837341308594, + "step": 2770 + }, + { + "epoch": 0.5331412103746398, + "grad_norm": 0.658456563949585, + "learning_rate": 0.0002631593557924097, + "loss": 3.232289123535156, + "step": 2775 + }, + { + "epoch": 0.5341018251681076, + "grad_norm": 1.1292603015899658, + "learning_rate": 0.0002630028117453936, + "loss": 3.2309852600097657, + "step": 2780 + }, + { + "epoch": 0.5350624399615754, + "grad_norm": 0.8553731441497803, + "learning_rate": 0.00026284598258069074, + "loss": 3.235504150390625, + "step": 2785 + }, + { + "epoch": 0.5360230547550432, + "grad_norm": 0.8991851210594177, + "learning_rate": 0.00026268886869399723, + "loss": 3.233042907714844, + "step": 2790 + }, + { + "epoch": 0.5369836695485111, + "grad_norm": 0.916896402835846, + "learning_rate": 0.0002625314704817276, + "loss": 3.2310935974121096, + "step": 2795 + }, + { + "epoch": 0.5379442843419788, + "grad_norm": 1.1668161153793335, + "learning_rate": 0.0002623737883410136, + "loss": 3.232282257080078, + "step": 2800 + }, + { + "epoch": 0.5389048991354467, + "grad_norm": 1.223141074180603, + "learning_rate": 0.00026221582266970346, + "loss": 3.2328109741210938, + "step": 2805 + }, + { + "epoch": 0.5398655139289145, + "grad_norm": 1.0517135858535767, + "learning_rate": 0.00026205757386636085, + "loss": 3.2323383331298827, + "step": 2810 + }, + { + "epoch": 0.5408261287223823, + "grad_norm": 0.7666857242584229, + "learning_rate": 0.0002618990423302636, + "loss": 3.234081268310547, + "step": 2815 + }, + { + "epoch": 0.5417867435158501, + "grad_norm": 1.253327488899231, + "learning_rate": 0.0002617402284614031, + "loss": 3.233143997192383, + "step": 2820 + }, + { + "epoch": 0.542747358309318, + "grad_norm": 0.7495427131652832, + "learning_rate": 0.00026158113266048317, + "loss": 3.236143112182617, + "step": 2825 + }, + { + "epoch": 0.5437079731027857, + "grad_norm": 1.1047379970550537, + "learning_rate": 0.0002614217553289186, + "loss": 3.234874725341797, + "step": 2830 + }, + { + "epoch": 0.5446685878962536, + "grad_norm": 1.2154449224472046, + "learning_rate": 0.0002612620968688349, + "loss": 3.2348342895507813, + "step": 2835 + }, + { + "epoch": 0.5456292026897214, + "grad_norm": 1.1259676218032837, + "learning_rate": 0.0002611021576830667, + "loss": 3.232221221923828, + "step": 2840 + }, + { + "epoch": 0.5465898174831892, + "grad_norm": 1.2192593812942505, + "learning_rate": 0.00026094193817515697, + "loss": 3.2312171936035154, + "step": 2845 + }, + { + "epoch": 0.547550432276657, + "grad_norm": 1.3673197031021118, + "learning_rate": 0.00026078143874935604, + "loss": 3.229291534423828, + "step": 2850 + }, + { + "epoch": 0.5485110470701249, + "grad_norm": 1.3008506298065186, + "learning_rate": 0.00026062065981062024, + "loss": 3.2330032348632813, + "step": 2855 + }, + { + "epoch": 0.5494716618635928, + "grad_norm": 1.1764352321624756, + "learning_rate": 0.00026045960176461155, + "loss": 3.232608413696289, + "step": 2860 + }, + { + "epoch": 0.5504322766570605, + "grad_norm": 1.0877325534820557, + "learning_rate": 0.0002602982650176958, + "loss": 3.232305908203125, + "step": 2865 + }, + { + "epoch": 0.5513928914505284, + "grad_norm": 1.1681047677993774, + "learning_rate": 0.0002601366499769422, + "loss": 3.233287811279297, + "step": 2870 + }, + { + "epoch": 0.5523535062439962, + "grad_norm": 1.0753793716430664, + "learning_rate": 0.00025997475705012203, + "loss": 3.2348102569580077, + "step": 2875 + }, + { + "epoch": 0.553314121037464, + "grad_norm": 0.9505957365036011, + "learning_rate": 0.0002598125866457078, + "loss": 3.2324668884277346, + "step": 2880 + }, + { + "epoch": 0.5542747358309318, + "grad_norm": 1.2333307266235352, + "learning_rate": 0.0002596501391728721, + "loss": 3.235234832763672, + "step": 2885 + }, + { + "epoch": 0.5552353506243997, + "grad_norm": 1.0360201597213745, + "learning_rate": 0.00025948741504148656, + "loss": 3.234451675415039, + "step": 2890 + }, + { + "epoch": 0.5561959654178674, + "grad_norm": 0.9275029301643372, + "learning_rate": 0.0002593244146621208, + "loss": 3.2318180084228514, + "step": 2895 + }, + { + "epoch": 0.5571565802113353, + "grad_norm": 0.8183401823043823, + "learning_rate": 0.00025916113844604157, + "loss": 3.2336944580078124, + "step": 2900 + }, + { + "epoch": 0.5581171950048031, + "grad_norm": 0.7255450487136841, + "learning_rate": 0.0002589975868052114, + "loss": 3.2311965942382814, + "step": 2905 + }, + { + "epoch": 0.5590778097982709, + "grad_norm": 1.1232422590255737, + "learning_rate": 0.000258833760152288, + "loss": 3.2349414825439453, + "step": 2910 + }, + { + "epoch": 0.5600384245917387, + "grad_norm": 1.311683177947998, + "learning_rate": 0.00025866965890062276, + "loss": 3.233409881591797, + "step": 2915 + }, + { + "epoch": 0.5609990393852066, + "grad_norm": 1.104836106300354, + "learning_rate": 0.0002585052834642599, + "loss": 3.228401947021484, + "step": 2920 + }, + { + "epoch": 0.5619596541786743, + "grad_norm": 1.170188069343567, + "learning_rate": 0.0002583406342579357, + "loss": 3.2292003631591797, + "step": 2925 + }, + { + "epoch": 0.5629202689721422, + "grad_norm": 1.1829124689102173, + "learning_rate": 0.0002581757116970768, + "loss": 3.2359298706054687, + "step": 2930 + }, + { + "epoch": 0.56388088376561, + "grad_norm": 0.885960042476654, + "learning_rate": 0.00025801051619779987, + "loss": 3.2320457458496095, + "step": 2935 + }, + { + "epoch": 0.5648414985590778, + "grad_norm": 1.0792263746261597, + "learning_rate": 0.0002578450481769099, + "loss": 3.2303844451904298, + "step": 2940 + }, + { + "epoch": 0.5658021133525456, + "grad_norm": 1.2442518472671509, + "learning_rate": 0.0002576793080519, + "loss": 3.2286468505859376, + "step": 2945 + }, + { + "epoch": 0.5667627281460135, + "grad_norm": 0.976241409778595, + "learning_rate": 0.00025751329624094926, + "loss": 3.226253128051758, + "step": 2950 + }, + { + "epoch": 0.5677233429394812, + "grad_norm": 1.2932735681533813, + "learning_rate": 0.00025734701316292263, + "loss": 3.2291339874267577, + "step": 2955 + }, + { + "epoch": 0.5686839577329491, + "grad_norm": 1.17381751537323, + "learning_rate": 0.0002571804592373693, + "loss": 3.2318046569824217, + "step": 2960 + }, + { + "epoch": 0.5696445725264169, + "grad_norm": 1.1502513885498047, + "learning_rate": 0.00025701363488452196, + "loss": 3.230632781982422, + "step": 2965 + }, + { + "epoch": 0.5706051873198847, + "grad_norm": 1.2376641035079956, + "learning_rate": 0.00025684654052529556, + "loss": 3.2285648345947267, + "step": 2970 + }, + { + "epoch": 0.5715658021133525, + "grad_norm": 1.2463771104812622, + "learning_rate": 0.0002566791765812862, + "loss": 3.232989501953125, + "step": 2975 + }, + { + "epoch": 0.5725264169068204, + "grad_norm": 1.148836612701416, + "learning_rate": 0.0002565115434747705, + "loss": 3.2307807922363283, + "step": 2980 + }, + { + "epoch": 0.5734870317002881, + "grad_norm": 1.313881516456604, + "learning_rate": 0.00025634364162870386, + "loss": 3.2307106018066407, + "step": 2985 + }, + { + "epoch": 0.574447646493756, + "grad_norm": 1.0571279525756836, + "learning_rate": 0.0002561754714667198, + "loss": 3.230280303955078, + "step": 2990 + }, + { + "epoch": 0.5754082612872238, + "grad_norm": 0.9680122137069702, + "learning_rate": 0.000256007033413129, + "loss": 3.2283321380615235, + "step": 2995 + }, + { + "epoch": 0.5763688760806917, + "grad_norm": 1.11472749710083, + "learning_rate": 0.00025583832789291794, + "loss": 3.229805755615234, + "step": 3000 + }, + { + "epoch": 0.5773294908741594, + "grad_norm": 1.0987354516983032, + "learning_rate": 0.0002556693553317479, + "loss": 3.2307418823242187, + "step": 3005 + }, + { + "epoch": 0.5782901056676273, + "grad_norm": 0.8845497369766235, + "learning_rate": 0.0002555001161559539, + "loss": 3.226715850830078, + "step": 3010 + }, + { + "epoch": 0.579250720461095, + "grad_norm": 1.37449049949646, + "learning_rate": 0.000255330610792544, + "loss": 3.230916976928711, + "step": 3015 + }, + { + "epoch": 0.5802113352545629, + "grad_norm": 1.2311089038848877, + "learning_rate": 0.00025516083966919744, + "loss": 3.228615570068359, + "step": 3020 + }, + { + "epoch": 0.5811719500480308, + "grad_norm": 1.0169094800949097, + "learning_rate": 0.00025499080321426417, + "loss": 3.2289390563964844, + "step": 3025 + }, + { + "epoch": 0.5821325648414986, + "grad_norm": 1.280392050743103, + "learning_rate": 0.00025482050185676367, + "loss": 3.227009963989258, + "step": 3030 + }, + { + "epoch": 0.5830931796349664, + "grad_norm": 0.9485029578208923, + "learning_rate": 0.0002546499360263837, + "loss": 3.231037139892578, + "step": 3035 + }, + { + "epoch": 0.5840537944284342, + "grad_norm": 1.0025532245635986, + "learning_rate": 0.0002544791061534794, + "loss": 3.231890869140625, + "step": 3040 + }, + { + "epoch": 0.5850144092219021, + "grad_norm": 0.9485952854156494, + "learning_rate": 0.000254308012669072, + "loss": 3.2329727172851563, + "step": 3045 + }, + { + "epoch": 0.5859750240153698, + "grad_norm": 0.9976146817207336, + "learning_rate": 0.00025413665600484796, + "loss": 3.228594207763672, + "step": 3050 + }, + { + "epoch": 0.5869356388088377, + "grad_norm": 0.9106408953666687, + "learning_rate": 0.0002539650365931577, + "loss": 3.225025177001953, + "step": 3055 + }, + { + "epoch": 0.5878962536023055, + "grad_norm": 0.823363721370697, + "learning_rate": 0.00025379315486701454, + "loss": 3.227878189086914, + "step": 3060 + }, + { + "epoch": 0.5888568683957733, + "grad_norm": 0.9073953032493591, + "learning_rate": 0.00025362101126009387, + "loss": 3.2288444519042967, + "step": 3065 + }, + { + "epoch": 0.5898174831892411, + "grad_norm": 1.0130420923233032, + "learning_rate": 0.0002534486062067316, + "loss": 3.228134536743164, + "step": 3070 + }, + { + "epoch": 0.590778097982709, + "grad_norm": 0.8417114615440369, + "learning_rate": 0.00025327594014192326, + "loss": 3.2314247131347655, + "step": 3075 + }, + { + "epoch": 0.5917387127761767, + "grad_norm": 0.8221418261528015, + "learning_rate": 0.0002531030135013232, + "loss": 3.226838302612305, + "step": 3080 + }, + { + "epoch": 0.5926993275696446, + "grad_norm": 1.0958610773086548, + "learning_rate": 0.0002529298267212429, + "loss": 3.2264358520507814, + "step": 3085 + }, + { + "epoch": 0.5936599423631124, + "grad_norm": 1.3347340822219849, + "learning_rate": 0.00025275638023865055, + "loss": 3.2262367248535155, + "step": 3090 + }, + { + "epoch": 0.5946205571565802, + "grad_norm": 0.8695631623268127, + "learning_rate": 0.0002525826744911693, + "loss": 3.228030776977539, + "step": 3095 + }, + { + "epoch": 0.595581171950048, + "grad_norm": 0.8885836601257324, + "learning_rate": 0.00025240870991707665, + "loss": 3.229400634765625, + "step": 3100 + }, + { + "epoch": 0.5965417867435159, + "grad_norm": 0.8682327270507812, + "learning_rate": 0.000252234486955303, + "loss": 3.2293159484863283, + "step": 3105 + }, + { + "epoch": 0.5975024015369836, + "grad_norm": 1.130051851272583, + "learning_rate": 0.00025206000604543083, + "loss": 3.2304660797119142, + "step": 3110 + }, + { + "epoch": 0.5984630163304515, + "grad_norm": 1.0737762451171875, + "learning_rate": 0.00025188526762769326, + "loss": 3.2278984069824217, + "step": 3115 + }, + { + "epoch": 0.5994236311239193, + "grad_norm": 1.096584677696228, + "learning_rate": 0.0002517102721429734, + "loss": 3.2275615692138673, + "step": 3120 + }, + { + "epoch": 0.6003842459173871, + "grad_norm": 1.1471421718597412, + "learning_rate": 0.00025153502003280267, + "loss": 3.225168991088867, + "step": 3125 + }, + { + "epoch": 0.6013448607108549, + "grad_norm": 0.8774202466011047, + "learning_rate": 0.0002513595117393602, + "loss": 3.225673294067383, + "step": 3130 + }, + { + "epoch": 0.6023054755043228, + "grad_norm": 1.1272785663604736, + "learning_rate": 0.00025118374770547136, + "loss": 3.228855514526367, + "step": 3135 + }, + { + "epoch": 0.6032660902977905, + "grad_norm": 1.0108392238616943, + "learning_rate": 0.0002510077283746069, + "loss": 3.226605224609375, + "step": 3140 + }, + { + "epoch": 0.6042267050912584, + "grad_norm": 0.9391249418258667, + "learning_rate": 0.00025083145419088165, + "loss": 3.2272396087646484, + "step": 3145 + }, + { + "epoch": 0.6051873198847262, + "grad_norm": 1.0647284984588623, + "learning_rate": 0.0002506549255990534, + "loss": 3.2265804290771483, + "step": 3150 + }, + { + "epoch": 0.6061479346781941, + "grad_norm": 1.143602967262268, + "learning_rate": 0.00025047814304452197, + "loss": 3.2242530822753905, + "step": 3155 + }, + { + "epoch": 0.6071085494716618, + "grad_norm": 1.098870873451233, + "learning_rate": 0.0002503011069733278, + "loss": 3.225702667236328, + "step": 3160 + }, + { + "epoch": 0.6080691642651297, + "grad_norm": 1.1901029348373413, + "learning_rate": 0.00025012381783215117, + "loss": 3.2249305725097654, + "step": 3165 + }, + { + "epoch": 0.6090297790585975, + "grad_norm": 1.254163384437561, + "learning_rate": 0.0002499462760683106, + "loss": 3.2239353179931642, + "step": 3170 + }, + { + "epoch": 0.6099903938520653, + "grad_norm": 1.3494505882263184, + "learning_rate": 0.00024976848212976237, + "loss": 3.225982666015625, + "step": 3175 + }, + { + "epoch": 0.6109510086455331, + "grad_norm": 0.9600538015365601, + "learning_rate": 0.00024959043646509866, + "loss": 3.2219696044921875, + "step": 3180 + }, + { + "epoch": 0.611911623439001, + "grad_norm": 1.1406611204147339, + "learning_rate": 0.00024941213952354703, + "loss": 3.225663757324219, + "step": 3185 + }, + { + "epoch": 0.6128722382324687, + "grad_norm": 1.276875376701355, + "learning_rate": 0.000249233591754969, + "loss": 3.226406478881836, + "step": 3190 + }, + { + "epoch": 0.6138328530259366, + "grad_norm": 1.0921087265014648, + "learning_rate": 0.00024905479360985884, + "loss": 3.227063751220703, + "step": 3195 + }, + { + "epoch": 0.6147934678194045, + "grad_norm": 1.0305696725845337, + "learning_rate": 0.0002488757455393426, + "loss": 3.2252223968505858, + "step": 3200 + }, + { + "epoch": 0.6157540826128722, + "grad_norm": 1.0612218379974365, + "learning_rate": 0.0002486964479951769, + "loss": 3.2239383697509765, + "step": 3205 + }, + { + "epoch": 0.6167146974063401, + "grad_norm": 1.011340856552124, + "learning_rate": 0.00024851690142974796, + "loss": 3.2294452667236326, + "step": 3210 + }, + { + "epoch": 0.6176753121998079, + "grad_norm": 1.0465643405914307, + "learning_rate": 0.0002483371062960701, + "loss": 3.2301746368408204, + "step": 3215 + }, + { + "epoch": 0.6186359269932757, + "grad_norm": 0.8956519961357117, + "learning_rate": 0.0002481570630477849, + "loss": 3.226789093017578, + "step": 3220 + }, + { + "epoch": 0.6195965417867435, + "grad_norm": 0.8143766522407532, + "learning_rate": 0.00024797677213915997, + "loss": 3.2227046966552733, + "step": 3225 + }, + { + "epoch": 0.6205571565802114, + "grad_norm": 1.0714408159255981, + "learning_rate": 0.00024779623402508766, + "loss": 3.225188064575195, + "step": 3230 + }, + { + "epoch": 0.6215177713736791, + "grad_norm": 1.0253653526306152, + "learning_rate": 0.0002476154491610843, + "loss": 3.225335693359375, + "step": 3235 + }, + { + "epoch": 0.622478386167147, + "grad_norm": 1.0260114669799805, + "learning_rate": 0.00024743441800328866, + "loss": 3.2250225067138674, + "step": 3240 + }, + { + "epoch": 0.6234390009606148, + "grad_norm": 1.0167043209075928, + "learning_rate": 0.0002472531410084607, + "loss": 3.222640609741211, + "step": 3245 + }, + { + "epoch": 0.6243996157540826, + "grad_norm": 1.0281862020492554, + "learning_rate": 0.00024707161863398114, + "loss": 3.223117446899414, + "step": 3250 + }, + { + "epoch": 0.6253602305475504, + "grad_norm": 1.1766680479049683, + "learning_rate": 0.0002468898513378494, + "loss": 3.2250850677490233, + "step": 3255 + }, + { + "epoch": 0.6263208453410183, + "grad_norm": 0.9479981064796448, + "learning_rate": 0.0002467078395786831, + "loss": 3.2229190826416017, + "step": 3260 + }, + { + "epoch": 0.627281460134486, + "grad_norm": 0.9690991640090942, + "learning_rate": 0.00024652558381571646, + "loss": 3.2238792419433593, + "step": 3265 + }, + { + "epoch": 0.6282420749279539, + "grad_norm": 1.127388596534729, + "learning_rate": 0.00024634308450879953, + "loss": 3.22578125, + "step": 3270 + }, + { + "epoch": 0.6292026897214217, + "grad_norm": 1.0700501203536987, + "learning_rate": 0.0002461603421183968, + "loss": 3.222941589355469, + "step": 3275 + }, + { + "epoch": 0.6301633045148896, + "grad_norm": 1.1317108869552612, + "learning_rate": 0.00024597735710558613, + "loss": 3.228685760498047, + "step": 3280 + }, + { + "epoch": 0.6311239193083573, + "grad_norm": 1.0256825685501099, + "learning_rate": 0.0002457941299320574, + "loss": 3.2236000061035157, + "step": 3285 + }, + { + "epoch": 0.6320845341018252, + "grad_norm": 1.0949205160140991, + "learning_rate": 0.0002456106610601115, + "loss": 3.223644256591797, + "step": 3290 + }, + { + "epoch": 0.633045148895293, + "grad_norm": 1.147100567817688, + "learning_rate": 0.0002454269509526593, + "loss": 3.222336196899414, + "step": 3295 + }, + { + "epoch": 0.6340057636887608, + "grad_norm": 1.0118300914764404, + "learning_rate": 0.0002452430000732203, + "loss": 3.2236583709716795, + "step": 3300 + }, + { + "epoch": 0.6349663784822286, + "grad_norm": 1.4460432529449463, + "learning_rate": 0.00024505880888592134, + "loss": 3.225933074951172, + "step": 3305 + }, + { + "epoch": 0.6359269932756965, + "grad_norm": 1.0727514028549194, + "learning_rate": 0.0002448743778554957, + "loss": 3.222777557373047, + "step": 3310 + }, + { + "epoch": 0.6368876080691642, + "grad_norm": 0.9514535069465637, + "learning_rate": 0.0002446897074472819, + "loss": 3.2218833923339845, + "step": 3315 + }, + { + "epoch": 0.6378482228626321, + "grad_norm": 0.8115454912185669, + "learning_rate": 0.0002445047981272221, + "loss": 3.2228424072265627, + "step": 3320 + }, + { + "epoch": 0.6388088376560999, + "grad_norm": 1.2015719413757324, + "learning_rate": 0.00024431965036186175, + "loss": 3.2233604431152343, + "step": 3325 + }, + { + "epoch": 0.6397694524495677, + "grad_norm": 1.0239508152008057, + "learning_rate": 0.00024413426461834746, + "loss": 3.223210906982422, + "step": 3330 + }, + { + "epoch": 0.6407300672430355, + "grad_norm": 0.8938039541244507, + "learning_rate": 0.0002439486413644265, + "loss": 3.2242591857910154, + "step": 3335 + }, + { + "epoch": 0.6416906820365034, + "grad_norm": 1.0734831094741821, + "learning_rate": 0.00024376278106844548, + "loss": 3.221482849121094, + "step": 3340 + }, + { + "epoch": 0.6426512968299711, + "grad_norm": 0.9349012970924377, + "learning_rate": 0.00024357668419934894, + "loss": 3.2233955383300783, + "step": 3345 + }, + { + "epoch": 0.643611911623439, + "grad_norm": 1.0495328903198242, + "learning_rate": 0.00024339035122667837, + "loss": 3.2246463775634764, + "step": 3350 + }, + { + "epoch": 0.6445725264169068, + "grad_norm": 1.152726173400879, + "learning_rate": 0.00024320378262057102, + "loss": 3.221533203125, + "step": 3355 + }, + { + "epoch": 0.6455331412103746, + "grad_norm": 1.1071127653121948, + "learning_rate": 0.00024301697885175847, + "loss": 3.2202079772949217, + "step": 3360 + }, + { + "epoch": 0.6464937560038425, + "grad_norm": 1.143314003944397, + "learning_rate": 0.00024282994039156598, + "loss": 3.228096771240234, + "step": 3365 + }, + { + "epoch": 0.6474543707973103, + "grad_norm": 1.0042517185211182, + "learning_rate": 0.00024264266771191066, + "loss": 3.2260887145996096, + "step": 3370 + }, + { + "epoch": 0.6484149855907781, + "grad_norm": 0.9083545804023743, + "learning_rate": 0.00024245516128530077, + "loss": 3.2202564239501954, + "step": 3375 + }, + { + "epoch": 0.6493756003842459, + "grad_norm": 0.8550533056259155, + "learning_rate": 0.0002422674215848342, + "loss": 3.22034912109375, + "step": 3380 + }, + { + "epoch": 0.6503362151777138, + "grad_norm": 0.9592113494873047, + "learning_rate": 0.0002420794490841975, + "loss": 3.22401123046875, + "step": 3385 + }, + { + "epoch": 0.6512968299711815, + "grad_norm": 0.8267397880554199, + "learning_rate": 0.00024189124425766454, + "loss": 3.2178253173828124, + "step": 3390 + }, + { + "epoch": 0.6522574447646494, + "grad_norm": 0.9734513759613037, + "learning_rate": 0.00024170280758009546, + "loss": 3.2190723419189453, + "step": 3395 + }, + { + "epoch": 0.6532180595581172, + "grad_norm": 1.0415617227554321, + "learning_rate": 0.00024151413952693524, + "loss": 3.2226608276367186, + "step": 3400 + }, + { + "epoch": 0.654178674351585, + "grad_norm": 0.9090771675109863, + "learning_rate": 0.00024132524057421287, + "loss": 3.221820831298828, + "step": 3405 + }, + { + "epoch": 0.6551392891450528, + "grad_norm": 0.8799076676368713, + "learning_rate": 0.0002411361111985396, + "loss": 3.2193443298339846, + "step": 3410 + }, + { + "epoch": 0.6560999039385207, + "grad_norm": 1.0817033052444458, + "learning_rate": 0.00024094675187710847, + "loss": 3.2185131072998048, + "step": 3415 + }, + { + "epoch": 0.6570605187319885, + "grad_norm": 0.9831625819206238, + "learning_rate": 0.0002407571630876923, + "loss": 3.2214286804199217, + "step": 3420 + }, + { + "epoch": 0.6580211335254563, + "grad_norm": 0.987235963344574, + "learning_rate": 0.00024056734530864313, + "loss": 3.219521713256836, + "step": 3425 + }, + { + "epoch": 0.6589817483189241, + "grad_norm": 1.047991394996643, + "learning_rate": 0.00024037729901889068, + "loss": 3.2219139099121095, + "step": 3430 + }, + { + "epoch": 0.659942363112392, + "grad_norm": 0.8672451376914978, + "learning_rate": 0.00024018702469794125, + "loss": 3.2201393127441404, + "step": 3435 + }, + { + "epoch": 0.6609029779058597, + "grad_norm": 1.1878256797790527, + "learning_rate": 0.0002399965228258764, + "loss": 3.2203353881835937, + "step": 3440 + }, + { + "epoch": 0.6618635926993276, + "grad_norm": 1.1251111030578613, + "learning_rate": 0.00023980579388335193, + "loss": 3.21944580078125, + "step": 3445 + }, + { + "epoch": 0.6628242074927954, + "grad_norm": 0.9487177729606628, + "learning_rate": 0.0002396148383515966, + "loss": 3.2208953857421876, + "step": 3450 + }, + { + "epoch": 0.6637848222862632, + "grad_norm": 0.9404230713844299, + "learning_rate": 0.00023942365671241072, + "loss": 3.22216796875, + "step": 3455 + }, + { + "epoch": 0.664745437079731, + "grad_norm": 1.1221790313720703, + "learning_rate": 0.00023923224944816522, + "loss": 3.220103454589844, + "step": 3460 + }, + { + "epoch": 0.6657060518731989, + "grad_norm": 1.1422487497329712, + "learning_rate": 0.00023904061704180027, + "loss": 3.2190887451171877, + "step": 3465 + }, + { + "epoch": 0.6666666666666666, + "grad_norm": 0.9574413299560547, + "learning_rate": 0.00023884875997682404, + "loss": 3.215955352783203, + "step": 3470 + }, + { + "epoch": 0.6676272814601345, + "grad_norm": 1.1348673105239868, + "learning_rate": 0.00023865667873731168, + "loss": 3.2160018920898437, + "step": 3475 + }, + { + "epoch": 0.6685878962536023, + "grad_norm": 1.0261040925979614, + "learning_rate": 0.00023846437380790368, + "loss": 3.221507263183594, + "step": 3480 + }, + { + "epoch": 0.6695485110470701, + "grad_norm": 1.009157657623291, + "learning_rate": 0.0002382718456738053, + "loss": 3.2182361602783205, + "step": 3485 + }, + { + "epoch": 0.6705091258405379, + "grad_norm": 0.9334181547164917, + "learning_rate": 0.00023807909482078475, + "loss": 3.221875762939453, + "step": 3490 + }, + { + "epoch": 0.6714697406340058, + "grad_norm": 1.0333514213562012, + "learning_rate": 0.0002378861217351721, + "loss": 3.2202327728271483, + "step": 3495 + }, + { + "epoch": 0.6724303554274735, + "grad_norm": 0.8510705828666687, + "learning_rate": 0.00023769292690385831, + "loss": 3.2205406188964845, + "step": 3500 + }, + { + "epoch": 0.6733909702209414, + "grad_norm": 1.0097713470458984, + "learning_rate": 0.0002374995108142938, + "loss": 3.2197288513183593, + "step": 3505 + }, + { + "epoch": 0.6743515850144092, + "grad_norm": 1.3777557611465454, + "learning_rate": 0.0002373058739544871, + "loss": 3.2197029113769533, + "step": 3510 + }, + { + "epoch": 0.675312199807877, + "grad_norm": 1.1440997123718262, + "learning_rate": 0.0002371120168130039, + "loss": 3.2237571716308593, + "step": 3515 + }, + { + "epoch": 0.6762728146013448, + "grad_norm": 1.1652058362960815, + "learning_rate": 0.0002369179398789657, + "loss": 3.216419982910156, + "step": 3520 + }, + { + "epoch": 0.6772334293948127, + "grad_norm": 0.7069166302680969, + "learning_rate": 0.00023672364364204853, + "loss": 3.2144538879394533, + "step": 3525 + }, + { + "epoch": 0.6781940441882806, + "grad_norm": 1.2215242385864258, + "learning_rate": 0.00023652912859248166, + "loss": 3.221971130371094, + "step": 3530 + }, + { + "epoch": 0.6791546589817483, + "grad_norm": 0.978817880153656, + "learning_rate": 0.00023633439522104658, + "loss": 3.21834716796875, + "step": 3535 + }, + { + "epoch": 0.6801152737752162, + "grad_norm": 1.0850077867507935, + "learning_rate": 0.0002361394440190755, + "loss": 3.216111755371094, + "step": 3540 + }, + { + "epoch": 0.681075888568684, + "grad_norm": 1.0549561977386475, + "learning_rate": 0.00023594427547845033, + "loss": 3.2226284027099608, + "step": 3545 + }, + { + "epoch": 0.6820365033621518, + "grad_norm": 1.0915031433105469, + "learning_rate": 0.00023574889009160134, + "loss": 3.220775604248047, + "step": 3550 + }, + { + "epoch": 0.6829971181556196, + "grad_norm": 1.2648931741714478, + "learning_rate": 0.00023555328835150587, + "loss": 3.219307708740234, + "step": 3555 + }, + { + "epoch": 0.6839577329490875, + "grad_norm": 0.8500558733940125, + "learning_rate": 0.00023535747075168726, + "loss": 3.2166213989257812, + "step": 3560 + }, + { + "epoch": 0.6849183477425552, + "grad_norm": 0.9885119795799255, + "learning_rate": 0.00023516143778621327, + "loss": 3.219788360595703, + "step": 3565 + }, + { + "epoch": 0.6858789625360231, + "grad_norm": 1.1124404668807983, + "learning_rate": 0.00023496518994969528, + "loss": 3.218476104736328, + "step": 3570 + }, + { + "epoch": 0.6868395773294909, + "grad_norm": 1.0537521839141846, + "learning_rate": 0.00023476872773728666, + "loss": 3.217354583740234, + "step": 3575 + }, + { + "epoch": 0.6878001921229587, + "grad_norm": 1.0711948871612549, + "learning_rate": 0.00023457205164468173, + "loss": 3.2162689208984374, + "step": 3580 + }, + { + "epoch": 0.6887608069164265, + "grad_norm": 1.0931755304336548, + "learning_rate": 0.0002343751621681145, + "loss": 3.217786407470703, + "step": 3585 + }, + { + "epoch": 0.6897214217098944, + "grad_norm": 0.9976537227630615, + "learning_rate": 0.00023417805980435736, + "loss": 3.216781997680664, + "step": 3590 + }, + { + "epoch": 0.6906820365033621, + "grad_norm": 0.8797841668128967, + "learning_rate": 0.00023398074505071964, + "loss": 3.21632080078125, + "step": 3595 + }, + { + "epoch": 0.69164265129683, + "grad_norm": 1.0468313694000244, + "learning_rate": 0.0002337832184050468, + "loss": 3.2173301696777346, + "step": 3600 + }, + { + "epoch": 0.6926032660902978, + "grad_norm": 1.2071698904037476, + "learning_rate": 0.0002335854803657188, + "loss": 3.2191944122314453, + "step": 3605 + }, + { + "epoch": 0.6935638808837656, + "grad_norm": 1.0584746599197388, + "learning_rate": 0.00023338753143164906, + "loss": 3.2196414947509764, + "step": 3610 + }, + { + "epoch": 0.6945244956772334, + "grad_norm": 0.9132346510887146, + "learning_rate": 0.00023318937210228295, + "loss": 3.215652847290039, + "step": 3615 + }, + { + "epoch": 0.6954851104707013, + "grad_norm": 0.9715785384178162, + "learning_rate": 0.00023299100287759686, + "loss": 3.215423583984375, + "step": 3620 + }, + { + "epoch": 0.696445725264169, + "grad_norm": 1.1859434843063354, + "learning_rate": 0.00023279242425809667, + "loss": 3.211594009399414, + "step": 3625 + }, + { + "epoch": 0.6974063400576369, + "grad_norm": 1.0725688934326172, + "learning_rate": 0.00023259363674481666, + "loss": 3.2200057983398436, + "step": 3630 + }, + { + "epoch": 0.6983669548511047, + "grad_norm": 1.0033189058303833, + "learning_rate": 0.00023239464083931802, + "loss": 3.217108154296875, + "step": 3635 + }, + { + "epoch": 0.6993275696445725, + "grad_norm": 1.0625147819519043, + "learning_rate": 0.00023219543704368792, + "loss": 3.2169322967529297, + "step": 3640 + }, + { + "epoch": 0.7002881844380403, + "grad_norm": 0.9788894653320312, + "learning_rate": 0.00023199602586053793, + "loss": 3.2101593017578125, + "step": 3645 + }, + { + "epoch": 0.7012487992315082, + "grad_norm": 0.9568902254104614, + "learning_rate": 0.0002317964077930029, + "loss": 3.215932846069336, + "step": 3650 + }, + { + "epoch": 0.7022094140249759, + "grad_norm": 1.3549705743789673, + "learning_rate": 0.00023159658334473974, + "loss": 3.2176162719726564, + "step": 3655 + }, + { + "epoch": 0.7031700288184438, + "grad_norm": 0.9975321292877197, + "learning_rate": 0.00023139655301992603, + "loss": 3.218878173828125, + "step": 3660 + }, + { + "epoch": 0.7041306436119116, + "grad_norm": 0.8793084621429443, + "learning_rate": 0.00023119631732325882, + "loss": 3.2155433654785157, + "step": 3665 + }, + { + "epoch": 0.7050912584053795, + "grad_norm": 1.2263166904449463, + "learning_rate": 0.0002309958767599532, + "loss": 3.2179286956787108, + "step": 3670 + }, + { + "epoch": 0.7060518731988472, + "grad_norm": 1.0385061502456665, + "learning_rate": 0.0002307952318357414, + "loss": 3.2134330749511717, + "step": 3675 + }, + { + "epoch": 0.7070124879923151, + "grad_norm": 1.019902229309082, + "learning_rate": 0.0002305943830568711, + "loss": 3.2176845550537108, + "step": 3680 + }, + { + "epoch": 0.7079731027857828, + "grad_norm": 0.9984716773033142, + "learning_rate": 0.00023039333093010433, + "loss": 3.218429946899414, + "step": 3685 + }, + { + "epoch": 0.7089337175792507, + "grad_norm": 1.0424875020980835, + "learning_rate": 0.00023019207596271634, + "loss": 3.2165111541748046, + "step": 3690 + }, + { + "epoch": 0.7098943323727186, + "grad_norm": 1.0895493030548096, + "learning_rate": 0.00022999061866249397, + "loss": 3.2162498474121093, + "step": 3695 + }, + { + "epoch": 0.7108549471661864, + "grad_norm": 1.0710501670837402, + "learning_rate": 0.00022978895953773464, + "loss": 3.2143184661865236, + "step": 3700 + }, + { + "epoch": 0.7118155619596542, + "grad_norm": 1.1927725076675415, + "learning_rate": 0.00022958709909724512, + "loss": 3.2180755615234373, + "step": 3705 + }, + { + "epoch": 0.712776176753122, + "grad_norm": 0.9517913460731506, + "learning_rate": 0.00022938503785033991, + "loss": 3.2159660339355467, + "step": 3710 + }, + { + "epoch": 0.7137367915465899, + "grad_norm": 1.0126185417175293, + "learning_rate": 0.00022918277630684032, + "loss": 3.2123146057128906, + "step": 3715 + }, + { + "epoch": 0.7146974063400576, + "grad_norm": 1.0625087022781372, + "learning_rate": 0.0002289803149770729, + "loss": 3.215711212158203, + "step": 3720 + }, + { + "epoch": 0.7156580211335255, + "grad_norm": 1.2499691247940063, + "learning_rate": 0.00022877765437186843, + "loss": 3.215605545043945, + "step": 3725 + }, + { + "epoch": 0.7166186359269933, + "grad_norm": 1.0240933895111084, + "learning_rate": 0.0002285747950025604, + "loss": 3.2169639587402346, + "step": 3730 + }, + { + "epoch": 0.7175792507204611, + "grad_norm": 1.0153264999389648, + "learning_rate": 0.00022837173738098375, + "loss": 3.2149864196777345, + "step": 3735 + }, + { + "epoch": 0.7185398655139289, + "grad_norm": 0.7960460782051086, + "learning_rate": 0.00022816848201947376, + "loss": 3.217276382446289, + "step": 3740 + }, + { + "epoch": 0.7195004803073968, + "grad_norm": 0.886752724647522, + "learning_rate": 0.0002279650294308645, + "loss": 3.2126708984375, + "step": 3745 + }, + { + "epoch": 0.7204610951008645, + "grad_norm": 0.9281694889068604, + "learning_rate": 0.00022776138012848776, + "loss": 3.2142318725585937, + "step": 3750 + }, + { + "epoch": 0.7214217098943324, + "grad_norm": 0.9548509120941162, + "learning_rate": 0.00022755753462617162, + "loss": 3.2146896362304687, + "step": 3755 + }, + { + "epoch": 0.7223823246878002, + "grad_norm": 1.099318265914917, + "learning_rate": 0.0002273534934382392, + "loss": 3.2121913909912108, + "step": 3760 + }, + { + "epoch": 0.723342939481268, + "grad_norm": 1.0480355024337769, + "learning_rate": 0.00022714925707950734, + "loss": 3.2157440185546875, + "step": 3765 + }, + { + "epoch": 0.7243035542747358, + "grad_norm": 0.8996202349662781, + "learning_rate": 0.0002269448260652853, + "loss": 3.2162132263183594, + "step": 3770 + }, + { + "epoch": 0.7252641690682037, + "grad_norm": 1.0007187128067017, + "learning_rate": 0.00022674020091137357, + "loss": 3.216075897216797, + "step": 3775 + }, + { + "epoch": 0.7262247838616714, + "grad_norm": 1.0066853761672974, + "learning_rate": 0.0002265353821340623, + "loss": 3.2114944458007812, + "step": 3780 + }, + { + "epoch": 0.7271853986551393, + "grad_norm": 0.8113794922828674, + "learning_rate": 0.00022633037025013034, + "loss": 3.212854766845703, + "step": 3785 + }, + { + "epoch": 0.7281460134486071, + "grad_norm": 0.8382619619369507, + "learning_rate": 0.00022612516577684373, + "loss": 3.2154186248779295, + "step": 3790 + }, + { + "epoch": 0.729106628242075, + "grad_norm": 0.7889550924301147, + "learning_rate": 0.0002259197692319544, + "loss": 3.212788391113281, + "step": 3795 + }, + { + "epoch": 0.7300672430355427, + "grad_norm": 0.988488495349884, + "learning_rate": 0.00022571418113369885, + "loss": 3.2164848327636717, + "step": 3800 + }, + { + "epoch": 0.7310278578290106, + "grad_norm": 1.2173198461532593, + "learning_rate": 0.00022550840200079696, + "loss": 3.216341018676758, + "step": 3805 + }, + { + "epoch": 0.7319884726224783, + "grad_norm": 0.9247362017631531, + "learning_rate": 0.00022530243235245067, + "loss": 3.214260482788086, + "step": 3810 + }, + { + "epoch": 0.7329490874159462, + "grad_norm": 1.0379600524902344, + "learning_rate": 0.0002250962727083424, + "loss": 3.211573028564453, + "step": 3815 + }, + { + "epoch": 0.733909702209414, + "grad_norm": 1.0794739723205566, + "learning_rate": 0.00022488992358863416, + "loss": 3.2143383026123047, + "step": 3820 + }, + { + "epoch": 0.7348703170028819, + "grad_norm": 0.7588670253753662, + "learning_rate": 0.00022468338551396598, + "loss": 3.2139129638671875, + "step": 3825 + }, + { + "epoch": 0.7358309317963496, + "grad_norm": 0.9358298182487488, + "learning_rate": 0.00022447665900545452, + "loss": 3.2157051086425783, + "step": 3830 + }, + { + "epoch": 0.7367915465898175, + "grad_norm": 1.0112359523773193, + "learning_rate": 0.00022426974458469206, + "loss": 3.2114688873291017, + "step": 3835 + }, + { + "epoch": 0.7377521613832853, + "grad_norm": 1.0056138038635254, + "learning_rate": 0.00022406264277374487, + "loss": 3.211741638183594, + "step": 3840 + }, + { + "epoch": 0.7387127761767531, + "grad_norm": 1.0002721548080444, + "learning_rate": 0.00022385535409515203, + "loss": 3.214643096923828, + "step": 3845 + }, + { + "epoch": 0.7396733909702209, + "grad_norm": 1.0511516332626343, + "learning_rate": 0.0002236478790719242, + "loss": 3.2131919860839844, + "step": 3850 + }, + { + "epoch": 0.7406340057636888, + "grad_norm": 1.0013073682785034, + "learning_rate": 0.00022344021822754212, + "loss": 3.2114883422851563, + "step": 3855 + }, + { + "epoch": 0.7415946205571565, + "grad_norm": 0.9666048288345337, + "learning_rate": 0.0002232323720859554, + "loss": 3.2139522552490236, + "step": 3860 + }, + { + "epoch": 0.7425552353506244, + "grad_norm": 0.9765871167182922, + "learning_rate": 0.0002230243411715812, + "loss": 3.216571044921875, + "step": 3865 + }, + { + "epoch": 0.7435158501440923, + "grad_norm": 0.8558294773101807, + "learning_rate": 0.00022281612600930282, + "loss": 3.211492156982422, + "step": 3870 + }, + { + "epoch": 0.74447646493756, + "grad_norm": 0.9790803790092468, + "learning_rate": 0.00022260772712446848, + "loss": 3.212940979003906, + "step": 3875 + }, + { + "epoch": 0.7454370797310279, + "grad_norm": 1.115885853767395, + "learning_rate": 0.00022239914504289002, + "loss": 3.215241241455078, + "step": 3880 + }, + { + "epoch": 0.7463976945244957, + "grad_norm": 0.8012208342552185, + "learning_rate": 0.00022219038029084134, + "loss": 3.2129531860351563, + "step": 3885 + }, + { + "epoch": 0.7473583093179635, + "grad_norm": 0.9198108315467834, + "learning_rate": 0.00022198143339505736, + "loss": 3.2131282806396486, + "step": 3890 + }, + { + "epoch": 0.7483189241114313, + "grad_norm": 0.9384759068489075, + "learning_rate": 0.00022177230488273255, + "loss": 3.2148651123046874, + "step": 3895 + }, + { + "epoch": 0.7492795389048992, + "grad_norm": 0.9330629706382751, + "learning_rate": 0.00022156299528151957, + "loss": 3.2130111694335937, + "step": 3900 + }, + { + "epoch": 0.7502401536983669, + "grad_norm": 1.0961406230926514, + "learning_rate": 0.00022135350511952805, + "loss": 3.209431457519531, + "step": 3905 + }, + { + "epoch": 0.7512007684918348, + "grad_norm": 1.0598336458206177, + "learning_rate": 0.0002211438349253231, + "loss": 3.2150421142578125, + "step": 3910 + }, + { + "epoch": 0.7521613832853026, + "grad_norm": 0.8614126443862915, + "learning_rate": 0.00022093398522792418, + "loss": 3.2113304138183594, + "step": 3915 + }, + { + "epoch": 0.7531219980787704, + "grad_norm": 1.2163376808166504, + "learning_rate": 0.0002207239565568036, + "loss": 3.213220977783203, + "step": 3920 + }, + { + "epoch": 0.7540826128722382, + "grad_norm": 1.068865180015564, + "learning_rate": 0.00022051374944188524, + "loss": 3.2152915954589845, + "step": 3925 + }, + { + "epoch": 0.7550432276657061, + "grad_norm": 0.8628993034362793, + "learning_rate": 0.0002203033644135432, + "loss": 3.2100101470947267, + "step": 3930 + }, + { + "epoch": 0.7560038424591738, + "grad_norm": 1.0070104598999023, + "learning_rate": 0.00022009280200260053, + "loss": 3.214314651489258, + "step": 3935 + }, + { + "epoch": 0.7569644572526417, + "grad_norm": 1.1911405324935913, + "learning_rate": 0.00021988206274032776, + "loss": 3.2116607666015624, + "step": 3940 + }, + { + "epoch": 0.7579250720461095, + "grad_norm": 1.0605741739273071, + "learning_rate": 0.00021967114715844163, + "loss": 3.210451126098633, + "step": 3945 + }, + { + "epoch": 0.7588856868395774, + "grad_norm": 1.010385274887085, + "learning_rate": 0.0002194600557891039, + "loss": 3.2117488861083983, + "step": 3950 + }, + { + "epoch": 0.7598463016330451, + "grad_norm": 0.9285851716995239, + "learning_rate": 0.00021924878916491962, + "loss": 3.2095130920410155, + "step": 3955 + }, + { + "epoch": 0.760806916426513, + "grad_norm": 0.9447981715202332, + "learning_rate": 0.00021903734781893625, + "loss": 3.214548873901367, + "step": 3960 + }, + { + "epoch": 0.7617675312199808, + "grad_norm": 1.0844841003417969, + "learning_rate": 0.000218825732284642, + "loss": 3.2142662048339843, + "step": 3965 + }, + { + "epoch": 0.7627281460134486, + "grad_norm": 0.9654716849327087, + "learning_rate": 0.00021861394309596446, + "loss": 3.214899444580078, + "step": 3970 + }, + { + "epoch": 0.7636887608069164, + "grad_norm": 0.8219507336616516, + "learning_rate": 0.00021840198078726964, + "loss": 3.2075408935546874, + "step": 3975 + }, + { + "epoch": 0.7646493756003843, + "grad_norm": 1.0514135360717773, + "learning_rate": 0.00021818984589336006, + "loss": 3.2124794006347654, + "step": 3980 + }, + { + "epoch": 0.765609990393852, + "grad_norm": 0.764786958694458, + "learning_rate": 0.0002179775389494739, + "loss": 3.210370635986328, + "step": 3985 + }, + { + "epoch": 0.7665706051873199, + "grad_norm": 0.9463275671005249, + "learning_rate": 0.0002177650604912833, + "loss": 3.214555358886719, + "step": 3990 + }, + { + "epoch": 0.7675312199807877, + "grad_norm": 0.9087357521057129, + "learning_rate": 0.0002175524110548932, + "loss": 3.211328125, + "step": 3995 + }, + { + "epoch": 0.7684918347742555, + "grad_norm": 0.8839055299758911, + "learning_rate": 0.00021733959117684008, + "loss": 3.208209991455078, + "step": 4000 + }, + { + "epoch": 0.7694524495677233, + "grad_norm": 1.0021069049835205, + "learning_rate": 0.00021712660139409015, + "loss": 3.213267517089844, + "step": 4005 + }, + { + "epoch": 0.7704130643611912, + "grad_norm": 1.089827060699463, + "learning_rate": 0.0002169134422440386, + "loss": 3.209787368774414, + "step": 4010 + }, + { + "epoch": 0.7713736791546589, + "grad_norm": 0.9058536887168884, + "learning_rate": 0.00021670011426450772, + "loss": 3.2052375793457033, + "step": 4015 + }, + { + "epoch": 0.7723342939481268, + "grad_norm": 0.8882672190666199, + "learning_rate": 0.000216486617993746, + "loss": 3.2103240966796873, + "step": 4020 + }, + { + "epoch": 0.7732949087415946, + "grad_norm": 0.7981977462768555, + "learning_rate": 0.00021627295397042635, + "loss": 3.2096282958984377, + "step": 4025 + }, + { + "epoch": 0.7742555235350624, + "grad_norm": 0.9157530665397644, + "learning_rate": 0.00021605912273364513, + "loss": 3.213287353515625, + "step": 4030 + }, + { + "epoch": 0.7752161383285303, + "grad_norm": 0.7963248491287231, + "learning_rate": 0.00021584512482292038, + "loss": 3.2098575592041017, + "step": 4035 + }, + { + "epoch": 0.7761767531219981, + "grad_norm": 1.0376338958740234, + "learning_rate": 0.00021563096077819083, + "loss": 3.2072002410888674, + "step": 4040 + }, + { + "epoch": 0.777137367915466, + "grad_norm": 1.159191608428955, + "learning_rate": 0.00021541663113981433, + "loss": 3.2127113342285156, + "step": 4045 + }, + { + "epoch": 0.7780979827089337, + "grad_norm": 1.213563323020935, + "learning_rate": 0.0002152021364485665, + "loss": 3.2094375610351564, + "step": 4050 + }, + { + "epoch": 0.7790585975024016, + "grad_norm": 1.1174225807189941, + "learning_rate": 0.00021498747724563953, + "loss": 3.2087074279785157, + "step": 4055 + }, + { + "epoch": 0.7800192122958693, + "grad_norm": 1.0365922451019287, + "learning_rate": 0.00021477265407264051, + "loss": 3.2079147338867187, + "step": 4060 + }, + { + "epoch": 0.7809798270893372, + "grad_norm": 1.0354453325271606, + "learning_rate": 0.00021455766747159044, + "loss": 3.2089813232421873, + "step": 4065 + }, + { + "epoch": 0.781940441882805, + "grad_norm": 1.0720642805099487, + "learning_rate": 0.0002143425179849226, + "loss": 3.207681655883789, + "step": 4070 + }, + { + "epoch": 0.7829010566762729, + "grad_norm": 0.9447735548019409, + "learning_rate": 0.00021412720615548105, + "loss": 3.2113746643066405, + "step": 4075 + }, + { + "epoch": 0.7838616714697406, + "grad_norm": 0.9376322627067566, + "learning_rate": 0.00021391173252651978, + "loss": 3.2086830139160156, + "step": 4080 + }, + { + "epoch": 0.7848222862632085, + "grad_norm": 1.0265626907348633, + "learning_rate": 0.00021369609764170075, + "loss": 3.2098388671875, + "step": 4085 + }, + { + "epoch": 0.7857829010566763, + "grad_norm": 1.1230889558792114, + "learning_rate": 0.00021348030204509303, + "loss": 3.208446502685547, + "step": 4090 + }, + { + "epoch": 0.7867435158501441, + "grad_norm": 0.8800603747367859, + "learning_rate": 0.00021326434628117088, + "loss": 3.208513641357422, + "step": 4095 + }, + { + "epoch": 0.7877041306436119, + "grad_norm": 1.0674163103103638, + "learning_rate": 0.00021304823089481298, + "loss": 3.2110557556152344, + "step": 4100 + }, + { + "epoch": 0.7886647454370798, + "grad_norm": 0.8815765976905823, + "learning_rate": 0.00021283195643130058, + "loss": 3.210008239746094, + "step": 4105 + }, + { + "epoch": 0.7896253602305475, + "grad_norm": 1.0902719497680664, + "learning_rate": 0.00021261552343631633, + "loss": 3.2107887268066406, + "step": 4110 + }, + { + "epoch": 0.7905859750240154, + "grad_norm": 0.859643816947937, + "learning_rate": 0.00021239893245594287, + "loss": 3.210626220703125, + "step": 4115 + }, + { + "epoch": 0.7915465898174832, + "grad_norm": 0.9734321236610413, + "learning_rate": 0.00021218218403666148, + "loss": 3.207859420776367, + "step": 4120 + }, + { + "epoch": 0.792507204610951, + "grad_norm": 0.8096666932106018, + "learning_rate": 0.00021196527872535068, + "loss": 3.2104820251464843, + "step": 4125 + }, + { + "epoch": 0.7934678194044188, + "grad_norm": 1.0135759115219116, + "learning_rate": 0.0002117482170692847, + "loss": 3.2095088958740234, + "step": 4130 + }, + { + "epoch": 0.7944284341978867, + "grad_norm": 0.9503042101860046, + "learning_rate": 0.00021153099961613257, + "loss": 3.2078937530517577, + "step": 4135 + }, + { + "epoch": 0.7953890489913544, + "grad_norm": 0.8183991312980652, + "learning_rate": 0.00021131362691395608, + "loss": 3.2082897186279298, + "step": 4140 + }, + { + "epoch": 0.7963496637848223, + "grad_norm": 0.9838517308235168, + "learning_rate": 0.00021109609951120887, + "loss": 3.2070068359375, + "step": 4145 + }, + { + "epoch": 0.7973102785782901, + "grad_norm": 1.0721310377120972, + "learning_rate": 0.00021087841795673488, + "loss": 3.206983947753906, + "step": 4150 + }, + { + "epoch": 0.7982708933717579, + "grad_norm": 1.0762015581130981, + "learning_rate": 0.00021066058279976704, + "loss": 3.2114967346191405, + "step": 4155 + }, + { + "epoch": 0.7992315081652257, + "grad_norm": 0.7675787806510925, + "learning_rate": 0.0002104425945899258, + "loss": 3.2034461975097654, + "step": 4160 + }, + { + "epoch": 0.8001921229586936, + "grad_norm": 1.0134726762771606, + "learning_rate": 0.00021022445387721767, + "loss": 3.207468032836914, + "step": 4165 + }, + { + "epoch": 0.8011527377521613, + "grad_norm": 0.9967617988586426, + "learning_rate": 0.00021000616121203422, + "loss": 3.2145057678222657, + "step": 4170 + }, + { + "epoch": 0.8021133525456292, + "grad_norm": 1.0330275297164917, + "learning_rate": 0.00020978771714515016, + "loss": 3.205029296875, + "step": 4175 + }, + { + "epoch": 0.803073967339097, + "grad_norm": 0.8626118898391724, + "learning_rate": 0.00020956912222772222, + "loss": 3.2084095001220705, + "step": 4180 + }, + { + "epoch": 0.8040345821325648, + "grad_norm": 1.186023235321045, + "learning_rate": 0.0002093503770112879, + "loss": 3.209725189208984, + "step": 4185 + }, + { + "epoch": 0.8049951969260326, + "grad_norm": 0.943437397480011, + "learning_rate": 0.00020913148204776378, + "loss": 3.209857177734375, + "step": 4190 + }, + { + "epoch": 0.8059558117195005, + "grad_norm": 1.167366623878479, + "learning_rate": 0.00020891243788944428, + "loss": 3.2103233337402344, + "step": 4195 + }, + { + "epoch": 0.8069164265129684, + "grad_norm": 0.9998597502708435, + "learning_rate": 0.00020869324508900026, + "loss": 3.211322784423828, + "step": 4200 + }, + { + "epoch": 0.8078770413064361, + "grad_norm": 0.9104480147361755, + "learning_rate": 0.00020847390419947778, + "loss": 3.2085220336914064, + "step": 4205 + }, + { + "epoch": 0.808837656099904, + "grad_norm": 1.0278260707855225, + "learning_rate": 0.0002082544157742963, + "loss": 3.2045223236083986, + "step": 4210 + }, + { + "epoch": 0.8097982708933718, + "grad_norm": 0.9279727339744568, + "learning_rate": 0.0002080347803672476, + "loss": 3.2089134216308595, + "step": 4215 + }, + { + "epoch": 0.8107588856868396, + "grad_norm": 0.9136155247688293, + "learning_rate": 0.00020781499853249441, + "loss": 3.2083240509033204, + "step": 4220 + }, + { + "epoch": 0.8117195004803074, + "grad_norm": 0.8539382219314575, + "learning_rate": 0.00020759507082456885, + "loss": 3.204279327392578, + "step": 4225 + }, + { + "epoch": 0.8126801152737753, + "grad_norm": 1.1511064767837524, + "learning_rate": 0.00020737499779837105, + "loss": 3.2057140350341795, + "step": 4230 + }, + { + "epoch": 0.813640730067243, + "grad_norm": 0.8153745532035828, + "learning_rate": 0.00020715478000916783, + "loss": 3.2046875, + "step": 4235 + }, + { + "epoch": 0.8146013448607109, + "grad_norm": 0.8916031122207642, + "learning_rate": 0.00020693441801259135, + "loss": 3.210984802246094, + "step": 4240 + }, + { + "epoch": 0.8155619596541787, + "grad_norm": 0.9601288437843323, + "learning_rate": 0.00020671391236463742, + "loss": 3.2044132232666014, + "step": 4245 + }, + { + "epoch": 0.8165225744476465, + "grad_norm": 0.9480450749397278, + "learning_rate": 0.00020649326362166449, + "loss": 3.2047096252441407, + "step": 4250 + }, + { + "epoch": 0.8174831892411143, + "grad_norm": 1.3009711503982544, + "learning_rate": 0.0002062724723403919, + "loss": 3.210066223144531, + "step": 4255 + }, + { + "epoch": 0.8184438040345822, + "grad_norm": 0.8482257127761841, + "learning_rate": 0.00020605153907789876, + "loss": 3.2040420532226563, + "step": 4260 + }, + { + "epoch": 0.8194044188280499, + "grad_norm": 1.1215349435806274, + "learning_rate": 0.00020583046439162232, + "loss": 3.209978485107422, + "step": 4265 + }, + { + "epoch": 0.8203650336215178, + "grad_norm": 1.0416927337646484, + "learning_rate": 0.00020560924883935674, + "loss": 3.20634765625, + "step": 4270 + }, + { + "epoch": 0.8213256484149856, + "grad_norm": 0.9002500176429749, + "learning_rate": 0.00020538789297925154, + "loss": 3.210291290283203, + "step": 4275 + }, + { + "epoch": 0.8222862632084534, + "grad_norm": 0.7959718108177185, + "learning_rate": 0.00020516639736981027, + "loss": 3.205914306640625, + "step": 4280 + }, + { + "epoch": 0.8232468780019212, + "grad_norm": 0.9458216428756714, + "learning_rate": 0.000204944762569889, + "loss": 3.2048385620117186, + "step": 4285 + }, + { + "epoch": 0.8242074927953891, + "grad_norm": 1.0240002870559692, + "learning_rate": 0.00020472298913869514, + "loss": 3.2106887817382814, + "step": 4290 + }, + { + "epoch": 0.8251681075888568, + "grad_norm": 0.9878635406494141, + "learning_rate": 0.00020450107763578582, + "loss": 3.206460952758789, + "step": 4295 + }, + { + "epoch": 0.8261287223823247, + "grad_norm": 0.961821973323822, + "learning_rate": 0.00020427902862106645, + "loss": 3.208537292480469, + "step": 4300 + }, + { + "epoch": 0.8270893371757925, + "grad_norm": 1.0129098892211914, + "learning_rate": 0.00020405684265478955, + "loss": 3.2066085815429686, + "step": 4305 + }, + { + "epoch": 0.8280499519692603, + "grad_norm": 1.2535499334335327, + "learning_rate": 0.00020383452029755308, + "loss": 3.206789016723633, + "step": 4310 + }, + { + "epoch": 0.8290105667627281, + "grad_norm": 1.0553781986236572, + "learning_rate": 0.00020361206211029915, + "loss": 3.208069610595703, + "step": 4315 + }, + { + "epoch": 0.829971181556196, + "grad_norm": 1.033327341079712, + "learning_rate": 0.00020338946865431257, + "loss": 3.208452606201172, + "step": 4320 + }, + { + "epoch": 0.8309317963496637, + "grad_norm": 1.1007755994796753, + "learning_rate": 0.00020316674049121954, + "loss": 3.205517578125, + "step": 4325 + }, + { + "epoch": 0.8318924111431316, + "grad_norm": 1.0913561582565308, + "learning_rate": 0.00020294387818298596, + "loss": 3.205488586425781, + "step": 4330 + }, + { + "epoch": 0.8328530259365994, + "grad_norm": 1.073475956916809, + "learning_rate": 0.00020272088229191638, + "loss": 3.2070037841796877, + "step": 4335 + }, + { + "epoch": 0.8338136407300673, + "grad_norm": 0.8359492421150208, + "learning_rate": 0.00020249775338065224, + "loss": 3.2025604248046875, + "step": 4340 + }, + { + "epoch": 0.834774255523535, + "grad_norm": 1.1163265705108643, + "learning_rate": 0.0002022744920121707, + "loss": 3.205812454223633, + "step": 4345 + }, + { + "epoch": 0.8357348703170029, + "grad_norm": 1.0195002555847168, + "learning_rate": 0.0002020510987497832, + "loss": 3.2072071075439452, + "step": 4350 + }, + { + "epoch": 0.8366954851104706, + "grad_norm": 1.06333327293396, + "learning_rate": 0.0002018275741571337, + "loss": 3.2045135498046875, + "step": 4355 + }, + { + "epoch": 0.8376560999039385, + "grad_norm": 1.07416570186615, + "learning_rate": 0.00020160391879819775, + "loss": 3.2041862487792967, + "step": 4360 + }, + { + "epoch": 0.8386167146974063, + "grad_norm": 0.8247685432434082, + "learning_rate": 0.00020138013323728072, + "loss": 3.204395294189453, + "step": 4365 + }, + { + "epoch": 0.8395773294908742, + "grad_norm": 0.7512625455856323, + "learning_rate": 0.00020115621803901658, + "loss": 3.2026763916015626, + "step": 4370 + }, + { + "epoch": 0.840537944284342, + "grad_norm": 0.6788026094436646, + "learning_rate": 0.0002009321737683664, + "loss": 3.202665328979492, + "step": 4375 + }, + { + "epoch": 0.8414985590778098, + "grad_norm": 1.178909182548523, + "learning_rate": 0.00020070800099061676, + "loss": 3.203816604614258, + "step": 4380 + }, + { + "epoch": 0.8424591738712777, + "grad_norm": 1.056028127670288, + "learning_rate": 0.00020048370027137864, + "loss": 3.203491973876953, + "step": 4385 + }, + { + "epoch": 0.8434197886647454, + "grad_norm": 0.9850196838378906, + "learning_rate": 0.0002002592721765857, + "loss": 3.203214645385742, + "step": 4390 + }, + { + "epoch": 0.8443804034582133, + "grad_norm": 0.812002956867218, + "learning_rate": 0.00020003471727249317, + "loss": 3.2064640045166017, + "step": 4395 + }, + { + "epoch": 0.8453410182516811, + "grad_norm": 0.7780818343162537, + "learning_rate": 0.00019981003612567594, + "loss": 3.2039905548095704, + "step": 4400 + }, + { + "epoch": 0.8463016330451489, + "grad_norm": 0.7282761931419373, + "learning_rate": 0.0001995852293030277, + "loss": 3.2027183532714845, + "step": 4405 + }, + { + "epoch": 0.8472622478386167, + "grad_norm": 1.597205400466919, + "learning_rate": 0.0001993602973717591, + "loss": 3.2074440002441404, + "step": 4410 + }, + { + "epoch": 0.8482228626320846, + "grad_norm": 0.7743385434150696, + "learning_rate": 0.0001991352408993965, + "loss": 3.20322265625, + "step": 4415 + }, + { + "epoch": 0.8491834774255523, + "grad_norm": 0.8802571296691895, + "learning_rate": 0.00019891006045378047, + "loss": 3.2041389465332033, + "step": 4420 + }, + { + "epoch": 0.8501440922190202, + "grad_norm": 0.9674323797225952, + "learning_rate": 0.00019868475660306435, + "loss": 3.2038337707519533, + "step": 4425 + }, + { + "epoch": 0.851104707012488, + "grad_norm": 0.8163370490074158, + "learning_rate": 0.00019845932991571284, + "loss": 3.204595947265625, + "step": 4430 + }, + { + "epoch": 0.8520653218059558, + "grad_norm": 0.8586170077323914, + "learning_rate": 0.00019823378096050067, + "loss": 3.204967498779297, + "step": 4435 + }, + { + "epoch": 0.8530259365994236, + "grad_norm": 0.9709478616714478, + "learning_rate": 0.00019800811030651095, + "loss": 3.200875091552734, + "step": 4440 + }, + { + "epoch": 0.8539865513928915, + "grad_norm": 0.8611642718315125, + "learning_rate": 0.00019778231852313386, + "loss": 3.2068603515625, + "step": 4445 + }, + { + "epoch": 0.8549471661863592, + "grad_norm": 0.8717654347419739, + "learning_rate": 0.00019755640618006532, + "loss": 3.2012969970703127, + "step": 4450 + }, + { + "epoch": 0.8559077809798271, + "grad_norm": 0.9735947251319885, + "learning_rate": 0.0001973303738473053, + "loss": 3.2008651733398437, + "step": 4455 + }, + { + "epoch": 0.8568683957732949, + "grad_norm": 0.9653201699256897, + "learning_rate": 0.0001971042220951565, + "loss": 3.2013267517089843, + "step": 4460 + }, + { + "epoch": 0.8578290105667628, + "grad_norm": 0.8183594346046448, + "learning_rate": 0.00019687795149422315, + "loss": 3.204022979736328, + "step": 4465 + }, + { + "epoch": 0.8587896253602305, + "grad_norm": 0.8697395920753479, + "learning_rate": 0.00019665156261540898, + "loss": 3.202830505371094, + "step": 4470 + }, + { + "epoch": 0.8597502401536984, + "grad_norm": 0.829225480556488, + "learning_rate": 0.00019642505602991654, + "loss": 3.2021392822265624, + "step": 4475 + }, + { + "epoch": 0.8607108549471661, + "grad_norm": 1.1149920225143433, + "learning_rate": 0.0001961984323092451, + "loss": 3.205643081665039, + "step": 4480 + }, + { + "epoch": 0.861671469740634, + "grad_norm": 0.8502190113067627, + "learning_rate": 0.00019597169202518954, + "loss": 3.2061065673828124, + "step": 4485 + }, + { + "epoch": 0.8626320845341018, + "grad_norm": 0.7430519461631775, + "learning_rate": 0.0001957448357498389, + "loss": 3.2033714294433593, + "step": 4490 + }, + { + "epoch": 0.8635926993275697, + "grad_norm": 0.8692427277565002, + "learning_rate": 0.0001955178640555748, + "loss": 3.2003097534179688, + "step": 4495 + }, + { + "epoch": 0.8645533141210374, + "grad_norm": 0.809934139251709, + "learning_rate": 0.00019529077751507016, + "loss": 3.2053035736083983, + "step": 4500 + }, + { + "epoch": 0.8655139289145053, + "grad_norm": 0.6470674276351929, + "learning_rate": 0.00019506357670128749, + "loss": 3.2052642822265627, + "step": 4505 + }, + { + "epoch": 0.866474543707973, + "grad_norm": 0.9643762111663818, + "learning_rate": 0.00019483626218747794, + "loss": 3.203343963623047, + "step": 4510 + }, + { + "epoch": 0.8674351585014409, + "grad_norm": 1.0639313459396362, + "learning_rate": 0.00019460883454717922, + "loss": 3.2024681091308596, + "step": 4515 + }, + { + "epoch": 0.8683957732949087, + "grad_norm": 0.8861836194992065, + "learning_rate": 0.00019438129435421462, + "loss": 3.201047897338867, + "step": 4520 + }, + { + "epoch": 0.8693563880883766, + "grad_norm": 1.0320836305618286, + "learning_rate": 0.00019415364218269146, + "loss": 3.2003639221191404, + "step": 4525 + }, + { + "epoch": 0.8703170028818443, + "grad_norm": 0.9835689663887024, + "learning_rate": 0.00019392587860699942, + "loss": 3.2048595428466795, + "step": 4530 + }, + { + "epoch": 0.8712776176753122, + "grad_norm": 0.7003117799758911, + "learning_rate": 0.00019369800420180943, + "loss": 3.204084014892578, + "step": 4535 + }, + { + "epoch": 0.8722382324687801, + "grad_norm": 0.9389500021934509, + "learning_rate": 0.00019347001954207193, + "loss": 3.2038955688476562, + "step": 4540 + }, + { + "epoch": 0.8731988472622478, + "grad_norm": 1.1108181476593018, + "learning_rate": 0.00019324192520301566, + "loss": 3.200461196899414, + "step": 4545 + }, + { + "epoch": 0.8741594620557157, + "grad_norm": 0.84727942943573, + "learning_rate": 0.00019301372176014605, + "loss": 3.2037403106689455, + "step": 4550 + }, + { + "epoch": 0.8751200768491835, + "grad_norm": 0.776715874671936, + "learning_rate": 0.00019278540978924378, + "loss": 3.1999557495117186, + "step": 4555 + }, + { + "epoch": 0.8760806916426513, + "grad_norm": 0.873356282711029, + "learning_rate": 0.0001925569898663633, + "loss": 3.201426696777344, + "step": 4560 + }, + { + "epoch": 0.8770413064361191, + "grad_norm": 1.0719802379608154, + "learning_rate": 0.00019232846256783163, + "loss": 3.2020469665527345, + "step": 4565 + }, + { + "epoch": 0.878001921229587, + "grad_norm": 0.8912343978881836, + "learning_rate": 0.00019209982847024655, + "loss": 3.200605010986328, + "step": 4570 + }, + { + "epoch": 0.8789625360230547, + "grad_norm": 0.8289459347724915, + "learning_rate": 0.00019187108815047523, + "loss": 3.2027835845947266, + "step": 4575 + }, + { + "epoch": 0.8799231508165226, + "grad_norm": 0.7044758200645447, + "learning_rate": 0.00019164224218565313, + "loss": 3.2026336669921873, + "step": 4580 + }, + { + "epoch": 0.8808837656099904, + "grad_norm": 0.8316428661346436, + "learning_rate": 0.00019141329115318203, + "loss": 3.204214096069336, + "step": 4585 + }, + { + "epoch": 0.8818443804034583, + "grad_norm": 0.9814732074737549, + "learning_rate": 0.00019118423563072885, + "loss": 3.2025177001953127, + "step": 4590 + }, + { + "epoch": 0.882804995196926, + "grad_norm": 0.6341441869735718, + "learning_rate": 0.0001909550761962242, + "loss": 3.2031497955322266, + "step": 4595 + }, + { + "epoch": 0.8837656099903939, + "grad_norm": 0.977326512336731, + "learning_rate": 0.00019072581342786084, + "loss": 3.199177551269531, + "step": 4600 + }, + { + "epoch": 0.8847262247838616, + "grad_norm": 0.7784291505813599, + "learning_rate": 0.00019049644790409225, + "loss": 3.198267936706543, + "step": 4605 + }, + { + "epoch": 0.8856868395773295, + "grad_norm": 1.1206945180892944, + "learning_rate": 0.00019026698020363107, + "loss": 3.205250549316406, + "step": 4610 + }, + { + "epoch": 0.8866474543707973, + "grad_norm": 0.8689852356910706, + "learning_rate": 0.00019003741090544804, + "loss": 3.2022071838378907, + "step": 4615 + }, + { + "epoch": 0.8876080691642652, + "grad_norm": 1.1133830547332764, + "learning_rate": 0.00018980774058876995, + "loss": 3.2035804748535157, + "step": 4620 + }, + { + "epoch": 0.8885686839577329, + "grad_norm": 0.8632625937461853, + "learning_rate": 0.00018957796983307858, + "loss": 3.2034885406494142, + "step": 4625 + }, + { + "epoch": 0.8895292987512008, + "grad_norm": 0.8747223615646362, + "learning_rate": 0.0001893480992181091, + "loss": 3.2007545471191405, + "step": 4630 + }, + { + "epoch": 0.8904899135446686, + "grad_norm": 1.075469732284546, + "learning_rate": 0.00018911812932384872, + "loss": 3.2042964935302733, + "step": 4635 + }, + { + "epoch": 0.8914505283381364, + "grad_norm": 1.0559624433517456, + "learning_rate": 0.000188888060730535, + "loss": 3.199691963195801, + "step": 4640 + }, + { + "epoch": 0.8924111431316042, + "grad_norm": 1.0888924598693848, + "learning_rate": 0.00018865789401865458, + "loss": 3.202299118041992, + "step": 4645 + }, + { + "epoch": 0.8933717579250721, + "grad_norm": 0.7396308779716492, + "learning_rate": 0.0001884276297689418, + "loss": 3.19757080078125, + "step": 4650 + }, + { + "epoch": 0.8943323727185398, + "grad_norm": 0.8118330240249634, + "learning_rate": 0.00018819726856237685, + "loss": 3.2041015625, + "step": 4655 + }, + { + "epoch": 0.8952929875120077, + "grad_norm": 0.9625673890113831, + "learning_rate": 0.00018796681098018486, + "loss": 3.2014694213867188, + "step": 4660 + }, + { + "epoch": 0.8962536023054755, + "grad_norm": 1.0025596618652344, + "learning_rate": 0.00018773625760383375, + "loss": 3.2010726928710938, + "step": 4665 + }, + { + "epoch": 0.8972142170989433, + "grad_norm": 1.0931344032287598, + "learning_rate": 0.00018750560901503353, + "loss": 3.2006298065185548, + "step": 4670 + }, + { + "epoch": 0.8981748318924111, + "grad_norm": 1.0470645427703857, + "learning_rate": 0.00018727486579573407, + "loss": 3.1973636627197264, + "step": 4675 + }, + { + "epoch": 0.899135446685879, + "grad_norm": 0.9046833515167236, + "learning_rate": 0.00018704402852812431, + "loss": 3.2017478942871094, + "step": 4680 + }, + { + "epoch": 0.9000960614793467, + "grad_norm": 0.9769369959831238, + "learning_rate": 0.00018681309779463033, + "loss": 3.2017196655273437, + "step": 4685 + }, + { + "epoch": 0.9010566762728146, + "grad_norm": 0.8879526853561401, + "learning_rate": 0.00018658207417791405, + "loss": 3.2009010314941406, + "step": 4690 + }, + { + "epoch": 0.9020172910662824, + "grad_norm": 0.7592387795448303, + "learning_rate": 0.00018635095826087175, + "loss": 3.2003753662109373, + "step": 4695 + }, + { + "epoch": 0.9029779058597502, + "grad_norm": 0.9972590804100037, + "learning_rate": 0.00018611975062663263, + "loss": 3.1992095947265624, + "step": 4700 + }, + { + "epoch": 0.9039385206532181, + "grad_norm": 0.9059445261955261, + "learning_rate": 0.0001858884518585572, + "loss": 3.195656585693359, + "step": 4705 + }, + { + "epoch": 0.9048991354466859, + "grad_norm": 0.8296246528625488, + "learning_rate": 0.00018565706254023605, + "loss": 3.196135711669922, + "step": 4710 + }, + { + "epoch": 0.9058597502401537, + "grad_norm": 0.9912506937980652, + "learning_rate": 0.00018542558325548814, + "loss": 3.1977779388427736, + "step": 4715 + }, + { + "epoch": 0.9068203650336215, + "grad_norm": 0.7257289290428162, + "learning_rate": 0.00018519401458835948, + "loss": 3.197937774658203, + "step": 4720 + }, + { + "epoch": 0.9077809798270894, + "grad_norm": 1.0594534873962402, + "learning_rate": 0.00018496235712312154, + "loss": 3.193804168701172, + "step": 4725 + }, + { + "epoch": 0.9087415946205571, + "grad_norm": 1.0351169109344482, + "learning_rate": 0.00018473061144426986, + "loss": 3.196752166748047, + "step": 4730 + }, + { + "epoch": 0.909702209414025, + "grad_norm": 0.8392277359962463, + "learning_rate": 0.0001844987781365226, + "loss": 3.2042442321777345, + "step": 4735 + }, + { + "epoch": 0.9106628242074928, + "grad_norm": 0.6720691323280334, + "learning_rate": 0.00018426685778481897, + "loss": 3.2033058166503907, + "step": 4740 + }, + { + "epoch": 0.9116234390009607, + "grad_norm": 1.009544014930725, + "learning_rate": 0.00018403485097431778, + "loss": 3.200358200073242, + "step": 4745 + }, + { + "epoch": 0.9125840537944284, + "grad_norm": 0.8520932793617249, + "learning_rate": 0.00018380275829039602, + "loss": 3.201191711425781, + "step": 4750 + }, + { + "epoch": 0.9135446685878963, + "grad_norm": 0.9514408111572266, + "learning_rate": 0.0001835705803186474, + "loss": 3.194804000854492, + "step": 4755 + }, + { + "epoch": 0.914505283381364, + "grad_norm": 0.8069214224815369, + "learning_rate": 0.00018333831764488065, + "loss": 3.1965736389160155, + "step": 4760 + }, + { + "epoch": 0.9154658981748319, + "grad_norm": 0.8542043566703796, + "learning_rate": 0.00018310597085511844, + "loss": 3.1960010528564453, + "step": 4765 + }, + { + "epoch": 0.9164265129682997, + "grad_norm": 0.7166752815246582, + "learning_rate": 0.00018287354053559546, + "loss": 3.1996959686279296, + "step": 4770 + }, + { + "epoch": 0.9173871277617676, + "grad_norm": 1.2815790176391602, + "learning_rate": 0.0001826410272727574, + "loss": 3.200403594970703, + "step": 4775 + }, + { + "epoch": 0.9183477425552353, + "grad_norm": 0.9036345481872559, + "learning_rate": 0.00018240843165325882, + "loss": 3.203938674926758, + "step": 4780 + }, + { + "epoch": 0.9193083573487032, + "grad_norm": 0.9565702080726624, + "learning_rate": 0.00018217575426396256, + "loss": 3.197901153564453, + "step": 4785 + }, + { + "epoch": 0.920268972142171, + "grad_norm": 0.8386558890342712, + "learning_rate": 0.00018194299569193747, + "loss": 3.194584274291992, + "step": 4790 + }, + { + "epoch": 0.9212295869356388, + "grad_norm": 1.0965979099273682, + "learning_rate": 0.0001817101565244573, + "loss": 3.20220947265625, + "step": 4795 + }, + { + "epoch": 0.9221902017291066, + "grad_norm": 1.1644456386566162, + "learning_rate": 0.00018147723734899916, + "loss": 3.1982275009155274, + "step": 4800 + }, + { + "epoch": 0.9231508165225745, + "grad_norm": 1.0380982160568237, + "learning_rate": 0.00018124423875324198, + "loss": 3.1980897903442385, + "step": 4805 + }, + { + "epoch": 0.9241114313160422, + "grad_norm": 0.823890209197998, + "learning_rate": 0.00018101116132506522, + "loss": 3.1994667053222656, + "step": 4810 + }, + { + "epoch": 0.9250720461095101, + "grad_norm": 0.5867325663566589, + "learning_rate": 0.00018077800565254702, + "loss": 3.1975902557373046, + "step": 4815 + }, + { + "epoch": 0.9260326609029779, + "grad_norm": 1.1908833980560303, + "learning_rate": 0.00018054477232396312, + "loss": 3.198208808898926, + "step": 4820 + }, + { + "epoch": 0.9269932756964457, + "grad_norm": 1.0711902379989624, + "learning_rate": 0.00018031146192778517, + "loss": 3.1953447341918944, + "step": 4825 + }, + { + "epoch": 0.9279538904899135, + "grad_norm": 0.9716858863830566, + "learning_rate": 0.0001800780750526792, + "loss": 3.1982431411743164, + "step": 4830 + }, + { + "epoch": 0.9289145052833814, + "grad_norm": 0.9620060920715332, + "learning_rate": 0.00017984461228750422, + "loss": 3.196470260620117, + "step": 4835 + }, + { + "epoch": 0.9298751200768491, + "grad_norm": 1.015932321548462, + "learning_rate": 0.00017961107422131075, + "loss": 3.199555206298828, + "step": 4840 + }, + { + "epoch": 0.930835734870317, + "grad_norm": 0.9855188727378845, + "learning_rate": 0.00017937746144333934, + "loss": 3.2012577056884766, + "step": 4845 + }, + { + "epoch": 0.9317963496637848, + "grad_norm": 0.8330841660499573, + "learning_rate": 0.0001791437745430189, + "loss": 3.1955251693725586, + "step": 4850 + }, + { + "epoch": 0.9327569644572526, + "grad_norm": 0.8580245971679688, + "learning_rate": 0.00017891001410996554, + "loss": 3.1982753753662108, + "step": 4855 + }, + { + "epoch": 0.9337175792507204, + "grad_norm": 1.1361298561096191, + "learning_rate": 0.00017867618073398074, + "loss": 3.202219009399414, + "step": 4860 + }, + { + "epoch": 0.9346781940441883, + "grad_norm": 0.9495155215263367, + "learning_rate": 0.00017844227500505016, + "loss": 3.1957366943359373, + "step": 4865 + }, + { + "epoch": 0.9356388088376562, + "grad_norm": 0.8877089619636536, + "learning_rate": 0.00017820829751334194, + "loss": 3.198602867126465, + "step": 4870 + }, + { + "epoch": 0.9365994236311239, + "grad_norm": 0.7768687605857849, + "learning_rate": 0.0001779742488492052, + "loss": 3.1972635269165037, + "step": 4875 + }, + { + "epoch": 0.9375600384245918, + "grad_norm": 1.0583457946777344, + "learning_rate": 0.00017774012960316887, + "loss": 3.1948143005371095, + "step": 4880 + }, + { + "epoch": 0.9385206532180596, + "grad_norm": 0.9406375288963318, + "learning_rate": 0.00017750594036593974, + "loss": 3.1969860076904295, + "step": 4885 + }, + { + "epoch": 0.9394812680115274, + "grad_norm": 1.0013177394866943, + "learning_rate": 0.00017727168172840133, + "loss": 3.1973243713378907, + "step": 4890 + }, + { + "epoch": 0.9404418828049952, + "grad_norm": 0.7509773373603821, + "learning_rate": 0.00017703735428161216, + "loss": 3.198769760131836, + "step": 4895 + }, + { + "epoch": 0.9414024975984631, + "grad_norm": 0.8109461069107056, + "learning_rate": 0.00017680295861680445, + "loss": 3.1947399139404298, + "step": 4900 + }, + { + "epoch": 0.9423631123919308, + "grad_norm": 0.8327275514602661, + "learning_rate": 0.0001765684953253825, + "loss": 3.196604919433594, + "step": 4905 + }, + { + "epoch": 0.9433237271853987, + "grad_norm": 0.8777264356613159, + "learning_rate": 0.00017633396499892125, + "loss": 3.197701644897461, + "step": 4910 + }, + { + "epoch": 0.9442843419788665, + "grad_norm": 1.045627474784851, + "learning_rate": 0.00017609936822916475, + "loss": 3.204631805419922, + "step": 4915 + }, + { + "epoch": 0.9452449567723343, + "grad_norm": 1.0442372560501099, + "learning_rate": 0.0001758647056080248, + "loss": 3.2004959106445314, + "step": 4920 + }, + { + "epoch": 0.9462055715658021, + "grad_norm": 0.7651126384735107, + "learning_rate": 0.00017562997772757916, + "loss": 3.1948537826538086, + "step": 4925 + }, + { + "epoch": 0.94716618635927, + "grad_norm": 0.790465235710144, + "learning_rate": 0.00017539518518007043, + "loss": 3.196913719177246, + "step": 4930 + }, + { + "epoch": 0.9481268011527377, + "grad_norm": 0.9765505194664001, + "learning_rate": 0.00017516032855790424, + "loss": 3.1958955764770507, + "step": 4935 + }, + { + "epoch": 0.9490874159462056, + "grad_norm": 1.0204505920410156, + "learning_rate": 0.00017492540845364798, + "loss": 3.197770690917969, + "step": 4940 + }, + { + "epoch": 0.9500480307396734, + "grad_norm": 0.7875639796257019, + "learning_rate": 0.00017469042546002913, + "loss": 3.1972192764282226, + "step": 4945 + }, + { + "epoch": 0.9510086455331412, + "grad_norm": 0.8441899418830872, + "learning_rate": 0.00017445538016993393, + "loss": 3.1970443725585938, + "step": 4950 + }, + { + "epoch": 0.951969260326609, + "grad_norm": 0.9154574275016785, + "learning_rate": 0.00017422027317640575, + "loss": 3.1967355728149416, + "step": 4955 + }, + { + "epoch": 0.9529298751200769, + "grad_norm": 0.6774746775627136, + "learning_rate": 0.00017398510507264363, + "loss": 3.1978105545043944, + "step": 4960 + }, + { + "epoch": 0.9538904899135446, + "grad_norm": 0.9463170170783997, + "learning_rate": 0.0001737498764520009, + "loss": 3.195131301879883, + "step": 4965 + }, + { + "epoch": 0.9548511047070125, + "grad_norm": 1.0156409740447998, + "learning_rate": 0.00017351458790798338, + "loss": 3.197483253479004, + "step": 4970 + }, + { + "epoch": 0.9558117195004803, + "grad_norm": 0.8678054213523865, + "learning_rate": 0.0001732792400342483, + "loss": 3.1987661361694335, + "step": 4975 + }, + { + "epoch": 0.9567723342939481, + "grad_norm": 0.8004742860794067, + "learning_rate": 0.00017304383342460242, + "loss": 3.1965717315673827, + "step": 4980 + }, + { + "epoch": 0.9577329490874159, + "grad_norm": 1.0178972482681274, + "learning_rate": 0.0001728083686730008, + "loss": 3.197255325317383, + "step": 4985 + }, + { + "epoch": 0.9586935638808838, + "grad_norm": 0.7633374333381653, + "learning_rate": 0.00017257284637354524, + "loss": 3.1934268951416014, + "step": 4990 + }, + { + "epoch": 0.9596541786743515, + "grad_norm": 0.735908031463623, + "learning_rate": 0.00017233726712048253, + "loss": 3.194230079650879, + "step": 4995 + }, + { + "epoch": 0.9606147934678194, + "grad_norm": 0.936837911605835, + "learning_rate": 0.0001721016315082034, + "loss": 3.1968971252441407, + "step": 5000 + }, + { + "epoch": 0.9615754082612872, + "grad_norm": 0.8966473937034607, + "learning_rate": 0.00017186594013124057, + "loss": 3.2022254943847654, + "step": 5005 + }, + { + "epoch": 0.962536023054755, + "grad_norm": 0.8859227299690247, + "learning_rate": 0.00017163019358426763, + "loss": 3.19702205657959, + "step": 5010 + }, + { + "epoch": 0.9634966378482228, + "grad_norm": 0.6885091662406921, + "learning_rate": 0.00017139439246209728, + "loss": 3.1955480575561523, + "step": 5015 + }, + { + "epoch": 0.9644572526416907, + "grad_norm": 0.6677486896514893, + "learning_rate": 0.00017115853735967995, + "loss": 3.1977615356445312, + "step": 5020 + }, + { + "epoch": 0.9654178674351584, + "grad_norm": 1.0436347723007202, + "learning_rate": 0.00017092262887210232, + "loss": 3.195005416870117, + "step": 5025 + }, + { + "epoch": 0.9663784822286263, + "grad_norm": 0.9044774174690247, + "learning_rate": 0.0001706866675945856, + "loss": 3.1969493865966796, + "step": 5030 + }, + { + "epoch": 0.9673390970220941, + "grad_norm": 0.9574539661407471, + "learning_rate": 0.00017045065412248434, + "loss": 3.194741058349609, + "step": 5035 + }, + { + "epoch": 0.968299711815562, + "grad_norm": 0.9477970004081726, + "learning_rate": 0.00017021458905128477, + "loss": 3.1963451385498045, + "step": 5040 + }, + { + "epoch": 0.9692603266090298, + "grad_norm": 0.8571361899375916, + "learning_rate": 0.00016997847297660324, + "loss": 3.194466972351074, + "step": 5045 + }, + { + "epoch": 0.9702209414024976, + "grad_norm": 0.8901039361953735, + "learning_rate": 0.00016974230649418487, + "loss": 3.1929143905639648, + "step": 5050 + }, + { + "epoch": 0.9711815561959655, + "grad_norm": 1.027086615562439, + "learning_rate": 0.00016950609019990187, + "loss": 3.191559982299805, + "step": 5055 + }, + { + "epoch": 0.9721421709894332, + "grad_norm": 0.8591023683547974, + "learning_rate": 0.00016926982468975225, + "loss": 3.199761962890625, + "step": 5060 + }, + { + "epoch": 0.9731027857829011, + "grad_norm": 1.0826267004013062, + "learning_rate": 0.00016903351055985806, + "loss": 3.1961299896240236, + "step": 5065 + }, + { + "epoch": 0.9740634005763689, + "grad_norm": 1.0031397342681885, + "learning_rate": 0.0001687971484064642, + "loss": 3.1959823608398437, + "step": 5070 + }, + { + "epoch": 0.9750240153698367, + "grad_norm": 0.7752434015274048, + "learning_rate": 0.00016856073882593646, + "loss": 3.1921688079833985, + "step": 5075 + }, + { + "epoch": 0.9759846301633045, + "grad_norm": 0.911412239074707, + "learning_rate": 0.00016832428241476063, + "loss": 3.191498947143555, + "step": 5080 + }, + { + "epoch": 0.9769452449567724, + "grad_norm": 0.9647287726402283, + "learning_rate": 0.00016808777976954042, + "loss": 3.1929500579833983, + "step": 5085 + }, + { + "epoch": 0.9779058597502401, + "grad_norm": 0.7081168293952942, + "learning_rate": 0.00016785123148699624, + "loss": 3.1945589065551756, + "step": 5090 + }, + { + "epoch": 0.978866474543708, + "grad_norm": 0.8361444473266602, + "learning_rate": 0.00016761463816396374, + "loss": 3.1967111587524415, + "step": 5095 + }, + { + "epoch": 0.9798270893371758, + "grad_norm": 0.8908597230911255, + "learning_rate": 0.00016737800039739212, + "loss": 3.1946334838867188, + "step": 5100 + }, + { + "epoch": 0.9807877041306436, + "grad_norm": 0.8121721744537354, + "learning_rate": 0.0001671413187843427, + "loss": 3.195412826538086, + "step": 5105 + }, + { + "epoch": 0.9817483189241114, + "grad_norm": 0.6848476529121399, + "learning_rate": 0.00016690459392198752, + "loss": 3.191677284240723, + "step": 5110 + }, + { + "epoch": 0.9827089337175793, + "grad_norm": 0.6763195991516113, + "learning_rate": 0.00016666782640760766, + "loss": 3.1958782196044924, + "step": 5115 + }, + { + "epoch": 0.983669548511047, + "grad_norm": 0.8343232870101929, + "learning_rate": 0.00016643101683859186, + "loss": 3.197767639160156, + "step": 5120 + }, + { + "epoch": 0.9846301633045149, + "grad_norm": 0.9125416874885559, + "learning_rate": 0.00016619416581243498, + "loss": 3.1921886444091796, + "step": 5125 + }, + { + "epoch": 0.9855907780979827, + "grad_norm": 0.8966071009635925, + "learning_rate": 0.00016595727392673643, + "loss": 3.1952947616577148, + "step": 5130 + }, + { + "epoch": 0.9865513928914506, + "grad_norm": 0.6765633225440979, + "learning_rate": 0.00016572034177919875, + "loss": 3.1933788299560546, + "step": 5135 + }, + { + "epoch": 0.9875120076849183, + "grad_norm": 0.814222514629364, + "learning_rate": 0.00016548336996762608, + "loss": 3.1904666900634764, + "step": 5140 + }, + { + "epoch": 0.9884726224783862, + "grad_norm": 0.9140084981918335, + "learning_rate": 0.00016524635908992253, + "loss": 3.1964300155639647, + "step": 5145 + }, + { + "epoch": 0.989433237271854, + "grad_norm": 0.9220796227455139, + "learning_rate": 0.00016500930974409092, + "loss": 3.191013526916504, + "step": 5150 + }, + { + "epoch": 0.9903938520653218, + "grad_norm": 0.7357509732246399, + "learning_rate": 0.00016477222252823107, + "loss": 3.192600059509277, + "step": 5155 + }, + { + "epoch": 0.9913544668587896, + "grad_norm": 0.9331154227256775, + "learning_rate": 0.00016453509804053833, + "loss": 3.195221519470215, + "step": 5160 + }, + { + "epoch": 0.9923150816522575, + "grad_norm": 0.8373104929924011, + "learning_rate": 0.00016429793687930215, + "loss": 3.195628356933594, + "step": 5165 + }, + { + "epoch": 0.9932756964457252, + "grad_norm": 0.8984804153442383, + "learning_rate": 0.0001640607396429044, + "loss": 3.192228889465332, + "step": 5170 + }, + { + "epoch": 0.9942363112391931, + "grad_norm": 0.8978838324546814, + "learning_rate": 0.00016382350692981812, + "loss": 3.1939905166625975, + "step": 5175 + }, + { + "epoch": 0.9951969260326609, + "grad_norm": 0.6365435719490051, + "learning_rate": 0.00016358623933860567, + "loss": 3.1925621032714844, + "step": 5180 + }, + { + "epoch": 0.9961575408261287, + "grad_norm": 0.7753834128379822, + "learning_rate": 0.00016334893746791762, + "loss": 3.192973327636719, + "step": 5185 + }, + { + "epoch": 0.9971181556195965, + "grad_norm": 0.9735800623893738, + "learning_rate": 0.0001631116019164909, + "loss": 3.1959808349609373, + "step": 5190 + }, + { + "epoch": 0.9980787704130644, + "grad_norm": 0.6787250638008118, + "learning_rate": 0.00016287423328314746, + "loss": 3.193808746337891, + "step": 5195 + }, + { + "epoch": 0.9990393852065321, + "grad_norm": 0.7312700152397156, + "learning_rate": 0.00016263683216679274, + "loss": 3.1927206039428713, + "step": 5200 + }, + { + "epoch": 1.0, + "grad_norm": 0.7704550623893738, + "learning_rate": 0.00016239939916641404, + "loss": 3.1963579177856447, + "step": 5205 + }, + { + "epoch": 1.0009606147934678, + "grad_norm": 0.7807424664497375, + "learning_rate": 0.00016216193488107926, + "loss": 3.1920791625976563, + "step": 5210 + }, + { + "epoch": 1.0019212295869357, + "grad_norm": 0.9543339014053345, + "learning_rate": 0.0001619244399099351, + "loss": 3.1910972595214844, + "step": 5215 + }, + { + "epoch": 1.0028818443804035, + "grad_norm": 0.7766204476356506, + "learning_rate": 0.00016168691485220573, + "loss": 3.194061851501465, + "step": 5220 + }, + { + "epoch": 1.0038424591738713, + "grad_norm": 0.8086463809013367, + "learning_rate": 0.00016144936030719126, + "loss": 3.1898380279541017, + "step": 5225 + }, + { + "epoch": 1.004803073967339, + "grad_norm": 1.1027806997299194, + "learning_rate": 0.00016121177687426617, + "loss": 3.193951416015625, + "step": 5230 + }, + { + "epoch": 1.005763688760807, + "grad_norm": 0.996457576751709, + "learning_rate": 0.00016097416515287787, + "loss": 3.188690757751465, + "step": 5235 + }, + { + "epoch": 1.0067243035542748, + "grad_norm": 0.7975702881813049, + "learning_rate": 0.00016073652574254504, + "loss": 3.1924358367919923, + "step": 5240 + }, + { + "epoch": 1.0076849183477425, + "grad_norm": 0.7981067299842834, + "learning_rate": 0.00016049885924285638, + "loss": 3.1881412506103515, + "step": 5245 + }, + { + "epoch": 1.0086455331412103, + "grad_norm": 0.8105031251907349, + "learning_rate": 0.00016026116625346876, + "loss": 3.192718505859375, + "step": 5250 + }, + { + "epoch": 1.0096061479346783, + "grad_norm": 0.7112568616867065, + "learning_rate": 0.00016002344737410602, + "loss": 3.187815856933594, + "step": 5255 + }, + { + "epoch": 1.010566762728146, + "grad_norm": 0.8126310706138611, + "learning_rate": 0.0001597857032045573, + "loss": 3.191788673400879, + "step": 5260 + }, + { + "epoch": 1.0115273775216138, + "grad_norm": 0.733974814414978, + "learning_rate": 0.00015954793434467545, + "loss": 3.1942237854003905, + "step": 5265 + }, + { + "epoch": 1.0124879923150816, + "grad_norm": 0.6596633791923523, + "learning_rate": 0.0001593101413943758, + "loss": 3.189275932312012, + "step": 5270 + }, + { + "epoch": 1.0134486071085496, + "grad_norm": 0.8669559955596924, + "learning_rate": 0.0001590723249536343, + "loss": 3.1921119689941406, + "step": 5275 + }, + { + "epoch": 1.0144092219020173, + "grad_norm": 1.2266993522644043, + "learning_rate": 0.00015883448562248624, + "loss": 3.191496467590332, + "step": 5280 + }, + { + "epoch": 1.015369836695485, + "grad_norm": 0.8078704476356506, + "learning_rate": 0.00015859662400102462, + "loss": 3.1911800384521483, + "step": 5285 + }, + { + "epoch": 1.0163304514889528, + "grad_norm": 0.6551411151885986, + "learning_rate": 0.00015835874068939878, + "loss": 3.188446044921875, + "step": 5290 + }, + { + "epoch": 1.0172910662824208, + "grad_norm": 0.8124401569366455, + "learning_rate": 0.0001581208362878126, + "loss": 3.1922882080078123, + "step": 5295 + }, + { + "epoch": 1.0182516810758886, + "grad_norm": 0.8051562309265137, + "learning_rate": 0.00015788291139652344, + "loss": 3.190486717224121, + "step": 5300 + }, + { + "epoch": 1.0192122958693564, + "grad_norm": 0.9743736386299133, + "learning_rate": 0.00015764496661584013, + "loss": 3.1894975662231446, + "step": 5305 + }, + { + "epoch": 1.0201729106628241, + "grad_norm": 0.7459861040115356, + "learning_rate": 0.00015740700254612177, + "loss": 3.193043518066406, + "step": 5310 + }, + { + "epoch": 1.021133525456292, + "grad_norm": 1.2092807292938232, + "learning_rate": 0.0001571690197877761, + "loss": 3.1933567047119142, + "step": 5315 + }, + { + "epoch": 1.0220941402497599, + "grad_norm": 0.7785642147064209, + "learning_rate": 0.00015693101894125806, + "loss": 3.1891590118408204, + "step": 5320 + }, + { + "epoch": 1.0230547550432276, + "grad_norm": 0.7812630534172058, + "learning_rate": 0.00015669300060706823, + "loss": 3.190948486328125, + "step": 5325 + }, + { + "epoch": 1.0240153698366954, + "grad_norm": 0.9370684027671814, + "learning_rate": 0.0001564549653857512, + "loss": 3.1883705139160154, + "step": 5330 + }, + { + "epoch": 1.0249759846301634, + "grad_norm": 0.8281055092811584, + "learning_rate": 0.00015621691387789432, + "loss": 3.1863418579101563, + "step": 5335 + }, + { + "epoch": 1.0259365994236311, + "grad_norm": 1.0316886901855469, + "learning_rate": 0.000155978846684126, + "loss": 3.19388370513916, + "step": 5340 + }, + { + "epoch": 1.026897214217099, + "grad_norm": 0.758702278137207, + "learning_rate": 0.00015574076440511407, + "loss": 3.1906862258911133, + "step": 5345 + }, + { + "epoch": 1.0278578290105667, + "grad_norm": 0.5779961347579956, + "learning_rate": 0.00015550266764156466, + "loss": 3.190965461730957, + "step": 5350 + }, + { + "epoch": 1.0288184438040346, + "grad_norm": 0.9982998371124268, + "learning_rate": 0.00015526455699422023, + "loss": 3.190749740600586, + "step": 5355 + }, + { + "epoch": 1.0297790585975024, + "grad_norm": 1.105873465538025, + "learning_rate": 0.00015502643306385846, + "loss": 3.1928503036499025, + "step": 5360 + }, + { + "epoch": 1.0307396733909702, + "grad_norm": 0.9287274479866028, + "learning_rate": 0.00015478829645129046, + "loss": 3.1860063552856444, + "step": 5365 + }, + { + "epoch": 1.031700288184438, + "grad_norm": 0.7572025656700134, + "learning_rate": 0.00015455014775735923, + "loss": 3.1904781341552733, + "step": 5370 + }, + { + "epoch": 1.032660902977906, + "grad_norm": 0.8866919279098511, + "learning_rate": 0.0001543119875829385, + "loss": 3.1870819091796876, + "step": 5375 + }, + { + "epoch": 1.0336215177713737, + "grad_norm": 0.7839595675468445, + "learning_rate": 0.00015407381652893066, + "loss": 3.1915761947631838, + "step": 5380 + }, + { + "epoch": 1.0345821325648414, + "grad_norm": 0.7243456244468689, + "learning_rate": 0.00015383563519626582, + "loss": 3.1893665313720705, + "step": 5385 + }, + { + "epoch": 1.0355427473583094, + "grad_norm": 1.0272892713546753, + "learning_rate": 0.0001535974441858999, + "loss": 3.187778854370117, + "step": 5390 + }, + { + "epoch": 1.0365033621517772, + "grad_norm": 1.069861650466919, + "learning_rate": 0.00015335924409881323, + "loss": 3.1904586791992187, + "step": 5395 + }, + { + "epoch": 1.037463976945245, + "grad_norm": 0.8298363089561462, + "learning_rate": 0.00015312103553600913, + "loss": 3.1916831970214843, + "step": 5400 + }, + { + "epoch": 1.0384245917387127, + "grad_norm": 0.7426701784133911, + "learning_rate": 0.00015288281909851213, + "loss": 3.1897172927856445, + "step": 5405 + }, + { + "epoch": 1.0393852065321807, + "grad_norm": 0.8290877938270569, + "learning_rate": 0.00015264459538736686, + "loss": 3.1861778259277345, + "step": 5410 + }, + { + "epoch": 1.0403458213256485, + "grad_norm": 0.5183548927307129, + "learning_rate": 0.00015240636500363607, + "loss": 3.185122489929199, + "step": 5415 + }, + { + "epoch": 1.0413064361191162, + "grad_norm": 0.9525085687637329, + "learning_rate": 0.00015216812854839957, + "loss": 3.187310791015625, + "step": 5420 + }, + { + "epoch": 1.042267050912584, + "grad_norm": 0.9072858691215515, + "learning_rate": 0.0001519298866227523, + "loss": 3.1853969573974608, + "step": 5425 + }, + { + "epoch": 1.043227665706052, + "grad_norm": 0.8549609780311584, + "learning_rate": 0.00015169163982780304, + "loss": 3.1908397674560547, + "step": 5430 + }, + { + "epoch": 1.0441882804995197, + "grad_norm": 0.5929837822914124, + "learning_rate": 0.00015145338876467303, + "loss": 3.1879791259765624, + "step": 5435 + }, + { + "epoch": 1.0451488952929875, + "grad_norm": 0.8749310970306396, + "learning_rate": 0.00015121513403449402, + "loss": 3.1909906387329103, + "step": 5440 + }, + { + "epoch": 1.0461095100864553, + "grad_norm": 0.6831730604171753, + "learning_rate": 0.00015097687623840726, + "loss": 3.188145637512207, + "step": 5445 + }, + { + "epoch": 1.0470701248799232, + "grad_norm": 0.7838398814201355, + "learning_rate": 0.00015073861597756144, + "loss": 3.1877973556518553, + "step": 5450 + }, + { + "epoch": 1.048030739673391, + "grad_norm": 0.6628211140632629, + "learning_rate": 0.00015050035385311182, + "loss": 3.1896636962890623, + "step": 5455 + }, + { + "epoch": 1.0489913544668588, + "grad_norm": 0.737322211265564, + "learning_rate": 0.00015026209046621803, + "loss": 3.1869300842285155, + "step": 5460 + }, + { + "epoch": 1.0499519692603265, + "grad_norm": 1.0209139585494995, + "learning_rate": 0.0001500238264180431, + "loss": 3.187608528137207, + "step": 5465 + }, + { + "epoch": 1.0509125840537945, + "grad_norm": 0.7242445945739746, + "learning_rate": 0.00014978556230975173, + "loss": 3.188776969909668, + "step": 5470 + }, + { + "epoch": 1.0518731988472623, + "grad_norm": 0.968614399433136, + "learning_rate": 0.0001495472987425086, + "loss": 3.1897449493408203, + "step": 5475 + }, + { + "epoch": 1.05283381364073, + "grad_norm": 1.0760611295700073, + "learning_rate": 0.00014930903631747724, + "loss": 3.1843488693237303, + "step": 5480 + }, + { + "epoch": 1.0537944284341978, + "grad_norm": 0.8381465673446655, + "learning_rate": 0.00014907077563581809, + "loss": 3.189168930053711, + "step": 5485 + }, + { + "epoch": 1.0547550432276658, + "grad_norm": 0.5522369742393494, + "learning_rate": 0.00014883251729868737, + "loss": 3.1856075286865235, + "step": 5490 + }, + { + "epoch": 1.0557156580211335, + "grad_norm": 0.8688974380493164, + "learning_rate": 0.00014859426190723523, + "loss": 3.1877534866333006, + "step": 5495 + }, + { + "epoch": 1.0566762728146013, + "grad_norm": 1.0312248468399048, + "learning_rate": 0.00014835601006260455, + "loss": 3.187721824645996, + "step": 5500 + }, + { + "epoch": 1.057636887608069, + "grad_norm": 1.0165605545043945, + "learning_rate": 0.00014811776236592914, + "loss": 3.1891069412231445, + "step": 5505 + }, + { + "epoch": 1.058597502401537, + "grad_norm": 0.6776289939880371, + "learning_rate": 0.00014787951941833236, + "loss": 3.188289451599121, + "step": 5510 + }, + { + "epoch": 1.0595581171950048, + "grad_norm": 0.6294692158699036, + "learning_rate": 0.00014764128182092557, + "loss": 3.191664123535156, + "step": 5515 + }, + { + "epoch": 1.0605187319884726, + "grad_norm": 0.8317521810531616, + "learning_rate": 0.00014740305017480674, + "loss": 3.1893966674804686, + "step": 5520 + }, + { + "epoch": 1.0614793467819403, + "grad_norm": 0.9894767999649048, + "learning_rate": 0.0001471648250810588, + "loss": 3.1885601043701173, + "step": 5525 + }, + { + "epoch": 1.0624399615754083, + "grad_norm": 0.7484564185142517, + "learning_rate": 0.00014692660714074796, + "loss": 3.18869571685791, + "step": 5530 + }, + { + "epoch": 1.063400576368876, + "grad_norm": 0.8479475975036621, + "learning_rate": 0.00014668839695492264, + "loss": 3.184678649902344, + "step": 5535 + }, + { + "epoch": 1.0643611911623438, + "grad_norm": 0.6609562039375305, + "learning_rate": 0.00014645019512461157, + "loss": 3.1846710205078126, + "step": 5540 + }, + { + "epoch": 1.0653218059558118, + "grad_norm": 0.955204963684082, + "learning_rate": 0.0001462120022508223, + "loss": 3.19210205078125, + "step": 5545 + }, + { + "epoch": 1.0662824207492796, + "grad_norm": 0.7516992092132568, + "learning_rate": 0.00014597381893453998, + "loss": 3.188296318054199, + "step": 5550 + }, + { + "epoch": 1.0672430355427474, + "grad_norm": 0.7409364581108093, + "learning_rate": 0.0001457356457767255, + "loss": 3.1862789154052735, + "step": 5555 + }, + { + "epoch": 1.0682036503362151, + "grad_norm": 0.6613131165504456, + "learning_rate": 0.00014549748337831423, + "loss": 3.185582733154297, + "step": 5560 + }, + { + "epoch": 1.069164265129683, + "grad_norm": 0.6710831522941589, + "learning_rate": 0.00014525933234021428, + "loss": 3.1856903076171874, + "step": 5565 + }, + { + "epoch": 1.0701248799231509, + "grad_norm": 0.6263184547424316, + "learning_rate": 0.00014502119326330516, + "loss": 3.186991310119629, + "step": 5570 + }, + { + "epoch": 1.0710854947166186, + "grad_norm": 0.838955283164978, + "learning_rate": 0.0001447830667484362, + "loss": 3.1853305816650392, + "step": 5575 + }, + { + "epoch": 1.0720461095100864, + "grad_norm": 0.6884387135505676, + "learning_rate": 0.00014454495339642504, + "loss": 3.189616584777832, + "step": 5580 + }, + { + "epoch": 1.0730067243035544, + "grad_norm": 0.8635831475257874, + "learning_rate": 0.00014430685380805605, + "loss": 3.185613822937012, + "step": 5585 + }, + { + "epoch": 1.0739673390970221, + "grad_norm": 0.842369794845581, + "learning_rate": 0.00014406876858407894, + "loss": 3.1873950958251953, + "step": 5590 + }, + { + "epoch": 1.07492795389049, + "grad_norm": 0.6380950212478638, + "learning_rate": 0.00014383069832520721, + "loss": 3.1872686386108398, + "step": 5595 + }, + { + "epoch": 1.0758885686839577, + "grad_norm": 0.6167466044425964, + "learning_rate": 0.00014359264363211647, + "loss": 3.1889183044433596, + "step": 5600 + }, + { + "epoch": 1.0768491834774256, + "grad_norm": 0.8871692419052124, + "learning_rate": 0.00014335460510544318, + "loss": 3.190326118469238, + "step": 5605 + }, + { + "epoch": 1.0778097982708934, + "grad_norm": 0.8428598642349243, + "learning_rate": 0.00014311658334578296, + "loss": 3.190188407897949, + "step": 5610 + }, + { + "epoch": 1.0787704130643612, + "grad_norm": 0.7600328922271729, + "learning_rate": 0.00014287857895368906, + "loss": 3.1856243133544924, + "step": 5615 + }, + { + "epoch": 1.079731027857829, + "grad_norm": 0.7061536908149719, + "learning_rate": 0.00014264059252967108, + "loss": 3.1858592987060548, + "step": 5620 + }, + { + "epoch": 1.080691642651297, + "grad_norm": 0.6776458024978638, + "learning_rate": 0.0001424026246741931, + "loss": 3.188393402099609, + "step": 5625 + }, + { + "epoch": 1.0816522574447647, + "grad_norm": 0.9490513205528259, + "learning_rate": 0.00014216467598767248, + "loss": 3.1876445770263673, + "step": 5630 + }, + { + "epoch": 1.0826128722382324, + "grad_norm": 0.8668431043624878, + "learning_rate": 0.0001419267470704781, + "loss": 3.1881153106689455, + "step": 5635 + }, + { + "epoch": 1.0835734870317002, + "grad_norm": 0.7393624782562256, + "learning_rate": 0.00014168883852292915, + "loss": 3.186983108520508, + "step": 5640 + }, + { + "epoch": 1.0845341018251682, + "grad_norm": 0.7788834571838379, + "learning_rate": 0.00014145095094529318, + "loss": 3.187802314758301, + "step": 5645 + }, + { + "epoch": 1.085494716618636, + "grad_norm": 0.6699324250221252, + "learning_rate": 0.00014121308493778503, + "loss": 3.1878355026245115, + "step": 5650 + }, + { + "epoch": 1.0864553314121037, + "grad_norm": 0.7733721733093262, + "learning_rate": 0.00014097524110056494, + "loss": 3.1843936920166014, + "step": 5655 + }, + { + "epoch": 1.0874159462055715, + "grad_norm": 0.711729884147644, + "learning_rate": 0.0001407374200337374, + "loss": 3.1883363723754883, + "step": 5660 + }, + { + "epoch": 1.0883765609990395, + "grad_norm": 0.6013932824134827, + "learning_rate": 0.0001404996223373494, + "loss": 3.1905609130859376, + "step": 5665 + }, + { + "epoch": 1.0893371757925072, + "grad_norm": 0.7252037525177002, + "learning_rate": 0.0001402618486113888, + "loss": 3.1879077911376954, + "step": 5670 + }, + { + "epoch": 1.090297790585975, + "grad_norm": 1.1090692281723022, + "learning_rate": 0.0001400240994557832, + "loss": 3.1889415740966798, + "step": 5675 + }, + { + "epoch": 1.0912584053794427, + "grad_norm": 1.0110430717468262, + "learning_rate": 0.0001397863754703981, + "loss": 3.185939407348633, + "step": 5680 + }, + { + "epoch": 1.0922190201729107, + "grad_norm": 0.8803393840789795, + "learning_rate": 0.00013954867725503542, + "loss": 3.1873733520507814, + "step": 5685 + }, + { + "epoch": 1.0931796349663785, + "grad_norm": 0.6381970047950745, + "learning_rate": 0.00013931100540943227, + "loss": 3.1853010177612306, + "step": 5690 + }, + { + "epoch": 1.0941402497598463, + "grad_norm": 0.5555366277694702, + "learning_rate": 0.00013907336053325896, + "loss": 3.184972381591797, + "step": 5695 + }, + { + "epoch": 1.0951008645533142, + "grad_norm": 1.0896075963974, + "learning_rate": 0.00013883574322611801, + "loss": 3.186721992492676, + "step": 5700 + }, + { + "epoch": 1.096061479346782, + "grad_norm": 0.8639760613441467, + "learning_rate": 0.00013859815408754218, + "loss": 3.1882835388183595, + "step": 5705 + }, + { + "epoch": 1.0970220941402498, + "grad_norm": 0.7950151562690735, + "learning_rate": 0.00013836059371699332, + "loss": 3.187421417236328, + "step": 5710 + }, + { + "epoch": 1.0979827089337175, + "grad_norm": 0.6552711725234985, + "learning_rate": 0.00013812306271386048, + "loss": 3.184224319458008, + "step": 5715 + }, + { + "epoch": 1.0989433237271853, + "grad_norm": 0.7803718447685242, + "learning_rate": 0.00013788556167745886, + "loss": 3.1838443756103514, + "step": 5720 + }, + { + "epoch": 1.0999039385206533, + "grad_norm": 0.8713034987449646, + "learning_rate": 0.00013764809120702797, + "loss": 3.188005828857422, + "step": 5725 + }, + { + "epoch": 1.100864553314121, + "grad_norm": 0.638888955116272, + "learning_rate": 0.00013741065190173008, + "loss": 3.183907890319824, + "step": 5730 + }, + { + "epoch": 1.1018251681075888, + "grad_norm": 0.8170698285102844, + "learning_rate": 0.00013717324436064897, + "loss": 3.184910011291504, + "step": 5735 + }, + { + "epoch": 1.1027857829010568, + "grad_norm": 0.8812254667282104, + "learning_rate": 0.0001369358691827882, + "loss": 3.1871992111206056, + "step": 5740 + }, + { + "epoch": 1.1037463976945245, + "grad_norm": 0.6346907615661621, + "learning_rate": 0.0001366985269670697, + "loss": 3.1861663818359376, + "step": 5745 + }, + { + "epoch": 1.1047070124879923, + "grad_norm": 0.930957019329071, + "learning_rate": 0.00013646121831233226, + "loss": 3.1821567535400392, + "step": 5750 + }, + { + "epoch": 1.10566762728146, + "grad_norm": 0.7975543141365051, + "learning_rate": 0.00013622394381732986, + "loss": 3.1874359130859373, + "step": 5755 + }, + { + "epoch": 1.106628242074928, + "grad_norm": 0.6229592561721802, + "learning_rate": 0.0001359867040807305, + "loss": 3.1849109649658205, + "step": 5760 + }, + { + "epoch": 1.1075888568683958, + "grad_norm": 0.7393524050712585, + "learning_rate": 0.0001357494997011143, + "loss": 3.1865949630737305, + "step": 5765 + }, + { + "epoch": 1.1085494716618636, + "grad_norm": 0.7825970649719238, + "learning_rate": 0.0001355123312769723, + "loss": 3.179779815673828, + "step": 5770 + }, + { + "epoch": 1.1095100864553313, + "grad_norm": 1.018021821975708, + "learning_rate": 0.00013527519940670474, + "loss": 3.1880813598632813, + "step": 5775 + }, + { + "epoch": 1.1104707012487993, + "grad_norm": 0.7721784114837646, + "learning_rate": 0.00013503810468861967, + "loss": 3.1860368728637694, + "step": 5780 + }, + { + "epoch": 1.111431316042267, + "grad_norm": 0.7420386075973511, + "learning_rate": 0.00013480104772093136, + "loss": 3.181845855712891, + "step": 5785 + }, + { + "epoch": 1.1123919308357348, + "grad_norm": 0.7086904644966125, + "learning_rate": 0.0001345640291017588, + "loss": 3.181869125366211, + "step": 5790 + }, + { + "epoch": 1.1133525456292026, + "grad_norm": 0.5291205048561096, + "learning_rate": 0.00013432704942912445, + "loss": 3.1842918395996094, + "step": 5795 + }, + { + "epoch": 1.1143131604226706, + "grad_norm": 0.6153315901756287, + "learning_rate": 0.0001340901093009522, + "loss": 3.180573654174805, + "step": 5800 + }, + { + "epoch": 1.1152737752161384, + "grad_norm": 0.8189062476158142, + "learning_rate": 0.00013385320931506636, + "loss": 3.181578826904297, + "step": 5805 + }, + { + "epoch": 1.1162343900096061, + "grad_norm": 0.8425748348236084, + "learning_rate": 0.00013361635006918987, + "loss": 3.1840728759765624, + "step": 5810 + }, + { + "epoch": 1.1171950048030739, + "grad_norm": 0.6964288353919983, + "learning_rate": 0.00013337953216094293, + "loss": 3.183258056640625, + "step": 5815 + }, + { + "epoch": 1.1181556195965419, + "grad_norm": 0.8112155795097351, + "learning_rate": 0.0001331427561878414, + "loss": 3.1822595596313477, + "step": 5820 + }, + { + "epoch": 1.1191162343900096, + "grad_norm": 0.5789711475372314, + "learning_rate": 0.0001329060227472953, + "loss": 3.1870792388916014, + "step": 5825 + }, + { + "epoch": 1.1200768491834774, + "grad_norm": 0.7857502102851868, + "learning_rate": 0.00013266933243660748, + "loss": 3.1816539764404297, + "step": 5830 + }, + { + "epoch": 1.1210374639769451, + "grad_norm": 0.8584936857223511, + "learning_rate": 0.0001324326858529718, + "loss": 3.1840972900390625, + "step": 5835 + }, + { + "epoch": 1.1219980787704131, + "grad_norm": 0.7861515879631042, + "learning_rate": 0.00013219608359347194, + "loss": 3.1865585327148436, + "step": 5840 + }, + { + "epoch": 1.122958693563881, + "grad_norm": 0.6675422787666321, + "learning_rate": 0.0001319595262550796, + "loss": 3.1822277069091798, + "step": 5845 + }, + { + "epoch": 1.1239193083573487, + "grad_norm": 0.7415298223495483, + "learning_rate": 0.00013172301443465327, + "loss": 3.18204345703125, + "step": 5850 + }, + { + "epoch": 1.1248799231508164, + "grad_norm": 0.7668789029121399, + "learning_rate": 0.00013148654872893647, + "loss": 3.1823986053466795, + "step": 5855 + }, + { + "epoch": 1.1258405379442844, + "grad_norm": 1.0284775495529175, + "learning_rate": 0.00013125012973455645, + "loss": 3.186284637451172, + "step": 5860 + }, + { + "epoch": 1.1268011527377522, + "grad_norm": 0.9501076936721802, + "learning_rate": 0.00013101375804802268, + "loss": 3.180257034301758, + "step": 5865 + }, + { + "epoch": 1.12776176753122, + "grad_norm": 0.6620021462440491, + "learning_rate": 0.00013077743426572508, + "loss": 3.1850921630859377, + "step": 5870 + }, + { + "epoch": 1.1287223823246877, + "grad_norm": 0.6377175450325012, + "learning_rate": 0.0001305411589839328, + "loss": 3.1847787857055665, + "step": 5875 + }, + { + "epoch": 1.1296829971181557, + "grad_norm": 0.8956018686294556, + "learning_rate": 0.00013030493279879267, + "loss": 3.1820547103881838, + "step": 5880 + }, + { + "epoch": 1.1306436119116234, + "grad_norm": 0.9648638367652893, + "learning_rate": 0.00013006875630632757, + "loss": 3.1855083465576173, + "step": 5885 + }, + { + "epoch": 1.1316042267050912, + "grad_norm": 0.7749320864677429, + "learning_rate": 0.000129832630102435, + "loss": 3.18371639251709, + "step": 5890 + }, + { + "epoch": 1.1325648414985592, + "grad_norm": 0.7162895202636719, + "learning_rate": 0.00012959655478288556, + "loss": 3.1814332962036134, + "step": 5895 + }, + { + "epoch": 1.133525456292027, + "grad_norm": 0.7946479320526123, + "learning_rate": 0.00012936053094332158, + "loss": 3.183310699462891, + "step": 5900 + }, + { + "epoch": 1.1344860710854947, + "grad_norm": 0.9230923652648926, + "learning_rate": 0.00012912455917925535, + "loss": 3.1882291793823243, + "step": 5905 + }, + { + "epoch": 1.1354466858789625, + "grad_norm": 0.6934447288513184, + "learning_rate": 0.0001288886400860679, + "loss": 3.1866472244262694, + "step": 5910 + }, + { + "epoch": 1.1364073006724302, + "grad_norm": 0.5946722626686096, + "learning_rate": 0.00012865277425900724, + "loss": 3.1836269378662108, + "step": 5915 + }, + { + "epoch": 1.1373679154658982, + "grad_norm": 0.6142674088478088, + "learning_rate": 0.00012841696229318712, + "loss": 3.1850711822509767, + "step": 5920 + }, + { + "epoch": 1.138328530259366, + "grad_norm": 0.5882613062858582, + "learning_rate": 0.00012818120478358522, + "loss": 3.1808708190917967, + "step": 5925 + }, + { + "epoch": 1.1392891450528337, + "grad_norm": 1.1265987157821655, + "learning_rate": 0.000127945502325042, + "loss": 3.1862953186035154, + "step": 5930 + }, + { + "epoch": 1.1402497598463017, + "grad_norm": 0.8293631076812744, + "learning_rate": 0.00012770985551225898, + "loss": 3.1817962646484377, + "step": 5935 + }, + { + "epoch": 1.1412103746397695, + "grad_norm": 0.6100813150405884, + "learning_rate": 0.00012747426493979713, + "loss": 3.182262420654297, + "step": 5940 + }, + { + "epoch": 1.1421709894332372, + "grad_norm": 0.6388895511627197, + "learning_rate": 0.00012723873120207575, + "loss": 3.183791732788086, + "step": 5945 + }, + { + "epoch": 1.143131604226705, + "grad_norm": 1.1153887510299683, + "learning_rate": 0.00012700325489337056, + "loss": 3.1867645263671873, + "step": 5950 + }, + { + "epoch": 1.144092219020173, + "grad_norm": 0.6275951862335205, + "learning_rate": 0.00012676783660781245, + "loss": 3.183899688720703, + "step": 5955 + }, + { + "epoch": 1.1450528338136408, + "grad_norm": 0.6649777293205261, + "learning_rate": 0.00012653247693938588, + "loss": 3.1856327056884766, + "step": 5960 + }, + { + "epoch": 1.1460134486071085, + "grad_norm": 0.7300867438316345, + "learning_rate": 0.00012629717648192747, + "loss": 3.1813177108764648, + "step": 5965 + }, + { + "epoch": 1.1469740634005763, + "grad_norm": 0.8946056365966797, + "learning_rate": 0.00012606193582912446, + "loss": 3.182071304321289, + "step": 5970 + }, + { + "epoch": 1.1479346781940443, + "grad_norm": 0.8307191729545593, + "learning_rate": 0.00012582675557451306, + "loss": 3.1808849334716798, + "step": 5975 + }, + { + "epoch": 1.148895292987512, + "grad_norm": 0.8783102631568909, + "learning_rate": 0.00012559163631147723, + "loss": 3.184329032897949, + "step": 5980 + }, + { + "epoch": 1.1498559077809798, + "grad_norm": 0.7081605792045593, + "learning_rate": 0.00012535657863324695, + "loss": 3.1825115203857424, + "step": 5985 + }, + { + "epoch": 1.1508165225744476, + "grad_norm": 0.5934545993804932, + "learning_rate": 0.0001251215831328969, + "loss": 3.1839290618896485, + "step": 5990 + }, + { + "epoch": 1.1517771373679155, + "grad_norm": 0.8041390180587769, + "learning_rate": 0.00012488665040334479, + "loss": 3.183102607727051, + "step": 5995 + }, + { + "epoch": 1.1527377521613833, + "grad_norm": 0.6526947617530823, + "learning_rate": 0.00012465178103735, + "loss": 3.1817670822143556, + "step": 6000 + }, + { + "epoch": 1.153698366954851, + "grad_norm": 0.6892653703689575, + "learning_rate": 0.00012441697562751215, + "loss": 3.180023193359375, + "step": 6005 + }, + { + "epoch": 1.154658981748319, + "grad_norm": 0.6279175877571106, + "learning_rate": 0.00012418223476626925, + "loss": 3.180769157409668, + "step": 6010 + }, + { + "epoch": 1.1556195965417868, + "grad_norm": 0.8991310596466064, + "learning_rate": 0.00012394755904589664, + "loss": 3.185033416748047, + "step": 6015 + }, + { + "epoch": 1.1565802113352546, + "grad_norm": 0.6720035672187805, + "learning_rate": 0.00012371294905850518, + "loss": 3.1769649505615236, + "step": 6020 + }, + { + "epoch": 1.1575408261287223, + "grad_norm": 0.8395649790763855, + "learning_rate": 0.00012347840539603998, + "loss": 3.1827545166015625, + "step": 6025 + }, + { + "epoch": 1.15850144092219, + "grad_norm": 0.8189507722854614, + "learning_rate": 0.00012324392865027873, + "loss": 3.1804759979248045, + "step": 6030 + }, + { + "epoch": 1.159462055715658, + "grad_norm": 0.9116944074630737, + "learning_rate": 0.00012300951941283036, + "loss": 3.185091018676758, + "step": 6035 + }, + { + "epoch": 1.1604226705091258, + "grad_norm": 0.6200742125511169, + "learning_rate": 0.00012277517827513341, + "loss": 3.1775371551513674, + "step": 6040 + }, + { + "epoch": 1.1613832853025936, + "grad_norm": 0.6865299344062805, + "learning_rate": 0.00012254090582845457, + "loss": 3.1813201904296875, + "step": 6045 + }, + { + "epoch": 1.1623439000960616, + "grad_norm": 0.7450284957885742, + "learning_rate": 0.00012230670266388728, + "loss": 3.18134765625, + "step": 6050 + }, + { + "epoch": 1.1633045148895294, + "grad_norm": 0.814132571220398, + "learning_rate": 0.00012207256937235014, + "loss": 3.182879638671875, + "step": 6055 + }, + { + "epoch": 1.1642651296829971, + "grad_norm": 0.8822999000549316, + "learning_rate": 0.00012183850654458553, + "loss": 3.1848114013671873, + "step": 6060 + }, + { + "epoch": 1.1652257444764649, + "grad_norm": 0.5908654928207397, + "learning_rate": 0.00012160451477115783, + "loss": 3.1844045639038088, + "step": 6065 + }, + { + "epoch": 1.1661863592699326, + "grad_norm": 0.5196489095687866, + "learning_rate": 0.00012137059464245242, + "loss": 3.181051254272461, + "step": 6070 + }, + { + "epoch": 1.1671469740634006, + "grad_norm": 0.7493326663970947, + "learning_rate": 0.00012113674674867379, + "loss": 3.180977630615234, + "step": 6075 + }, + { + "epoch": 1.1681075888568684, + "grad_norm": 0.8588430881500244, + "learning_rate": 0.00012090297167984409, + "loss": 3.183562088012695, + "step": 6080 + }, + { + "epoch": 1.1690682036503361, + "grad_norm": 0.707949697971344, + "learning_rate": 0.00012066927002580185, + "loss": 3.186185836791992, + "step": 6085 + }, + { + "epoch": 1.1700288184438041, + "grad_norm": 0.7008398771286011, + "learning_rate": 0.00012043564237620028, + "loss": 3.183705520629883, + "step": 6090 + }, + { + "epoch": 1.170989433237272, + "grad_norm": 0.5148990750312805, + "learning_rate": 0.00012020208932050595, + "loss": 3.180450439453125, + "step": 6095 + }, + { + "epoch": 1.1719500480307397, + "grad_norm": 1.0000126361846924, + "learning_rate": 0.00011996861144799712, + "loss": 3.1828609466552735, + "step": 6100 + }, + { + "epoch": 1.1729106628242074, + "grad_norm": 0.9941655993461609, + "learning_rate": 0.0001197352093477625, + "loss": 3.1851844787597656, + "step": 6105 + }, + { + "epoch": 1.1738712776176754, + "grad_norm": 0.6262122988700867, + "learning_rate": 0.00011950188360869947, + "loss": 3.186477279663086, + "step": 6110 + }, + { + "epoch": 1.1748318924111432, + "grad_norm": 0.7170460224151611, + "learning_rate": 0.00011926863481951279, + "loss": 3.181361770629883, + "step": 6115 + }, + { + "epoch": 1.175792507204611, + "grad_norm": 0.5651366114616394, + "learning_rate": 0.00011903546356871315, + "loss": 3.180009460449219, + "step": 6120 + }, + { + "epoch": 1.1767531219980787, + "grad_norm": 0.7348089814186096, + "learning_rate": 0.00011880237044461546, + "loss": 3.1843181610107423, + "step": 6125 + }, + { + "epoch": 1.1777137367915467, + "grad_norm": 0.8490554690361023, + "learning_rate": 0.00011856935603533759, + "loss": 3.1855024337768554, + "step": 6130 + }, + { + "epoch": 1.1786743515850144, + "grad_norm": 0.688126802444458, + "learning_rate": 0.0001183364209287989, + "loss": 3.1825450897216796, + "step": 6135 + }, + { + "epoch": 1.1796349663784822, + "grad_norm": 0.6043757796287537, + "learning_rate": 0.00011810356571271847, + "loss": 3.1808750152587892, + "step": 6140 + }, + { + "epoch": 1.18059558117195, + "grad_norm": 0.6653470396995544, + "learning_rate": 0.00011787079097461394, + "loss": 3.185806655883789, + "step": 6145 + }, + { + "epoch": 1.181556195965418, + "grad_norm": 0.5727336406707764, + "learning_rate": 0.0001176380973017998, + "loss": 3.182196617126465, + "step": 6150 + }, + { + "epoch": 1.1825168107588857, + "grad_norm": 0.4174487292766571, + "learning_rate": 0.00011740548528138613, + "loss": 3.1803747177124024, + "step": 6155 + }, + { + "epoch": 1.1834774255523535, + "grad_norm": 0.8878217339515686, + "learning_rate": 0.00011717295550027685, + "loss": 3.1840145111083986, + "step": 6160 + }, + { + "epoch": 1.1844380403458212, + "grad_norm": 0.8910025954246521, + "learning_rate": 0.00011694050854516853, + "loss": 3.181249237060547, + "step": 6165 + }, + { + "epoch": 1.1853986551392892, + "grad_norm": 0.7205166220664978, + "learning_rate": 0.00011670814500254858, + "loss": 3.1778148651123046, + "step": 6170 + }, + { + "epoch": 1.186359269932757, + "grad_norm": 0.6521825790405273, + "learning_rate": 0.00011647586545869418, + "loss": 3.17816276550293, + "step": 6175 + }, + { + "epoch": 1.1873198847262247, + "grad_norm": 0.6897614002227783, + "learning_rate": 0.00011624367049967037, + "loss": 3.1780323028564452, + "step": 6180 + }, + { + "epoch": 1.1882804995196925, + "grad_norm": 0.5899947881698608, + "learning_rate": 0.00011601156071132883, + "loss": 3.181020164489746, + "step": 6185 + }, + { + "epoch": 1.1892411143131605, + "grad_norm": 0.6569263935089111, + "learning_rate": 0.00011577953667930641, + "loss": 3.1823570251464846, + "step": 6190 + }, + { + "epoch": 1.1902017291066282, + "grad_norm": 0.6824354529380798, + "learning_rate": 0.00011554759898902355, + "loss": 3.1782264709472656, + "step": 6195 + }, + { + "epoch": 1.191162343900096, + "grad_norm": 0.5225342512130737, + "learning_rate": 0.00011531574822568278, + "loss": 3.1797101974487303, + "step": 6200 + }, + { + "epoch": 1.192122958693564, + "grad_norm": 0.5605508685112, + "learning_rate": 0.00011508398497426745, + "loss": 3.1792762756347654, + "step": 6205 + }, + { + "epoch": 1.1930835734870318, + "grad_norm": 0.8098820447921753, + "learning_rate": 0.00011485230981953997, + "loss": 3.183140182495117, + "step": 6210 + }, + { + "epoch": 1.1940441882804995, + "grad_norm": 0.6241822838783264, + "learning_rate": 0.00011462072334604053, + "loss": 3.1846290588378907, + "step": 6215 + }, + { + "epoch": 1.1950048030739673, + "grad_norm": 0.6163104772567749, + "learning_rate": 0.00011438922613808554, + "loss": 3.184892272949219, + "step": 6220 + }, + { + "epoch": 1.195965417867435, + "grad_norm": 0.6423901915550232, + "learning_rate": 0.00011415781877976628, + "loss": 3.180209922790527, + "step": 6225 + }, + { + "epoch": 1.196926032660903, + "grad_norm": 0.743615448474884, + "learning_rate": 0.00011392650185494712, + "loss": 3.181998634338379, + "step": 6230 + }, + { + "epoch": 1.1978866474543708, + "grad_norm": 0.7482926845550537, + "learning_rate": 0.00011369527594726452, + "loss": 3.178305244445801, + "step": 6235 + }, + { + "epoch": 1.1988472622478386, + "grad_norm": 0.6290075778961182, + "learning_rate": 0.0001134641416401251, + "loss": 3.182933807373047, + "step": 6240 + }, + { + "epoch": 1.1998078770413065, + "grad_norm": 0.5498515963554382, + "learning_rate": 0.00011323309951670444, + "loss": 3.1798477172851562, + "step": 6245 + }, + { + "epoch": 1.2007684918347743, + "grad_norm": 0.6985566020011902, + "learning_rate": 0.00011300215015994554, + "loss": 3.1814619064331056, + "step": 6250 + }, + { + "epoch": 1.201729106628242, + "grad_norm": 0.8303685784339905, + "learning_rate": 0.00011277129415255727, + "loss": 3.1828535079956053, + "step": 6255 + }, + { + "epoch": 1.2026897214217098, + "grad_norm": 0.7280902862548828, + "learning_rate": 0.00011254053207701308, + "loss": 3.17972354888916, + "step": 6260 + }, + { + "epoch": 1.2036503362151778, + "grad_norm": 0.6531400084495544, + "learning_rate": 0.00011230986451554932, + "loss": 3.1785831451416016, + "step": 6265 + }, + { + "epoch": 1.2046109510086456, + "grad_norm": 0.6152121424674988, + "learning_rate": 0.00011207929205016386, + "loss": 3.177712631225586, + "step": 6270 + }, + { + "epoch": 1.2055715658021133, + "grad_norm": 0.5569039583206177, + "learning_rate": 0.00011184881526261483, + "loss": 3.1826234817504884, + "step": 6275 + }, + { + "epoch": 1.206532180595581, + "grad_norm": 1.1272635459899902, + "learning_rate": 0.00011161843473441867, + "loss": 3.179414749145508, + "step": 6280 + }, + { + "epoch": 1.207492795389049, + "grad_norm": 0.6206833720207214, + "learning_rate": 0.0001113881510468492, + "loss": 3.184890365600586, + "step": 6285 + }, + { + "epoch": 1.2084534101825168, + "grad_norm": 0.6447250247001648, + "learning_rate": 0.00011115796478093569, + "loss": 3.1766143798828126, + "step": 6290 + }, + { + "epoch": 1.2094140249759846, + "grad_norm": 0.7120223641395569, + "learning_rate": 0.00011092787651746178, + "loss": 3.1796751022338867, + "step": 6295 + }, + { + "epoch": 1.2103746397694524, + "grad_norm": 0.620508074760437, + "learning_rate": 0.00011069788683696366, + "loss": 3.1746284484863283, + "step": 6300 + }, + { + "epoch": 1.2113352545629203, + "grad_norm": 0.6917296051979065, + "learning_rate": 0.00011046799631972903, + "loss": 3.1808353424072267, + "step": 6305 + }, + { + "epoch": 1.2122958693563881, + "grad_norm": 0.7654911875724792, + "learning_rate": 0.00011023820554579512, + "loss": 3.1830776214599608, + "step": 6310 + }, + { + "epoch": 1.2132564841498559, + "grad_norm": 0.7507694959640503, + "learning_rate": 0.00011000851509494771, + "loss": 3.181525230407715, + "step": 6315 + }, + { + "epoch": 1.2142170989433236, + "grad_norm": 0.6086501479148865, + "learning_rate": 0.00010977892554671935, + "loss": 3.1778955459594727, + "step": 6320 + }, + { + "epoch": 1.2151777137367916, + "grad_norm": 0.5555998682975769, + "learning_rate": 0.00010954943748038798, + "loss": 3.1771930694580077, + "step": 6325 + }, + { + "epoch": 1.2161383285302594, + "grad_norm": 0.6753772497177124, + "learning_rate": 0.00010932005147497559, + "loss": 3.178631591796875, + "step": 6330 + }, + { + "epoch": 1.2170989433237271, + "grad_norm": 0.5674150586128235, + "learning_rate": 0.00010909076810924654, + "loss": 3.183444786071777, + "step": 6335 + }, + { + "epoch": 1.218059558117195, + "grad_norm": 0.8342090249061584, + "learning_rate": 0.00010886158796170633, + "loss": 3.181511116027832, + "step": 6340 + }, + { + "epoch": 1.219020172910663, + "grad_norm": 0.9122212529182434, + "learning_rate": 0.00010863251161060002, + "loss": 3.1852695465087892, + "step": 6345 + }, + { + "epoch": 1.2199807877041307, + "grad_norm": 0.44202694296836853, + "learning_rate": 0.00010840353963391065, + "loss": 3.1827775955200197, + "step": 6350 + }, + { + "epoch": 1.2209414024975984, + "grad_norm": 0.5556421279907227, + "learning_rate": 0.0001081746726093581, + "loss": 3.176057815551758, + "step": 6355 + }, + { + "epoch": 1.2219020172910664, + "grad_norm": 0.5319374203681946, + "learning_rate": 0.00010794591111439729, + "loss": 3.177410125732422, + "step": 6360 + }, + { + "epoch": 1.2228626320845342, + "grad_norm": 0.7626725435256958, + "learning_rate": 0.000107717255726217, + "loss": 3.182592582702637, + "step": 6365 + }, + { + "epoch": 1.223823246878002, + "grad_norm": 0.7391705513000488, + "learning_rate": 0.00010748870702173815, + "loss": 3.178932952880859, + "step": 6370 + }, + { + "epoch": 1.2247838616714697, + "grad_norm": 0.4964456260204315, + "learning_rate": 0.00010726026557761264, + "loss": 3.1790546417236327, + "step": 6375 + }, + { + "epoch": 1.2257444764649374, + "grad_norm": 0.5090710520744324, + "learning_rate": 0.00010703193197022159, + "loss": 3.1741233825683595, + "step": 6380 + }, + { + "epoch": 1.2267050912584054, + "grad_norm": 0.8102352023124695, + "learning_rate": 0.00010680370677567423, + "loss": 3.184102249145508, + "step": 6385 + }, + { + "epoch": 1.2276657060518732, + "grad_norm": 0.7811996340751648, + "learning_rate": 0.00010657559056980603, + "loss": 3.1772319793701174, + "step": 6390 + }, + { + "epoch": 1.228626320845341, + "grad_norm": 0.7701186537742615, + "learning_rate": 0.00010634758392817762, + "loss": 3.1799732208251954, + "step": 6395 + }, + { + "epoch": 1.229586935638809, + "grad_norm": 0.685007631778717, + "learning_rate": 0.00010611968742607318, + "loss": 3.176517105102539, + "step": 6400 + }, + { + "epoch": 1.2305475504322767, + "grad_norm": 0.5316605567932129, + "learning_rate": 0.00010589190163849885, + "loss": 3.1800348281860353, + "step": 6405 + }, + { + "epoch": 1.2315081652257445, + "grad_norm": 0.6539739370346069, + "learning_rate": 0.00010566422714018167, + "loss": 3.1817600250244142, + "step": 6410 + }, + { + "epoch": 1.2324687800192122, + "grad_norm": 0.5445916056632996, + "learning_rate": 0.00010543666450556774, + "loss": 3.1751848220825196, + "step": 6415 + }, + { + "epoch": 1.23342939481268, + "grad_norm": 0.6994644403457642, + "learning_rate": 0.00010520921430882088, + "loss": 3.182989311218262, + "step": 6420 + }, + { + "epoch": 1.234390009606148, + "grad_norm": 0.6963391304016113, + "learning_rate": 0.00010498187712382132, + "loss": 3.1813974380493164, + "step": 6425 + }, + { + "epoch": 1.2353506243996157, + "grad_norm": 0.8707190752029419, + "learning_rate": 0.00010475465352416402, + "loss": 3.1847021102905275, + "step": 6430 + }, + { + "epoch": 1.2363112391930835, + "grad_norm": 0.628430962562561, + "learning_rate": 0.00010452754408315754, + "loss": 3.180718994140625, + "step": 6435 + }, + { + "epoch": 1.2372718539865515, + "grad_norm": 0.4727374017238617, + "learning_rate": 0.00010430054937382219, + "loss": 3.179873466491699, + "step": 6440 + }, + { + "epoch": 1.2382324687800192, + "grad_norm": 0.6456632614135742, + "learning_rate": 0.000104073669968889, + "loss": 3.1764720916748046, + "step": 6445 + }, + { + "epoch": 1.239193083573487, + "grad_norm": 0.9211506843566895, + "learning_rate": 0.0001038469064407979, + "loss": 3.175563430786133, + "step": 6450 + }, + { + "epoch": 1.2401536983669548, + "grad_norm": 0.7014700770378113, + "learning_rate": 0.00010362025936169664, + "loss": 3.1777694702148436, + "step": 6455 + }, + { + "epoch": 1.2411143131604228, + "grad_norm": 0.612933874130249, + "learning_rate": 0.00010339372930343896, + "loss": 3.177769088745117, + "step": 6460 + }, + { + "epoch": 1.2420749279538905, + "grad_norm": 0.47281619906425476, + "learning_rate": 0.00010316731683758346, + "loss": 3.177565574645996, + "step": 6465 + }, + { + "epoch": 1.2430355427473583, + "grad_norm": 0.6775256395339966, + "learning_rate": 0.00010294102253539204, + "loss": 3.1776294708251953, + "step": 6470 + }, + { + "epoch": 1.243996157540826, + "grad_norm": 0.7892163395881653, + "learning_rate": 0.00010271484696782837, + "loss": 3.175709342956543, + "step": 6475 + }, + { + "epoch": 1.244956772334294, + "grad_norm": 0.6570128202438354, + "learning_rate": 0.00010248879070555668, + "loss": 3.1776430130004885, + "step": 6480 + }, + { + "epoch": 1.2459173871277618, + "grad_norm": 0.5200539231300354, + "learning_rate": 0.00010226285431894013, + "loss": 3.176724433898926, + "step": 6485 + }, + { + "epoch": 1.2468780019212296, + "grad_norm": 0.7223069667816162, + "learning_rate": 0.00010203703837803934, + "loss": 3.176532745361328, + "step": 6490 + }, + { + "epoch": 1.2478386167146973, + "grad_norm": 0.8052407503128052, + "learning_rate": 0.00010181134345261115, + "loss": 3.176080322265625, + "step": 6495 + }, + { + "epoch": 1.2487992315081653, + "grad_norm": 0.48731186985969543, + "learning_rate": 0.00010158577011210695, + "loss": 3.173239517211914, + "step": 6500 + }, + { + "epoch": 1.249759846301633, + "grad_norm": 0.48055845499038696, + "learning_rate": 0.00010136031892567145, + "loss": 3.1757953643798826, + "step": 6505 + }, + { + "epoch": 1.2507204610951008, + "grad_norm": 0.9334460496902466, + "learning_rate": 0.00010113499046214113, + "loss": 3.173488998413086, + "step": 6510 + }, + { + "epoch": 1.2516810758885688, + "grad_norm": 0.747382402420044, + "learning_rate": 0.00010090978529004284, + "loss": 3.1791568756103517, + "step": 6515 + }, + { + "epoch": 1.2526416906820366, + "grad_norm": 0.5684623718261719, + "learning_rate": 0.00010068470397759226, + "loss": 3.1786075592041017, + "step": 6520 + }, + { + "epoch": 1.2536023054755043, + "grad_norm": 0.7101691365242004, + "learning_rate": 0.0001004597470926927, + "loss": 3.17834587097168, + "step": 6525 + }, + { + "epoch": 1.254562920268972, + "grad_norm": 0.6113804578781128, + "learning_rate": 0.0001002349152029334, + "loss": 3.1760730743408203, + "step": 6530 + }, + { + "epoch": 1.2555235350624399, + "grad_norm": 0.634519100189209, + "learning_rate": 0.00010001020887558837, + "loss": 3.1759071350097656, + "step": 6535 + }, + { + "epoch": 1.2564841498559078, + "grad_norm": 0.6819270253181458, + "learning_rate": 9.978562867761465e-05, + "loss": 3.1800254821777343, + "step": 6540 + }, + { + "epoch": 1.2574447646493756, + "grad_norm": 0.6920965313911438, + "learning_rate": 9.95611751756511e-05, + "loss": 3.179019546508789, + "step": 6545 + }, + { + "epoch": 1.2584053794428434, + "grad_norm": 0.7164304256439209, + "learning_rate": 9.9336848936017e-05, + "loss": 3.174900436401367, + "step": 6550 + }, + { + "epoch": 1.2593659942363113, + "grad_norm": 0.5404880046844482, + "learning_rate": 9.911265052471046e-05, + "loss": 3.1754596710205076, + "step": 6555 + }, + { + "epoch": 1.260326609029779, + "grad_norm": 0.6277545690536499, + "learning_rate": 9.888858050740703e-05, + "loss": 3.1767545700073243, + "step": 6560 + }, + { + "epoch": 1.2612872238232469, + "grad_norm": 0.4282405972480774, + "learning_rate": 9.866463944945839e-05, + "loss": 3.175259590148926, + "step": 6565 + }, + { + "epoch": 1.2622478386167146, + "grad_norm": 0.740606963634491, + "learning_rate": 9.844082791589075e-05, + "loss": 3.1768466949462892, + "step": 6570 + }, + { + "epoch": 1.2632084534101824, + "grad_norm": 0.6880332231521606, + "learning_rate": 9.821714647140363e-05, + "loss": 3.1770793914794924, + "step": 6575 + }, + { + "epoch": 1.2641690682036504, + "grad_norm": 0.5514328479766846, + "learning_rate": 9.799359568036821e-05, + "loss": 3.180289649963379, + "step": 6580 + }, + { + "epoch": 1.2651296829971181, + "grad_norm": 0.569648027420044, + "learning_rate": 9.777017610682612e-05, + "loss": 3.1760627746582033, + "step": 6585 + }, + { + "epoch": 1.266090297790586, + "grad_norm": 0.6740429401397705, + "learning_rate": 9.754688831448781e-05, + "loss": 3.173795700073242, + "step": 6590 + }, + { + "epoch": 1.267050912584054, + "grad_norm": 0.6410177946090698, + "learning_rate": 9.732373286673135e-05, + "loss": 3.178542709350586, + "step": 6595 + }, + { + "epoch": 1.2680115273775217, + "grad_norm": 0.5730255246162415, + "learning_rate": 9.710071032660076e-05, + "loss": 3.1776718139648437, + "step": 6600 + }, + { + "epoch": 1.2689721421709894, + "grad_norm": 0.6753981113433838, + "learning_rate": 9.687782125680484e-05, + "loss": 3.1791826248168946, + "step": 6605 + }, + { + "epoch": 1.2699327569644572, + "grad_norm": 0.5752228498458862, + "learning_rate": 9.66550662197155e-05, + "loss": 3.173518753051758, + "step": 6610 + }, + { + "epoch": 1.270893371757925, + "grad_norm": 0.5868836045265198, + "learning_rate": 9.643244577736667e-05, + "loss": 3.173557090759277, + "step": 6615 + }, + { + "epoch": 1.271853986551393, + "grad_norm": 0.5383107662200928, + "learning_rate": 9.620996049145247e-05, + "loss": 3.1773754119873048, + "step": 6620 + }, + { + "epoch": 1.2728146013448607, + "grad_norm": 0.5838056802749634, + "learning_rate": 9.598761092332616e-05, + "loss": 3.1778125762939453, + "step": 6625 + }, + { + "epoch": 1.2737752161383284, + "grad_norm": 0.5867598056793213, + "learning_rate": 9.576539763399847e-05, + "loss": 3.183711814880371, + "step": 6630 + }, + { + "epoch": 1.2747358309317964, + "grad_norm": 0.6111621856689453, + "learning_rate": 9.554332118413635e-05, + "loss": 3.1793869018554686, + "step": 6635 + }, + { + "epoch": 1.2756964457252642, + "grad_norm": 0.8043511509895325, + "learning_rate": 9.532138213406143e-05, + "loss": 3.175552749633789, + "step": 6640 + }, + { + "epoch": 1.276657060518732, + "grad_norm": 0.8225224018096924, + "learning_rate": 9.509958104374877e-05, + "loss": 3.1786468505859373, + "step": 6645 + }, + { + "epoch": 1.2776176753121997, + "grad_norm": 0.6611747741699219, + "learning_rate": 9.487791847282517e-05, + "loss": 3.1788875579833986, + "step": 6650 + }, + { + "epoch": 1.2785782901056677, + "grad_norm": 0.49061068892478943, + "learning_rate": 9.465639498056815e-05, + "loss": 3.1779201507568358, + "step": 6655 + }, + { + "epoch": 1.2795389048991355, + "grad_norm": 0.6303700804710388, + "learning_rate": 9.443501112590413e-05, + "loss": 3.176443862915039, + "step": 6660 + }, + { + "epoch": 1.2804995196926032, + "grad_norm": 0.7213342189788818, + "learning_rate": 9.421376746740733e-05, + "loss": 3.1703159332275392, + "step": 6665 + }, + { + "epoch": 1.2814601344860712, + "grad_norm": 0.6647571921348572, + "learning_rate": 9.399266456329815e-05, + "loss": 3.1765552520751954, + "step": 6670 + }, + { + "epoch": 1.282420749279539, + "grad_norm": 0.6062517762184143, + "learning_rate": 9.377170297144196e-05, + "loss": 3.1765026092529296, + "step": 6675 + }, + { + "epoch": 1.2833813640730067, + "grad_norm": 0.4922938048839569, + "learning_rate": 9.355088324934741e-05, + "loss": 3.176412582397461, + "step": 6680 + }, + { + "epoch": 1.2843419788664745, + "grad_norm": 0.5771006941795349, + "learning_rate": 9.333020595416548e-05, + "loss": 3.1751476287841798, + "step": 6685 + }, + { + "epoch": 1.2853025936599423, + "grad_norm": 0.6197171211242676, + "learning_rate": 9.310967164268749e-05, + "loss": 3.174680709838867, + "step": 6690 + }, + { + "epoch": 1.2862632084534102, + "grad_norm": 0.5435570478439331, + "learning_rate": 9.28892808713442e-05, + "loss": 3.1765819549560548, + "step": 6695 + }, + { + "epoch": 1.287223823246878, + "grad_norm": 0.5521527528762817, + "learning_rate": 9.266903419620411e-05, + "loss": 3.173064041137695, + "step": 6700 + }, + { + "epoch": 1.2881844380403458, + "grad_norm": 0.5107563138008118, + "learning_rate": 9.24489321729722e-05, + "loss": 3.1766895294189452, + "step": 6705 + }, + { + "epoch": 1.2891450528338138, + "grad_norm": 0.66566401720047, + "learning_rate": 9.222897535698841e-05, + "loss": 3.175777053833008, + "step": 6710 + }, + { + "epoch": 1.2901056676272815, + "grad_norm": 0.501441478729248, + "learning_rate": 9.20091643032264e-05, + "loss": 3.1773338317871094, + "step": 6715 + }, + { + "epoch": 1.2910662824207493, + "grad_norm": 0.7562515735626221, + "learning_rate": 9.178949956629195e-05, + "loss": 3.1776708602905273, + "step": 6720 + }, + { + "epoch": 1.292026897214217, + "grad_norm": 0.7048882246017456, + "learning_rate": 9.156998170042182e-05, + "loss": 3.1779850006103514, + "step": 6725 + }, + { + "epoch": 1.2929875120076848, + "grad_norm": 0.5082115530967712, + "learning_rate": 9.1350611259482e-05, + "loss": 3.174315071105957, + "step": 6730 + }, + { + "epoch": 1.2939481268011528, + "grad_norm": 0.5481780171394348, + "learning_rate": 9.113138879696667e-05, + "loss": 3.174916648864746, + "step": 6735 + }, + { + "epoch": 1.2949087415946205, + "grad_norm": 0.5947979688644409, + "learning_rate": 9.091231486599656e-05, + "loss": 3.177842903137207, + "step": 6740 + }, + { + "epoch": 1.2958693563880883, + "grad_norm": 0.5829041600227356, + "learning_rate": 9.069339001931775e-05, + "loss": 3.1755516052246096, + "step": 6745 + }, + { + "epoch": 1.2968299711815563, + "grad_norm": 0.7779812812805176, + "learning_rate": 9.047461480929996e-05, + "loss": 3.1758398056030273, + "step": 6750 + }, + { + "epoch": 1.297790585975024, + "grad_norm": 0.5685781836509705, + "learning_rate": 9.025598978793564e-05, + "loss": 3.1764198303222657, + "step": 6755 + }, + { + "epoch": 1.2987512007684918, + "grad_norm": 0.5285525918006897, + "learning_rate": 9.003751550683803e-05, + "loss": 3.1752038955688477, + "step": 6760 + }, + { + "epoch": 1.2997118155619596, + "grad_norm": 0.6162580251693726, + "learning_rate": 8.981919251724023e-05, + "loss": 3.172860336303711, + "step": 6765 + }, + { + "epoch": 1.3006724303554273, + "grad_norm": 0.6464502215385437, + "learning_rate": 8.960102136999346e-05, + "loss": 3.174723815917969, + "step": 6770 + }, + { + "epoch": 1.3016330451488953, + "grad_norm": 0.472971111536026, + "learning_rate": 8.938300261556599e-05, + "loss": 3.176887321472168, + "step": 6775 + }, + { + "epoch": 1.302593659942363, + "grad_norm": 0.46003440022468567, + "learning_rate": 8.916513680404145e-05, + "loss": 3.176510047912598, + "step": 6780 + }, + { + "epoch": 1.3035542747358309, + "grad_norm": 0.7004385590553284, + "learning_rate": 8.894742448511766e-05, + "loss": 3.1788423538208006, + "step": 6785 + }, + { + "epoch": 1.3045148895292988, + "grad_norm": 0.6769760847091675, + "learning_rate": 8.872986620810508e-05, + "loss": 3.175991439819336, + "step": 6790 + }, + { + "epoch": 1.3054755043227666, + "grad_norm": 0.593636691570282, + "learning_rate": 8.851246252192566e-05, + "loss": 3.1727279663085937, + "step": 6795 + }, + { + "epoch": 1.3064361191162344, + "grad_norm": 0.580614447593689, + "learning_rate": 8.82952139751111e-05, + "loss": 3.178467559814453, + "step": 6800 + }, + { + "epoch": 1.3073967339097021, + "grad_norm": 0.519120991230011, + "learning_rate": 8.807812111580186e-05, + "loss": 3.1754281997680662, + "step": 6805 + }, + { + "epoch": 1.30835734870317, + "grad_norm": 0.5653340816497803, + "learning_rate": 8.786118449174541e-05, + "loss": 3.1773509979248047, + "step": 6810 + }, + { + "epoch": 1.3093179634966379, + "grad_norm": 0.7600631713867188, + "learning_rate": 8.764440465029513e-05, + "loss": 3.1753978729248047, + "step": 6815 + }, + { + "epoch": 1.3102785782901056, + "grad_norm": 0.5709202885627747, + "learning_rate": 8.742778213840888e-05, + "loss": 3.176059341430664, + "step": 6820 + }, + { + "epoch": 1.3112391930835736, + "grad_norm": 0.5290617346763611, + "learning_rate": 8.721131750264737e-05, + "loss": 3.176310348510742, + "step": 6825 + }, + { + "epoch": 1.3121998078770414, + "grad_norm": 0.5382722020149231, + "learning_rate": 8.699501128917321e-05, + "loss": 3.1754171371459963, + "step": 6830 + }, + { + "epoch": 1.3131604226705091, + "grad_norm": 0.5834631323814392, + "learning_rate": 8.677886404374903e-05, + "loss": 3.176200103759766, + "step": 6835 + }, + { + "epoch": 1.314121037463977, + "grad_norm": 0.44905054569244385, + "learning_rate": 8.656287631173656e-05, + "loss": 3.178561210632324, + "step": 6840 + }, + { + "epoch": 1.3150816522574447, + "grad_norm": 0.4696488380432129, + "learning_rate": 8.634704863809501e-05, + "loss": 3.1775485992431642, + "step": 6845 + }, + { + "epoch": 1.3160422670509127, + "grad_norm": 0.5930847525596619, + "learning_rate": 8.61313815673798e-05, + "loss": 3.176204299926758, + "step": 6850 + }, + { + "epoch": 1.3170028818443804, + "grad_norm": 0.6078386902809143, + "learning_rate": 8.591587564374094e-05, + "loss": 3.175630569458008, + "step": 6855 + }, + { + "epoch": 1.3179634966378482, + "grad_norm": 0.6279536485671997, + "learning_rate": 8.570053141092209e-05, + "loss": 3.1791570663452147, + "step": 6860 + }, + { + "epoch": 1.3189241114313162, + "grad_norm": 0.4797205328941345, + "learning_rate": 8.548534941225877e-05, + "loss": 3.1787494659423827, + "step": 6865 + }, + { + "epoch": 1.319884726224784, + "grad_norm": 0.692002534866333, + "learning_rate": 8.527033019067736e-05, + "loss": 3.1755470275878905, + "step": 6870 + }, + { + "epoch": 1.3208453410182517, + "grad_norm": 0.609710693359375, + "learning_rate": 8.505547428869326e-05, + "loss": 3.1747051239013673, + "step": 6875 + }, + { + "epoch": 1.3218059558117194, + "grad_norm": 0.511199414730072, + "learning_rate": 8.484078224841002e-05, + "loss": 3.176365280151367, + "step": 6880 + }, + { + "epoch": 1.3227665706051872, + "grad_norm": 0.4836551547050476, + "learning_rate": 8.462625461151769e-05, + "loss": 3.1734235763549803, + "step": 6885 + }, + { + "epoch": 1.3237271853986552, + "grad_norm": 0.6896669268608093, + "learning_rate": 8.441189191929152e-05, + "loss": 3.1753162384033202, + "step": 6890 + }, + { + "epoch": 1.324687800192123, + "grad_norm": 0.5728012323379517, + "learning_rate": 8.419769471259053e-05, + "loss": 3.1740385055541993, + "step": 6895 + }, + { + "epoch": 1.3256484149855907, + "grad_norm": 0.6955984234809875, + "learning_rate": 8.398366353185633e-05, + "loss": 3.175432586669922, + "step": 6900 + }, + { + "epoch": 1.3266090297790587, + "grad_norm": 0.48423975706100464, + "learning_rate": 8.376979891711145e-05, + "loss": 3.1750982284545897, + "step": 6905 + }, + { + "epoch": 1.3275696445725265, + "grad_norm": 0.4743121862411499, + "learning_rate": 8.355610140795827e-05, + "loss": 3.1710472106933594, + "step": 6910 + }, + { + "epoch": 1.3285302593659942, + "grad_norm": 0.6560218334197998, + "learning_rate": 8.334257154357755e-05, + "loss": 3.1815349578857424, + "step": 6915 + }, + { + "epoch": 1.329490874159462, + "grad_norm": 0.536353588104248, + "learning_rate": 8.31292098627271e-05, + "loss": 3.1730472564697267, + "step": 6920 + }, + { + "epoch": 1.3304514889529298, + "grad_norm": 0.49340781569480896, + "learning_rate": 8.291601690374021e-05, + "loss": 3.1722347259521486, + "step": 6925 + }, + { + "epoch": 1.3314121037463977, + "grad_norm": 0.5234084129333496, + "learning_rate": 8.270299320452467e-05, + "loss": 3.172749710083008, + "step": 6930 + }, + { + "epoch": 1.3323727185398655, + "grad_norm": 0.6601616740226746, + "learning_rate": 8.249013930256108e-05, + "loss": 3.1766475677490233, + "step": 6935 + }, + { + "epoch": 1.3333333333333333, + "grad_norm": 0.7026968002319336, + "learning_rate": 8.227745573490176e-05, + "loss": 3.173655319213867, + "step": 6940 + }, + { + "epoch": 1.3342939481268012, + "grad_norm": 0.5302594304084778, + "learning_rate": 8.206494303816904e-05, + "loss": 3.178236389160156, + "step": 6945 + }, + { + "epoch": 1.335254562920269, + "grad_norm": 0.46280986070632935, + "learning_rate": 8.185260174855433e-05, + "loss": 3.173746871948242, + "step": 6950 + }, + { + "epoch": 1.3362151777137368, + "grad_norm": 0.5142591595649719, + "learning_rate": 8.164043240181646e-05, + "loss": 3.173639106750488, + "step": 6955 + }, + { + "epoch": 1.3371757925072045, + "grad_norm": 0.682403028011322, + "learning_rate": 8.142843553328048e-05, + "loss": 3.175322151184082, + "step": 6960 + }, + { + "epoch": 1.3381364073006723, + "grad_norm": 0.6334633231163025, + "learning_rate": 8.121661167783625e-05, + "loss": 3.17366943359375, + "step": 6965 + }, + { + "epoch": 1.3390970220941403, + "grad_norm": 0.606118381023407, + "learning_rate": 8.100496136993712e-05, + "loss": 3.1736032485961916, + "step": 6970 + }, + { + "epoch": 1.340057636887608, + "grad_norm": 0.5890917778015137, + "learning_rate": 8.07934851435984e-05, + "loss": 3.172843360900879, + "step": 6975 + }, + { + "epoch": 1.341018251681076, + "grad_norm": 0.47088661789894104, + "learning_rate": 8.05821835323964e-05, + "loss": 3.1722528457641603, + "step": 6980 + }, + { + "epoch": 1.3419788664745438, + "grad_norm": 0.48279380798339844, + "learning_rate": 8.037105706946673e-05, + "loss": 3.177801513671875, + "step": 6985 + }, + { + "epoch": 1.3429394812680115, + "grad_norm": 0.5353664755821228, + "learning_rate": 8.01601062875032e-05, + "loss": 3.1740434646606444, + "step": 6990 + }, + { + "epoch": 1.3439000960614793, + "grad_norm": 0.5050175189971924, + "learning_rate": 7.994933171875613e-05, + "loss": 3.1727882385253907, + "step": 6995 + }, + { + "epoch": 1.344860710854947, + "grad_norm": 0.44954636693000793, + "learning_rate": 7.973873389503149e-05, + "loss": 3.1749853134155273, + "step": 7000 + }, + { + "epoch": 1.345821325648415, + "grad_norm": 0.6471356749534607, + "learning_rate": 7.952831334768913e-05, + "loss": 3.1730751037597655, + "step": 7005 + }, + { + "epoch": 1.3467819404418828, + "grad_norm": 0.5839700102806091, + "learning_rate": 7.931807060764176e-05, + "loss": 3.172525405883789, + "step": 7010 + }, + { + "epoch": 1.3477425552353506, + "grad_norm": 0.5643643140792847, + "learning_rate": 7.910800620535329e-05, + "loss": 3.172414016723633, + "step": 7015 + }, + { + "epoch": 1.3487031700288186, + "grad_norm": 0.5194916129112244, + "learning_rate": 7.889812067083777e-05, + "loss": 3.1767629623413085, + "step": 7020 + }, + { + "epoch": 1.3496637848222863, + "grad_norm": 0.4932091236114502, + "learning_rate": 7.868841453365797e-05, + "loss": 3.1722953796386717, + "step": 7025 + }, + { + "epoch": 1.350624399615754, + "grad_norm": 0.5307772159576416, + "learning_rate": 7.847888832292397e-05, + "loss": 3.1748939514160157, + "step": 7030 + }, + { + "epoch": 1.3515850144092219, + "grad_norm": 0.5100224614143372, + "learning_rate": 7.826954256729188e-05, + "loss": 3.1737871170043945, + "step": 7035 + }, + { + "epoch": 1.3525456292026896, + "grad_norm": 0.5833878517150879, + "learning_rate": 7.806037779496264e-05, + "loss": 3.171864128112793, + "step": 7040 + }, + { + "epoch": 1.3535062439961576, + "grad_norm": 0.4274962842464447, + "learning_rate": 7.785139453368025e-05, + "loss": 3.174275207519531, + "step": 7045 + }, + { + "epoch": 1.3544668587896254, + "grad_norm": 0.43694546818733215, + "learning_rate": 7.764259331073096e-05, + "loss": 3.1718505859375, + "step": 7050 + }, + { + "epoch": 1.3554274735830931, + "grad_norm": 0.5787099003791809, + "learning_rate": 7.743397465294174e-05, + "loss": 3.171575164794922, + "step": 7055 + }, + { + "epoch": 1.356388088376561, + "grad_norm": 0.45274344086647034, + "learning_rate": 7.72255390866789e-05, + "loss": 3.169318199157715, + "step": 7060 + }, + { + "epoch": 1.3573487031700289, + "grad_norm": 0.490500807762146, + "learning_rate": 7.701728713784662e-05, + "loss": 3.172596740722656, + "step": 7065 + }, + { + "epoch": 1.3583093179634966, + "grad_norm": 0.8986951112747192, + "learning_rate": 7.6809219331886e-05, + "loss": 3.177069664001465, + "step": 7070 + }, + { + "epoch": 1.3592699327569644, + "grad_norm": 0.6794254183769226, + "learning_rate": 7.660133619377346e-05, + "loss": 3.1766523361206054, + "step": 7075 + }, + { + "epoch": 1.3602305475504322, + "grad_norm": 0.4294492304325104, + "learning_rate": 7.639363824801957e-05, + "loss": 3.171103668212891, + "step": 7080 + }, + { + "epoch": 1.3611911623439001, + "grad_norm": 0.4653557240962982, + "learning_rate": 7.618612601866738e-05, + "loss": 3.1739612579345704, + "step": 7085 + }, + { + "epoch": 1.362151777137368, + "grad_norm": 0.48339158296585083, + "learning_rate": 7.597880002929155e-05, + "loss": 3.174298095703125, + "step": 7090 + }, + { + "epoch": 1.3631123919308357, + "grad_norm": 0.4108154773712158, + "learning_rate": 7.577166080299704e-05, + "loss": 3.1711090087890623, + "step": 7095 + }, + { + "epoch": 1.3640730067243036, + "grad_norm": 0.5386460423469543, + "learning_rate": 7.556470886241715e-05, + "loss": 3.172881317138672, + "step": 7100 + }, + { + "epoch": 1.3650336215177714, + "grad_norm": 0.521826982498169, + "learning_rate": 7.535794472971292e-05, + "loss": 3.1787548065185547, + "step": 7105 + }, + { + "epoch": 1.3659942363112392, + "grad_norm": 0.5726050138473511, + "learning_rate": 7.515136892657152e-05, + "loss": 3.1746589660644533, + "step": 7110 + }, + { + "epoch": 1.366954851104707, + "grad_norm": 0.49367332458496094, + "learning_rate": 7.494498197420486e-05, + "loss": 3.171868896484375, + "step": 7115 + }, + { + "epoch": 1.3679154658981747, + "grad_norm": 0.6521613597869873, + "learning_rate": 7.473878439334833e-05, + "loss": 3.1738697052001954, + "step": 7120 + }, + { + "epoch": 1.3688760806916427, + "grad_norm": 0.4904996454715729, + "learning_rate": 7.453277670425964e-05, + "loss": 3.1751945495605467, + "step": 7125 + }, + { + "epoch": 1.3698366954851104, + "grad_norm": 0.5470160841941833, + "learning_rate": 7.432695942671736e-05, + "loss": 3.1728076934814453, + "step": 7130 + }, + { + "epoch": 1.3707973102785782, + "grad_norm": 0.6032967567443848, + "learning_rate": 7.412133308001952e-05, + "loss": 3.166953468322754, + "step": 7135 + }, + { + "epoch": 1.3717579250720462, + "grad_norm": 0.4871714413166046, + "learning_rate": 7.39158981829825e-05, + "loss": 3.1728256225585936, + "step": 7140 + }, + { + "epoch": 1.372718539865514, + "grad_norm": 0.42759451270103455, + "learning_rate": 7.371065525393965e-05, + "loss": 3.1744380950927735, + "step": 7145 + }, + { + "epoch": 1.3736791546589817, + "grad_norm": 0.5663604736328125, + "learning_rate": 7.350560481073996e-05, + "loss": 3.1777429580688477, + "step": 7150 + }, + { + "epoch": 1.3746397694524495, + "grad_norm": 0.5459421873092651, + "learning_rate": 7.330074737074665e-05, + "loss": 3.170821762084961, + "step": 7155 + }, + { + "epoch": 1.3756003842459175, + "grad_norm": 0.46091675758361816, + "learning_rate": 7.309608345083605e-05, + "loss": 3.1742374420166017, + "step": 7160 + }, + { + "epoch": 1.3765609990393852, + "grad_norm": 0.38046908378601074, + "learning_rate": 7.289161356739638e-05, + "loss": 3.1713001251220705, + "step": 7165 + }, + { + "epoch": 1.377521613832853, + "grad_norm": 0.468205064535141, + "learning_rate": 7.268733823632601e-05, + "loss": 3.1741426467895506, + "step": 7170 + }, + { + "epoch": 1.378482228626321, + "grad_norm": 0.6682205200195312, + "learning_rate": 7.248325797303256e-05, + "loss": 3.1712989807128906, + "step": 7175 + }, + { + "epoch": 1.3794428434197887, + "grad_norm": 0.5630512833595276, + "learning_rate": 7.227937329243149e-05, + "loss": 3.171531867980957, + "step": 7180 + }, + { + "epoch": 1.3804034582132565, + "grad_norm": 0.54999178647995, + "learning_rate": 7.20756847089448e-05, + "loss": 3.173867416381836, + "step": 7185 + }, + { + "epoch": 1.3813640730067243, + "grad_norm": 0.43921148777008057, + "learning_rate": 7.187219273649962e-05, + "loss": 3.170378494262695, + "step": 7190 + }, + { + "epoch": 1.382324687800192, + "grad_norm": 0.4150092601776123, + "learning_rate": 7.166889788852707e-05, + "loss": 3.170474624633789, + "step": 7195 + }, + { + "epoch": 1.38328530259366, + "grad_norm": 0.38194531202316284, + "learning_rate": 7.146580067796102e-05, + "loss": 3.1690601348876952, + "step": 7200 + }, + { + "epoch": 1.3842459173871278, + "grad_norm": 0.7853124141693115, + "learning_rate": 7.126290161723642e-05, + "loss": 3.1698192596435546, + "step": 7205 + }, + { + "epoch": 1.3852065321805955, + "grad_norm": 0.4423576593399048, + "learning_rate": 7.106020121828848e-05, + "loss": 3.1732460021972657, + "step": 7210 + }, + { + "epoch": 1.3861671469740635, + "grad_norm": 0.4560396671295166, + "learning_rate": 7.085769999255108e-05, + "loss": 3.1727630615234377, + "step": 7215 + }, + { + "epoch": 1.3871277617675313, + "grad_norm": 0.41349244117736816, + "learning_rate": 7.065539845095567e-05, + "loss": 3.1731325149536134, + "step": 7220 + }, + { + "epoch": 1.388088376560999, + "grad_norm": 0.47393569350242615, + "learning_rate": 7.045329710392967e-05, + "loss": 3.1727859497070314, + "step": 7225 + }, + { + "epoch": 1.3890489913544668, + "grad_norm": 0.7396698594093323, + "learning_rate": 7.025139646139553e-05, + "loss": 3.1715877532958983, + "step": 7230 + }, + { + "epoch": 1.3900096061479346, + "grad_norm": 0.7079951763153076, + "learning_rate": 7.004969703276941e-05, + "loss": 3.169532585144043, + "step": 7235 + }, + { + "epoch": 1.3909702209414025, + "grad_norm": 0.46009525656700134, + "learning_rate": 6.984819932695956e-05, + "loss": 3.169749450683594, + "step": 7240 + }, + { + "epoch": 1.3919308357348703, + "grad_norm": 0.4487673044204712, + "learning_rate": 6.964690385236534e-05, + "loss": 3.1695636749267577, + "step": 7245 + }, + { + "epoch": 1.392891450528338, + "grad_norm": 0.4718167781829834, + "learning_rate": 6.944581111687593e-05, + "loss": 3.168662452697754, + "step": 7250 + }, + { + "epoch": 1.393852065321806, + "grad_norm": 0.46017783880233765, + "learning_rate": 6.924492162786898e-05, + "loss": 3.169230842590332, + "step": 7255 + }, + { + "epoch": 1.3948126801152738, + "grad_norm": 0.45555198192596436, + "learning_rate": 6.904423589220917e-05, + "loss": 3.169890594482422, + "step": 7260 + }, + { + "epoch": 1.3957732949087416, + "grad_norm": 0.4948507845401764, + "learning_rate": 6.884375441624724e-05, + "loss": 3.173093795776367, + "step": 7265 + }, + { + "epoch": 1.3967339097022093, + "grad_norm": 0.43780985474586487, + "learning_rate": 6.864347770581859e-05, + "loss": 3.168751335144043, + "step": 7270 + }, + { + "epoch": 1.397694524495677, + "grad_norm": 0.4438318908214569, + "learning_rate": 6.844340626624178e-05, + "loss": 3.1701900482177736, + "step": 7275 + }, + { + "epoch": 1.398655139289145, + "grad_norm": 0.5062385201454163, + "learning_rate": 6.824354060231765e-05, + "loss": 3.1665103912353514, + "step": 7280 + }, + { + "epoch": 1.3996157540826129, + "grad_norm": 0.4086364507675171, + "learning_rate": 6.804388121832777e-05, + "loss": 3.1716842651367188, + "step": 7285 + }, + { + "epoch": 1.4005763688760806, + "grad_norm": 0.5124282836914062, + "learning_rate": 6.784442861803331e-05, + "loss": 3.1696632385253904, + "step": 7290 + }, + { + "epoch": 1.4015369836695486, + "grad_norm": 0.5204159617424011, + "learning_rate": 6.764518330467346e-05, + "loss": 3.1707353591918945, + "step": 7295 + }, + { + "epoch": 1.4024975984630164, + "grad_norm": 0.6575141549110413, + "learning_rate": 6.744614578096475e-05, + "loss": 3.171962547302246, + "step": 7300 + }, + { + "epoch": 1.4034582132564841, + "grad_norm": 0.5063356757164001, + "learning_rate": 6.72473165490993e-05, + "loss": 3.1709272384643556, + "step": 7305 + }, + { + "epoch": 1.4044188280499519, + "grad_norm": 0.4604819416999817, + "learning_rate": 6.704869611074351e-05, + "loss": 3.173125076293945, + "step": 7310 + }, + { + "epoch": 1.4053794428434199, + "grad_norm": 0.4766218960285187, + "learning_rate": 6.685028496703719e-05, + "loss": 3.1713457107543945, + "step": 7315 + }, + { + "epoch": 1.4063400576368876, + "grad_norm": 0.48007673025131226, + "learning_rate": 6.665208361859203e-05, + "loss": 3.1723331451416015, + "step": 7320 + }, + { + "epoch": 1.4073006724303554, + "grad_norm": 0.3352242410182953, + "learning_rate": 6.64540925654904e-05, + "loss": 3.171708679199219, + "step": 7325 + }, + { + "epoch": 1.4082612872238234, + "grad_norm": 0.6097206473350525, + "learning_rate": 6.625631230728393e-05, + "loss": 3.169724464416504, + "step": 7330 + }, + { + "epoch": 1.4092219020172911, + "grad_norm": 0.45327913761138916, + "learning_rate": 6.605874334299257e-05, + "loss": 3.172129821777344, + "step": 7335 + }, + { + "epoch": 1.410182516810759, + "grad_norm": 0.3918154537677765, + "learning_rate": 6.586138617110312e-05, + "loss": 3.1696929931640625, + "step": 7340 + }, + { + "epoch": 1.4111431316042267, + "grad_norm": 0.45734044909477234, + "learning_rate": 6.566424128956788e-05, + "loss": 3.169087600708008, + "step": 7345 + }, + { + "epoch": 1.4121037463976944, + "grad_norm": 0.41583287715911865, + "learning_rate": 6.546730919580365e-05, + "loss": 3.169902801513672, + "step": 7350 + }, + { + "epoch": 1.4130643611911624, + "grad_norm": 0.3783586919307709, + "learning_rate": 6.527059038669032e-05, + "loss": 3.173455810546875, + "step": 7355 + }, + { + "epoch": 1.4140249759846302, + "grad_norm": 0.5969278216362, + "learning_rate": 6.507408535856968e-05, + "loss": 3.171731185913086, + "step": 7360 + }, + { + "epoch": 1.414985590778098, + "grad_norm": 0.5529734492301941, + "learning_rate": 6.487779460724388e-05, + "loss": 3.169993591308594, + "step": 7365 + }, + { + "epoch": 1.415946205571566, + "grad_norm": 0.460746705532074, + "learning_rate": 6.468171862797484e-05, + "loss": 3.1684539794921873, + "step": 7370 + }, + { + "epoch": 1.4169068203650337, + "grad_norm": 0.40012434124946594, + "learning_rate": 6.448585791548231e-05, + "loss": 3.170420837402344, + "step": 7375 + }, + { + "epoch": 1.4178674351585014, + "grad_norm": 0.4528000056743622, + "learning_rate": 6.42902129639429e-05, + "loss": 3.1724185943603516, + "step": 7380 + }, + { + "epoch": 1.4188280499519692, + "grad_norm": 0.5503178834915161, + "learning_rate": 6.409478426698893e-05, + "loss": 3.170566749572754, + "step": 7385 + }, + { + "epoch": 1.419788664745437, + "grad_norm": 0.4782489240169525, + "learning_rate": 6.389957231770705e-05, + "loss": 3.1753549575805664, + "step": 7390 + }, + { + "epoch": 1.420749279538905, + "grad_norm": 0.5017015933990479, + "learning_rate": 6.370457760863708e-05, + "loss": 3.1712177276611326, + "step": 7395 + }, + { + "epoch": 1.4217098943323727, + "grad_norm": 0.4609530568122864, + "learning_rate": 6.35098006317706e-05, + "loss": 3.1715621948242188, + "step": 7400 + }, + { + "epoch": 1.4226705091258405, + "grad_norm": 0.43000540137290955, + "learning_rate": 6.331524187854992e-05, + "loss": 3.1696197509765627, + "step": 7405 + }, + { + "epoch": 1.4236311239193085, + "grad_norm": 0.6466848850250244, + "learning_rate": 6.31209018398668e-05, + "loss": 3.171028900146484, + "step": 7410 + }, + { + "epoch": 1.4245917387127762, + "grad_norm": 0.4881599247455597, + "learning_rate": 6.292678100606101e-05, + "loss": 3.1710824966430664, + "step": 7415 + }, + { + "epoch": 1.425552353506244, + "grad_norm": 0.42242470383644104, + "learning_rate": 6.273287986691934e-05, + "loss": 3.1717771530151366, + "step": 7420 + }, + { + "epoch": 1.4265129682997117, + "grad_norm": 0.40897586941719055, + "learning_rate": 6.253919891167427e-05, + "loss": 3.168376159667969, + "step": 7425 + }, + { + "epoch": 1.4274735830931795, + "grad_norm": 0.31631234288215637, + "learning_rate": 6.23457386290028e-05, + "loss": 3.167464828491211, + "step": 7430 + }, + { + "epoch": 1.4284341978866475, + "grad_norm": 0.423149973154068, + "learning_rate": 6.215249950702488e-05, + "loss": 3.1704282760620117, + "step": 7435 + }, + { + "epoch": 1.4293948126801153, + "grad_norm": 0.663411557674408, + "learning_rate": 6.195948203330282e-05, + "loss": 3.170808219909668, + "step": 7440 + }, + { + "epoch": 1.430355427473583, + "grad_norm": 0.35881131887435913, + "learning_rate": 6.176668669483948e-05, + "loss": 3.1693355560302736, + "step": 7445 + }, + { + "epoch": 1.431316042267051, + "grad_norm": 0.3775772452354431, + "learning_rate": 6.15741139780772e-05, + "loss": 3.1732666015625, + "step": 7450 + }, + { + "epoch": 1.4322766570605188, + "grad_norm": 0.40818941593170166, + "learning_rate": 6.138176436889672e-05, + "loss": 3.1699262619018556, + "step": 7455 + }, + { + "epoch": 1.4332372718539865, + "grad_norm": 0.31840816140174866, + "learning_rate": 6.118963835261588e-05, + "loss": 3.171256256103516, + "step": 7460 + }, + { + "epoch": 1.4341978866474543, + "grad_norm": 0.45461708307266235, + "learning_rate": 6.099773641398834e-05, + "loss": 3.1705402374267577, + "step": 7465 + }, + { + "epoch": 1.435158501440922, + "grad_norm": 0.49488258361816406, + "learning_rate": 6.080605903720229e-05, + "loss": 3.1715473175048827, + "step": 7470 + }, + { + "epoch": 1.43611911623439, + "grad_norm": 0.4354032278060913, + "learning_rate": 6.061460670587948e-05, + "loss": 3.170640563964844, + "step": 7475 + }, + { + "epoch": 1.4370797310278578, + "grad_norm": 0.39679399132728577, + "learning_rate": 6.0423379903073824e-05, + "loss": 3.170828437805176, + "step": 7480 + }, + { + "epoch": 1.4380403458213258, + "grad_norm": 0.37695419788360596, + "learning_rate": 6.023237911127005e-05, + "loss": 3.167501449584961, + "step": 7485 + }, + { + "epoch": 1.4390009606147935, + "grad_norm": 0.37574145197868347, + "learning_rate": 6.004160481238281e-05, + "loss": 3.172044372558594, + "step": 7490 + }, + { + "epoch": 1.4399615754082613, + "grad_norm": 0.5872917771339417, + "learning_rate": 5.9851057487755225e-05, + "loss": 3.167539596557617, + "step": 7495 + }, + { + "epoch": 1.440922190201729, + "grad_norm": 0.43774786591529846, + "learning_rate": 5.966073761815776e-05, + "loss": 3.170111656188965, + "step": 7500 + }, + { + "epoch": 1.4418828049951968, + "grad_norm": 0.319992333650589, + "learning_rate": 5.947064568378696e-05, + "loss": 3.1693355560302736, + "step": 7505 + }, + { + "epoch": 1.4428434197886648, + "grad_norm": 0.4077332019805908, + "learning_rate": 5.928078216426427e-05, + "loss": 3.1672155380249025, + "step": 7510 + }, + { + "epoch": 1.4438040345821326, + "grad_norm": 0.355669766664505, + "learning_rate": 5.909114753863488e-05, + "loss": 3.1689855575561525, + "step": 7515 + }, + { + "epoch": 1.4447646493756003, + "grad_norm": 0.5554761290550232, + "learning_rate": 5.8901742285366315e-05, + "loss": 3.171320343017578, + "step": 7520 + }, + { + "epoch": 1.4457252641690683, + "grad_norm": 0.4790358543395996, + "learning_rate": 5.8712566882347504e-05, + "loss": 3.1696239471435548, + "step": 7525 + }, + { + "epoch": 1.446685878962536, + "grad_norm": 0.4286898076534271, + "learning_rate": 5.8523621806887374e-05, + "loss": 3.169162368774414, + "step": 7530 + }, + { + "epoch": 1.4476464937560038, + "grad_norm": 0.42978528141975403, + "learning_rate": 5.8334907535713836e-05, + "loss": 3.172647476196289, + "step": 7535 + }, + { + "epoch": 1.4486071085494716, + "grad_norm": 0.4055570363998413, + "learning_rate": 5.8146424544972205e-05, + "loss": 3.1680522918701173, + "step": 7540 + }, + { + "epoch": 1.4495677233429394, + "grad_norm": 0.3212641775608063, + "learning_rate": 5.7958173310224465e-05, + "loss": 3.168552017211914, + "step": 7545 + }, + { + "epoch": 1.4505283381364074, + "grad_norm": 0.40570881962776184, + "learning_rate": 5.7770154306447866e-05, + "loss": 3.172994041442871, + "step": 7550 + }, + { + "epoch": 1.4514889529298751, + "grad_norm": 0.4512649476528168, + "learning_rate": 5.758236800803351e-05, + "loss": 3.171183395385742, + "step": 7555 + }, + { + "epoch": 1.4524495677233429, + "grad_norm": 0.5307871103286743, + "learning_rate": 5.739481488878558e-05, + "loss": 3.172830581665039, + "step": 7560 + }, + { + "epoch": 1.4534101825168109, + "grad_norm": 0.38034340739250183, + "learning_rate": 5.720749542191979e-05, + "loss": 3.1710861206054686, + "step": 7565 + }, + { + "epoch": 1.4543707973102786, + "grad_norm": 0.46515336632728577, + "learning_rate": 5.7020410080062416e-05, + "loss": 3.1718137741088865, + "step": 7570 + }, + { + "epoch": 1.4553314121037464, + "grad_norm": 0.4164998531341553, + "learning_rate": 5.683355933524894e-05, + "loss": 3.1689435958862306, + "step": 7575 + }, + { + "epoch": 1.4562920268972142, + "grad_norm": 0.48741379380226135, + "learning_rate": 5.6646943658922975e-05, + "loss": 3.167072296142578, + "step": 7580 + }, + { + "epoch": 1.457252641690682, + "grad_norm": 0.39116013050079346, + "learning_rate": 5.646056352193505e-05, + "loss": 3.171469497680664, + "step": 7585 + }, + { + "epoch": 1.45821325648415, + "grad_norm": 0.4745250642299652, + "learning_rate": 5.6274419394541296e-05, + "loss": 3.168760871887207, + "step": 7590 + }, + { + "epoch": 1.4591738712776177, + "grad_norm": 0.5160530805587769, + "learning_rate": 5.608851174640247e-05, + "loss": 3.168941307067871, + "step": 7595 + }, + { + "epoch": 1.4601344860710854, + "grad_norm": 0.3649737238883972, + "learning_rate": 5.590284104658264e-05, + "loss": 3.1685501098632813, + "step": 7600 + }, + { + "epoch": 1.4610951008645534, + "grad_norm": 0.37525928020477295, + "learning_rate": 5.571740776354811e-05, + "loss": 3.169935607910156, + "step": 7605 + }, + { + "epoch": 1.4620557156580212, + "grad_norm": 0.35180380940437317, + "learning_rate": 5.553221236516594e-05, + "loss": 3.1680299758911135, + "step": 7610 + }, + { + "epoch": 1.463016330451489, + "grad_norm": 0.5641520023345947, + "learning_rate": 5.534725531870317e-05, + "loss": 3.171617126464844, + "step": 7615 + }, + { + "epoch": 1.4639769452449567, + "grad_norm": 0.5281386375427246, + "learning_rate": 5.516253709082547e-05, + "loss": 3.1680675506591798, + "step": 7620 + }, + { + "epoch": 1.4649375600384245, + "grad_norm": 0.37240076065063477, + "learning_rate": 5.4978058147595796e-05, + "loss": 3.164999008178711, + "step": 7625 + }, + { + "epoch": 1.4658981748318924, + "grad_norm": 0.37205061316490173, + "learning_rate": 5.479381895447346e-05, + "loss": 3.1697301864624023, + "step": 7630 + }, + { + "epoch": 1.4668587896253602, + "grad_norm": 0.35913726687431335, + "learning_rate": 5.4609819976312854e-05, + "loss": 3.1634567260742186, + "step": 7635 + }, + { + "epoch": 1.4678194044188282, + "grad_norm": 0.42885205149650574, + "learning_rate": 5.4426061677362284e-05, + "loss": 3.173727798461914, + "step": 7640 + }, + { + "epoch": 1.468780019212296, + "grad_norm": 0.3557131588459015, + "learning_rate": 5.424254452126279e-05, + "loss": 3.1679935455322266, + "step": 7645 + }, + { + "epoch": 1.4697406340057637, + "grad_norm": 0.4141709804534912, + "learning_rate": 5.4059268971047e-05, + "loss": 3.1674577713012697, + "step": 7650 + }, + { + "epoch": 1.4707012487992315, + "grad_norm": 0.40807923674583435, + "learning_rate": 5.387623548913795e-05, + "loss": 3.171774673461914, + "step": 7655 + }, + { + "epoch": 1.4716618635926992, + "grad_norm": 0.3850151598453522, + "learning_rate": 5.36934445373478e-05, + "loss": 3.164442443847656, + "step": 7660 + }, + { + "epoch": 1.4726224783861672, + "grad_norm": 0.43482470512390137, + "learning_rate": 5.3510896576876924e-05, + "loss": 3.1658775329589846, + "step": 7665 + }, + { + "epoch": 1.473583093179635, + "grad_norm": 0.40651193261146545, + "learning_rate": 5.3328592068312565e-05, + "loss": 3.164666748046875, + "step": 7670 + }, + { + "epoch": 1.4745437079731027, + "grad_norm": 0.33272117376327515, + "learning_rate": 5.3146531471627737e-05, + "loss": 3.169040298461914, + "step": 7675 + }, + { + "epoch": 1.4755043227665707, + "grad_norm": 0.3396502435207367, + "learning_rate": 5.29647152461799e-05, + "loss": 3.1710128784179688, + "step": 7680 + }, + { + "epoch": 1.4764649375600385, + "grad_norm": 0.27867111563682556, + "learning_rate": 5.278314385071011e-05, + "loss": 3.169612693786621, + "step": 7685 + }, + { + "epoch": 1.4774255523535063, + "grad_norm": 0.5024138689041138, + "learning_rate": 5.260181774334165e-05, + "loss": 3.1680776596069338, + "step": 7690 + }, + { + "epoch": 1.478386167146974, + "grad_norm": 0.5927129983901978, + "learning_rate": 5.2420737381578814e-05, + "loss": 3.1725555419921876, + "step": 7695 + }, + { + "epoch": 1.4793467819404418, + "grad_norm": 0.2990652322769165, + "learning_rate": 5.223990322230596e-05, + "loss": 3.1685482025146485, + "step": 7700 + }, + { + "epoch": 1.4803073967339098, + "grad_norm": 0.41576090455055237, + "learning_rate": 5.205931572178625e-05, + "loss": 3.1704280853271483, + "step": 7705 + }, + { + "epoch": 1.4812680115273775, + "grad_norm": 0.33696505427360535, + "learning_rate": 5.187897533566047e-05, + "loss": 3.1678043365478517, + "step": 7710 + }, + { + "epoch": 1.4822286263208453, + "grad_norm": 0.33144330978393555, + "learning_rate": 5.169888251894587e-05, + "loss": 3.1680809020996095, + "step": 7715 + }, + { + "epoch": 1.4831892411143133, + "grad_norm": 0.35232898592948914, + "learning_rate": 5.151903772603517e-05, + "loss": 3.1694143295288084, + "step": 7720 + }, + { + "epoch": 1.484149855907781, + "grad_norm": 0.482440710067749, + "learning_rate": 5.1339441410695225e-05, + "loss": 3.167070007324219, + "step": 7725 + }, + { + "epoch": 1.4851104707012488, + "grad_norm": 0.44148990511894226, + "learning_rate": 5.116009402606591e-05, + "loss": 3.1692365646362304, + "step": 7730 + }, + { + "epoch": 1.4860710854947166, + "grad_norm": 0.4089405834674835, + "learning_rate": 5.0980996024659075e-05, + "loss": 3.165630912780762, + "step": 7735 + }, + { + "epoch": 1.4870317002881843, + "grad_norm": 0.3344769775867462, + "learning_rate": 5.0802147858357386e-05, + "loss": 3.1695476531982423, + "step": 7740 + }, + { + "epoch": 1.4879923150816523, + "grad_norm": 0.389249324798584, + "learning_rate": 5.0623549978413166e-05, + "loss": 3.171805000305176, + "step": 7745 + }, + { + "epoch": 1.48895292987512, + "grad_norm": 0.3440057635307312, + "learning_rate": 5.0445202835447056e-05, + "loss": 3.1670265197753906, + "step": 7750 + }, + { + "epoch": 1.4899135446685878, + "grad_norm": 0.3163948059082031, + "learning_rate": 5.026710687944728e-05, + "loss": 3.1670303344726562, + "step": 7755 + }, + { + "epoch": 1.4908741594620558, + "grad_norm": 0.3952416777610779, + "learning_rate": 5.0089262559768246e-05, + "loss": 3.171236038208008, + "step": 7760 + }, + { + "epoch": 1.4918347742555236, + "grad_norm": 0.3361209034919739, + "learning_rate": 4.9911670325129304e-05, + "loss": 3.1686822891235353, + "step": 7765 + }, + { + "epoch": 1.4927953890489913, + "grad_norm": 0.35905590653419495, + "learning_rate": 4.9734330623613924e-05, + "loss": 3.1694427490234376, + "step": 7770 + }, + { + "epoch": 1.493756003842459, + "grad_norm": 0.4923272132873535, + "learning_rate": 4.955724390266841e-05, + "loss": 3.171158218383789, + "step": 7775 + }, + { + "epoch": 1.4947166186359269, + "grad_norm": 0.3285467326641083, + "learning_rate": 4.9380410609100674e-05, + "loss": 3.1719661712646485, + "step": 7780 + }, + { + "epoch": 1.4956772334293948, + "grad_norm": 0.35682541131973267, + "learning_rate": 4.920383118907929e-05, + "loss": 3.1671146392822265, + "step": 7785 + }, + { + "epoch": 1.4966378482228626, + "grad_norm": 0.5638254880905151, + "learning_rate": 4.902750608813222e-05, + "loss": 3.1699798583984373, + "step": 7790 + }, + { + "epoch": 1.4975984630163304, + "grad_norm": 0.2942904531955719, + "learning_rate": 4.885143575114587e-05, + "loss": 3.1689422607421873, + "step": 7795 + }, + { + "epoch": 1.4985590778097984, + "grad_norm": 0.3493422269821167, + "learning_rate": 4.8675620622363645e-05, + "loss": 3.168619918823242, + "step": 7800 + }, + { + "epoch": 1.4995196926032661, + "grad_norm": 0.40712451934814453, + "learning_rate": 4.850006114538519e-05, + "loss": 3.169388771057129, + "step": 7805 + }, + { + "epoch": 1.5004803073967339, + "grad_norm": 0.31108608841896057, + "learning_rate": 4.8324757763165075e-05, + "loss": 3.1686771392822264, + "step": 7810 + }, + { + "epoch": 1.5014409221902016, + "grad_norm": 0.5020172595977783, + "learning_rate": 4.814971091801179e-05, + "loss": 3.1644512176513673, + "step": 7815 + }, + { + "epoch": 1.5024015369836694, + "grad_norm": 0.4801461696624756, + "learning_rate": 4.7974921051586385e-05, + "loss": 3.1699331283569334, + "step": 7820 + }, + { + "epoch": 1.5033621517771374, + "grad_norm": 0.31796136498451233, + "learning_rate": 4.780038860490164e-05, + "loss": 3.1699861526489257, + "step": 7825 + }, + { + "epoch": 1.5043227665706052, + "grad_norm": 0.3169126510620117, + "learning_rate": 4.762611401832089e-05, + "loss": 3.1705883026123045, + "step": 7830 + }, + { + "epoch": 1.5052833813640731, + "grad_norm": 0.3533933758735657, + "learning_rate": 4.745209773155671e-05, + "loss": 3.165675926208496, + "step": 7835 + }, + { + "epoch": 1.506243996157541, + "grad_norm": 0.4423210024833679, + "learning_rate": 4.727834018367007e-05, + "loss": 3.1672183990478517, + "step": 7840 + }, + { + "epoch": 1.5072046109510087, + "grad_norm": 0.33951136469841003, + "learning_rate": 4.710484181306912e-05, + "loss": 3.169887733459473, + "step": 7845 + }, + { + "epoch": 1.5081652257444764, + "grad_norm": 0.298172265291214, + "learning_rate": 4.693160305750801e-05, + "loss": 3.1656867980957033, + "step": 7850 + }, + { + "epoch": 1.5091258405379442, + "grad_norm": 0.3188192546367645, + "learning_rate": 4.675862435408591e-05, + "loss": 3.1662307739257813, + "step": 7855 + }, + { + "epoch": 1.510086455331412, + "grad_norm": 0.3231668174266815, + "learning_rate": 4.6585906139245834e-05, + "loss": 3.1684810638427736, + "step": 7860 + }, + { + "epoch": 1.51104707012488, + "grad_norm": 0.46347707509994507, + "learning_rate": 4.641344884877362e-05, + "loss": 3.1662145614624024, + "step": 7865 + }, + { + "epoch": 1.5120076849183477, + "grad_norm": 0.4303688704967499, + "learning_rate": 4.6241252917796576e-05, + "loss": 3.166950798034668, + "step": 7870 + }, + { + "epoch": 1.5129682997118157, + "grad_norm": 0.3407163918018341, + "learning_rate": 4.6069318780782765e-05, + "loss": 3.1677284240722656, + "step": 7875 + }, + { + "epoch": 1.5139289145052834, + "grad_norm": 0.2888704240322113, + "learning_rate": 4.589764687153967e-05, + "loss": 3.162090301513672, + "step": 7880 + }, + { + "epoch": 1.5148895292987512, + "grad_norm": 0.3005317747592926, + "learning_rate": 4.5726237623213155e-05, + "loss": 3.1687442779541017, + "step": 7885 + }, + { + "epoch": 1.515850144092219, + "grad_norm": 0.2978266477584839, + "learning_rate": 4.555509146828624e-05, + "loss": 3.1686517715454103, + "step": 7890 + }, + { + "epoch": 1.5168107588856867, + "grad_norm": 0.2926294505596161, + "learning_rate": 4.53842088385783e-05, + "loss": 3.170303726196289, + "step": 7895 + }, + { + "epoch": 1.5177713736791547, + "grad_norm": 0.47925063967704773, + "learning_rate": 4.521359016524376e-05, + "loss": 3.169039726257324, + "step": 7900 + }, + { + "epoch": 1.5187319884726225, + "grad_norm": 0.4029441475868225, + "learning_rate": 4.5043235878770965e-05, + "loss": 3.166218376159668, + "step": 7905 + }, + { + "epoch": 1.5196926032660905, + "grad_norm": 0.31125885248184204, + "learning_rate": 4.4873146408981295e-05, + "loss": 3.1654186248779297, + "step": 7910 + }, + { + "epoch": 1.5206532180595582, + "grad_norm": 0.3251868784427643, + "learning_rate": 4.4703322185027926e-05, + "loss": 3.1693817138671876, + "step": 7915 + }, + { + "epoch": 1.521613832853026, + "grad_norm": 0.31480100750923157, + "learning_rate": 4.453376363539481e-05, + "loss": 3.1616186141967773, + "step": 7920 + }, + { + "epoch": 1.5225744476464937, + "grad_norm": 0.3208870589733124, + "learning_rate": 4.436447118789555e-05, + "loss": 3.168798637390137, + "step": 7925 + }, + { + "epoch": 1.5235350624399615, + "grad_norm": 0.3242574632167816, + "learning_rate": 4.419544526967238e-05, + "loss": 3.171648406982422, + "step": 7930 + }, + { + "epoch": 1.5244956772334293, + "grad_norm": 0.3922894597053528, + "learning_rate": 4.402668630719504e-05, + "loss": 3.1695817947387694, + "step": 7935 + }, + { + "epoch": 1.5254562920268973, + "grad_norm": 0.3969297409057617, + "learning_rate": 4.385819472625963e-05, + "loss": 3.1659820556640623, + "step": 7940 + }, + { + "epoch": 1.526416906820365, + "grad_norm": 0.3534747362136841, + "learning_rate": 4.368997095198775e-05, + "loss": 3.164035415649414, + "step": 7945 + }, + { + "epoch": 1.527377521613833, + "grad_norm": 0.387599915266037, + "learning_rate": 4.352201540882523e-05, + "loss": 3.166557502746582, + "step": 7950 + }, + { + "epoch": 1.5283381364073008, + "grad_norm": 0.3567873537540436, + "learning_rate": 4.33543285205412e-05, + "loss": 3.1643226623535154, + "step": 7955 + }, + { + "epoch": 1.5292987512007685, + "grad_norm": 0.3381181061267853, + "learning_rate": 4.318691071022676e-05, + "loss": 3.1638862609863283, + "step": 7960 + }, + { + "epoch": 1.5302593659942363, + "grad_norm": 0.2984130084514618, + "learning_rate": 4.301976240029428e-05, + "loss": 3.1657569885253904, + "step": 7965 + }, + { + "epoch": 1.531219980787704, + "grad_norm": 0.2934906780719757, + "learning_rate": 4.285288401247614e-05, + "loss": 3.165040969848633, + "step": 7970 + }, + { + "epoch": 1.5321805955811718, + "grad_norm": 0.2868797779083252, + "learning_rate": 4.268627596782354e-05, + "loss": 3.166617202758789, + "step": 7975 + }, + { + "epoch": 1.5331412103746398, + "grad_norm": 0.40537068247795105, + "learning_rate": 4.251993868670569e-05, + "loss": 3.1657033920288087, + "step": 7980 + }, + { + "epoch": 1.5341018251681076, + "grad_norm": 0.3606870770454407, + "learning_rate": 4.235387258880871e-05, + "loss": 3.1681283950805663, + "step": 7985 + }, + { + "epoch": 1.5350624399615755, + "grad_norm": 0.3049268126487732, + "learning_rate": 4.218807809313428e-05, + "loss": 3.165813446044922, + "step": 7990 + }, + { + "epoch": 1.5360230547550433, + "grad_norm": 0.3047685921192169, + "learning_rate": 4.202255561799897e-05, + "loss": 3.1700771331787108, + "step": 7995 + }, + { + "epoch": 1.536983669548511, + "grad_norm": 0.32123667001724243, + "learning_rate": 4.1857305581032974e-05, + "loss": 3.16713924407959, + "step": 8000 + }, + { + "epoch": 1.5379442843419788, + "grad_norm": 0.31456029415130615, + "learning_rate": 4.1692328399179134e-05, + "loss": 3.1665233612060546, + "step": 8005 + }, + { + "epoch": 1.5389048991354466, + "grad_norm": 0.2823289632797241, + "learning_rate": 4.1527624488691706e-05, + "loss": 3.165792465209961, + "step": 8010 + }, + { + "epoch": 1.5398655139289144, + "grad_norm": 0.41505372524261475, + "learning_rate": 4.1363194265135584e-05, + "loss": 3.165696907043457, + "step": 8015 + }, + { + "epoch": 1.5408261287223823, + "grad_norm": 0.3873119056224823, + "learning_rate": 4.1199038143385114e-05, + "loss": 3.1616024017333983, + "step": 8020 + }, + { + "epoch": 1.54178674351585, + "grad_norm": 0.29045772552490234, + "learning_rate": 4.1035156537623056e-05, + "loss": 3.163679504394531, + "step": 8025 + }, + { + "epoch": 1.542747358309318, + "grad_norm": 0.31360870599746704, + "learning_rate": 4.087154986133944e-05, + "loss": 3.1689876556396483, + "step": 8030 + }, + { + "epoch": 1.5437079731027858, + "grad_norm": 0.34343641996383667, + "learning_rate": 4.070821852733074e-05, + "loss": 3.1644660949707033, + "step": 8035 + }, + { + "epoch": 1.5446685878962536, + "grad_norm": 0.3275468647480011, + "learning_rate": 4.054516294769871e-05, + "loss": 3.161859321594238, + "step": 8040 + }, + { + "epoch": 1.5456292026897214, + "grad_norm": 0.3492441773414612, + "learning_rate": 4.038238353384919e-05, + "loss": 3.166157531738281, + "step": 8045 + }, + { + "epoch": 1.5465898174831891, + "grad_norm": 0.3080005645751953, + "learning_rate": 4.021988069649138e-05, + "loss": 3.169676399230957, + "step": 8050 + }, + { + "epoch": 1.547550432276657, + "grad_norm": 0.2594451904296875, + "learning_rate": 4.0057654845636714e-05, + "loss": 3.170206642150879, + "step": 8055 + }, + { + "epoch": 1.5485110470701249, + "grad_norm": 0.2814268469810486, + "learning_rate": 3.989570639059753e-05, + "loss": 3.164537811279297, + "step": 8060 + }, + { + "epoch": 1.5494716618635929, + "grad_norm": 0.3419535756111145, + "learning_rate": 3.973403573998647e-05, + "loss": 3.169506645202637, + "step": 8065 + }, + { + "epoch": 1.5504322766570606, + "grad_norm": 0.3430168032646179, + "learning_rate": 3.957264330171511e-05, + "loss": 3.1708099365234377, + "step": 8070 + }, + { + "epoch": 1.5513928914505284, + "grad_norm": 0.3074200749397278, + "learning_rate": 3.941152948299321e-05, + "loss": 3.1663097381591796, + "step": 8075 + }, + { + "epoch": 1.5523535062439962, + "grad_norm": 0.29986321926116943, + "learning_rate": 3.9250694690327383e-05, + "loss": 3.1678709030151366, + "step": 8080 + }, + { + "epoch": 1.553314121037464, + "grad_norm": 0.305226594209671, + "learning_rate": 3.9090139329520327e-05, + "loss": 3.163295364379883, + "step": 8085 + }, + { + "epoch": 1.5542747358309317, + "grad_norm": 0.27455848455429077, + "learning_rate": 3.8929863805669706e-05, + "loss": 3.163608741760254, + "step": 8090 + }, + { + "epoch": 1.5552353506243997, + "grad_norm": 0.30126896500587463, + "learning_rate": 3.876986852316715e-05, + "loss": 3.1656238555908205, + "step": 8095 + }, + { + "epoch": 1.5561959654178674, + "grad_norm": 0.3914654552936554, + "learning_rate": 3.861015388569709e-05, + "loss": 3.1641334533691405, + "step": 8100 + }, + { + "epoch": 1.5571565802113354, + "grad_norm": 0.3612956404685974, + "learning_rate": 3.845072029623598e-05, + "loss": 3.1632358551025392, + "step": 8105 + }, + { + "epoch": 1.5581171950048032, + "grad_norm": 0.2885013222694397, + "learning_rate": 3.8291568157051154e-05, + "loss": 3.1646907806396483, + "step": 8110 + }, + { + "epoch": 1.559077809798271, + "grad_norm": 0.29719772934913635, + "learning_rate": 3.8132697869699705e-05, + "loss": 3.1659112930297852, + "step": 8115 + }, + { + "epoch": 1.5600384245917387, + "grad_norm": 0.308212012052536, + "learning_rate": 3.797410983502766e-05, + "loss": 3.1631364822387695, + "step": 8120 + }, + { + "epoch": 1.5609990393852065, + "grad_norm": 0.30908966064453125, + "learning_rate": 3.781580445316906e-05, + "loss": 3.1642078399658202, + "step": 8125 + }, + { + "epoch": 1.5619596541786742, + "grad_norm": 0.3072703182697296, + "learning_rate": 3.765778212354445e-05, + "loss": 3.1633129119873047, + "step": 8130 + }, + { + "epoch": 1.5629202689721422, + "grad_norm": 0.3363180458545685, + "learning_rate": 3.7500043244860475e-05, + "loss": 3.16600341796875, + "step": 8135 + }, + { + "epoch": 1.56388088376561, + "grad_norm": 0.41240906715393066, + "learning_rate": 3.7342588215108446e-05, + "loss": 3.168244743347168, + "step": 8140 + }, + { + "epoch": 1.564841498559078, + "grad_norm": 0.3254190683364868, + "learning_rate": 3.7185417431563644e-05, + "loss": 3.1647628784179687, + "step": 8145 + }, + { + "epoch": 1.5658021133525457, + "grad_norm": 0.3488558232784271, + "learning_rate": 3.702853129078398e-05, + "loss": 3.1683582305908202, + "step": 8150 + }, + { + "epoch": 1.5667627281460135, + "grad_norm": 0.34358084201812744, + "learning_rate": 3.6871930188609325e-05, + "loss": 3.1676990509033205, + "step": 8155 + }, + { + "epoch": 1.5677233429394812, + "grad_norm": 0.3370071053504944, + "learning_rate": 3.671561452016033e-05, + "loss": 3.1652057647705076, + "step": 8160 + }, + { + "epoch": 1.568683957732949, + "grad_norm": 0.23271813988685608, + "learning_rate": 3.655958467983749e-05, + "loss": 3.164310073852539, + "step": 8165 + }, + { + "epoch": 1.5696445725264168, + "grad_norm": 0.33882594108581543, + "learning_rate": 3.6403841061320026e-05, + "loss": 3.1641658782958983, + "step": 8170 + }, + { + "epoch": 1.5706051873198847, + "grad_norm": 0.25975945591926575, + "learning_rate": 3.6248384057565104e-05, + "loss": 3.1639142990112306, + "step": 8175 + }, + { + "epoch": 1.5715658021133525, + "grad_norm": 0.3010103404521942, + "learning_rate": 3.6093214060806686e-05, + "loss": 3.166419792175293, + "step": 8180 + }, + { + "epoch": 1.5725264169068205, + "grad_norm": 0.26971593499183655, + "learning_rate": 3.593833146255461e-05, + "loss": 3.1657012939453124, + "step": 8185 + }, + { + "epoch": 1.5734870317002883, + "grad_norm": 0.2859204411506653, + "learning_rate": 3.5783736653593546e-05, + "loss": 3.163465118408203, + "step": 8190 + }, + { + "epoch": 1.574447646493756, + "grad_norm": 0.2561207115650177, + "learning_rate": 3.56294300239821e-05, + "loss": 3.164503288269043, + "step": 8195 + }, + { + "epoch": 1.5754082612872238, + "grad_norm": 0.22760829329490662, + "learning_rate": 3.547541196305166e-05, + "loss": 3.1677932739257812, + "step": 8200 + }, + { + "epoch": 1.5763688760806915, + "grad_norm": 0.3261941969394684, + "learning_rate": 3.53216828594056e-05, + "loss": 3.164171600341797, + "step": 8205 + }, + { + "epoch": 1.5773294908741593, + "grad_norm": 0.3237496018409729, + "learning_rate": 3.5168243100918254e-05, + "loss": 3.1644121170043946, + "step": 8210 + }, + { + "epoch": 1.5782901056676273, + "grad_norm": 0.308450311422348, + "learning_rate": 3.501509307473391e-05, + "loss": 3.167286682128906, + "step": 8215 + }, + { + "epoch": 1.579250720461095, + "grad_norm": 0.2737598717212677, + "learning_rate": 3.486223316726569e-05, + "loss": 3.159839630126953, + "step": 8220 + }, + { + "epoch": 1.580211335254563, + "grad_norm": 0.2531875967979431, + "learning_rate": 3.470966376419489e-05, + "loss": 3.165401268005371, + "step": 8225 + }, + { + "epoch": 1.5811719500480308, + "grad_norm": 0.3886658847332001, + "learning_rate": 3.455738525046976e-05, + "loss": 3.164730453491211, + "step": 8230 + }, + { + "epoch": 1.5821325648414986, + "grad_norm": 0.32052338123321533, + "learning_rate": 3.440539801030463e-05, + "loss": 3.1636079788208007, + "step": 8235 + }, + { + "epoch": 1.5830931796349663, + "grad_norm": 0.3415239751338959, + "learning_rate": 3.425370242717887e-05, + "loss": 3.1651962280273436, + "step": 8240 + }, + { + "epoch": 1.584053794428434, + "grad_norm": 0.27457112073898315, + "learning_rate": 3.410229888383597e-05, + "loss": 3.169073486328125, + "step": 8245 + }, + { + "epoch": 1.585014409221902, + "grad_norm": 0.3607783615589142, + "learning_rate": 3.395118776228265e-05, + "loss": 3.1643707275390627, + "step": 8250 + }, + { + "epoch": 1.5859750240153698, + "grad_norm": 0.44193530082702637, + "learning_rate": 3.380036944378775e-05, + "loss": 3.165296936035156, + "step": 8255 + }, + { + "epoch": 1.5869356388088378, + "grad_norm": 0.27032527327537537, + "learning_rate": 3.364984430888138e-05, + "loss": 3.1653570175170898, + "step": 8260 + }, + { + "epoch": 1.5878962536023056, + "grad_norm": 0.3761611580848694, + "learning_rate": 3.34996127373539e-05, + "loss": 3.1637210845947266, + "step": 8265 + }, + { + "epoch": 1.5888568683957733, + "grad_norm": 0.2229749858379364, + "learning_rate": 3.3349675108254946e-05, + "loss": 3.165303421020508, + "step": 8270 + }, + { + "epoch": 1.589817483189241, + "grad_norm": 0.23744769394397736, + "learning_rate": 3.320003179989254e-05, + "loss": 3.16616268157959, + "step": 8275 + }, + { + "epoch": 1.5907780979827089, + "grad_norm": 0.28989988565444946, + "learning_rate": 3.305068318983211e-05, + "loss": 3.1658939361572265, + "step": 8280 + }, + { + "epoch": 1.5917387127761766, + "grad_norm": 0.2983841300010681, + "learning_rate": 3.290162965489558e-05, + "loss": 3.1631855010986327, + "step": 8285 + }, + { + "epoch": 1.5926993275696446, + "grad_norm": 0.2570306658744812, + "learning_rate": 3.275287157116021e-05, + "loss": 3.1675746917724608, + "step": 8290 + }, + { + "epoch": 1.5936599423631124, + "grad_norm": 0.3952568471431732, + "learning_rate": 3.260440931395794e-05, + "loss": 3.1651771545410154, + "step": 8295 + }, + { + "epoch": 1.5946205571565804, + "grad_norm": 0.3825446367263794, + "learning_rate": 3.24562432578743e-05, + "loss": 3.1654695510864257, + "step": 8300 + }, + { + "epoch": 1.5955811719500481, + "grad_norm": 0.2898649275302887, + "learning_rate": 3.230837377674746e-05, + "loss": 3.1621742248535156, + "step": 8305 + }, + { + "epoch": 1.5965417867435159, + "grad_norm": 0.24470216035842896, + "learning_rate": 3.216080124366724e-05, + "loss": 3.1656517028808593, + "step": 8310 + }, + { + "epoch": 1.5975024015369836, + "grad_norm": 0.23829184472560883, + "learning_rate": 3.20135260309743e-05, + "loss": 3.166254425048828, + "step": 8315 + }, + { + "epoch": 1.5984630163304514, + "grad_norm": 0.2703918218612671, + "learning_rate": 3.186654851025911e-05, + "loss": 3.1674873352050783, + "step": 8320 + }, + { + "epoch": 1.5994236311239192, + "grad_norm": 0.2798570394515991, + "learning_rate": 3.171986905236104e-05, + "loss": 3.16519718170166, + "step": 8325 + }, + { + "epoch": 1.6003842459173871, + "grad_norm": 0.29663723707199097, + "learning_rate": 3.15734880273674e-05, + "loss": 3.1651836395263673, + "step": 8330 + }, + { + "epoch": 1.601344860710855, + "grad_norm": 0.2850891351699829, + "learning_rate": 3.142740580461261e-05, + "loss": 3.1634387969970703, + "step": 8335 + }, + { + "epoch": 1.602305475504323, + "grad_norm": 0.3124440610408783, + "learning_rate": 3.128162275267697e-05, + "loss": 3.1671958923339845, + "step": 8340 + }, + { + "epoch": 1.6032660902977907, + "grad_norm": 0.3228466808795929, + "learning_rate": 3.113613923938614e-05, + "loss": 3.1625482559204103, + "step": 8345 + }, + { + "epoch": 1.6042267050912584, + "grad_norm": 0.23934929072856903, + "learning_rate": 3.0990955631809965e-05, + "loss": 3.1659454345703124, + "step": 8350 + }, + { + "epoch": 1.6051873198847262, + "grad_norm": 0.2973260283470154, + "learning_rate": 3.08460722962616e-05, + "loss": 3.167291259765625, + "step": 8355 + }, + { + "epoch": 1.606147934678194, + "grad_norm": 0.26656073331832886, + "learning_rate": 3.070148959829649e-05, + "loss": 3.165090560913086, + "step": 8360 + }, + { + "epoch": 1.6071085494716617, + "grad_norm": 0.25156885385513306, + "learning_rate": 3.055720790271164e-05, + "loss": 3.1673526763916016, + "step": 8365 + }, + { + "epoch": 1.6080691642651297, + "grad_norm": 0.2427086979150772, + "learning_rate": 3.0413227573544592e-05, + "loss": 3.167298698425293, + "step": 8370 + }, + { + "epoch": 1.6090297790585975, + "grad_norm": 0.33012548089027405, + "learning_rate": 3.026954897407252e-05, + "loss": 3.16772403717041, + "step": 8375 + }, + { + "epoch": 1.6099903938520654, + "grad_norm": 0.19765010476112366, + "learning_rate": 3.01261724668112e-05, + "loss": 3.1641408920288088, + "step": 8380 + }, + { + "epoch": 1.6109510086455332, + "grad_norm": 0.25813835859298706, + "learning_rate": 2.9983098413514284e-05, + "loss": 3.168695068359375, + "step": 8385 + }, + { + "epoch": 1.611911623439001, + "grad_norm": 0.2662206292152405, + "learning_rate": 2.9840327175172295e-05, + "loss": 3.1604537963867188, + "step": 8390 + }, + { + "epoch": 1.6128722382324687, + "grad_norm": 0.2579902410507202, + "learning_rate": 2.969785911201172e-05, + "loss": 3.1639808654785155, + "step": 8395 + }, + { + "epoch": 1.6138328530259365, + "grad_norm": 0.27471011877059937, + "learning_rate": 2.9555694583494095e-05, + "loss": 3.1644439697265625, + "step": 8400 + }, + { + "epoch": 1.6147934678194045, + "grad_norm": 0.2601883113384247, + "learning_rate": 2.9413833948315163e-05, + "loss": 3.1667598724365233, + "step": 8405 + }, + { + "epoch": 1.6157540826128722, + "grad_norm": 0.27512237429618835, + "learning_rate": 2.9272277564403746e-05, + "loss": 3.1641109466552733, + "step": 8410 + }, + { + "epoch": 1.6167146974063402, + "grad_norm": 0.25258708000183105, + "learning_rate": 2.9131025788921193e-05, + "loss": 3.1663232803344727, + "step": 8415 + }, + { + "epoch": 1.617675312199808, + "grad_norm": 0.22575309872627258, + "learning_rate": 2.8990078978260216e-05, + "loss": 3.166961669921875, + "step": 8420 + }, + { + "epoch": 1.6186359269932757, + "grad_norm": 0.24490588903427124, + "learning_rate": 2.8849437488044118e-05, + "loss": 3.1663583755493163, + "step": 8425 + }, + { + "epoch": 1.6195965417867435, + "grad_norm": 0.2155570238828659, + "learning_rate": 2.8709101673125728e-05, + "loss": 3.1656875610351562, + "step": 8430 + }, + { + "epoch": 1.6205571565802113, + "grad_norm": 0.2795376479625702, + "learning_rate": 2.856907188758674e-05, + "loss": 3.162942314147949, + "step": 8435 + }, + { + "epoch": 1.621517771373679, + "grad_norm": 0.3358766734600067, + "learning_rate": 2.8429348484736658e-05, + "loss": 3.1658599853515623, + "step": 8440 + }, + { + "epoch": 1.622478386167147, + "grad_norm": 0.2666330933570862, + "learning_rate": 2.8289931817112027e-05, + "loss": 3.1691190719604494, + "step": 8445 + }, + { + "epoch": 1.6234390009606148, + "grad_norm": 0.2384926676750183, + "learning_rate": 2.8150822236475283e-05, + "loss": 3.1659183502197266, + "step": 8450 + }, + { + "epoch": 1.6243996157540828, + "grad_norm": 0.25563499331474304, + "learning_rate": 2.8012020093814252e-05, + "loss": 3.1584651947021483, + "step": 8455 + }, + { + "epoch": 1.6253602305475505, + "grad_norm": 0.2299119532108307, + "learning_rate": 2.7873525739340973e-05, + "loss": 3.1645713806152345, + "step": 8460 + }, + { + "epoch": 1.6263208453410183, + "grad_norm": 0.26834285259246826, + "learning_rate": 2.773533952249088e-05, + "loss": 3.166657257080078, + "step": 8465 + }, + { + "epoch": 1.627281460134486, + "grad_norm": 0.2765657901763916, + "learning_rate": 2.7597461791921987e-05, + "loss": 3.165719223022461, + "step": 8470 + }, + { + "epoch": 1.6282420749279538, + "grad_norm": 0.26235702633857727, + "learning_rate": 2.745989289551401e-05, + "loss": 3.163796615600586, + "step": 8475 + }, + { + "epoch": 1.6292026897214216, + "grad_norm": 0.2585254907608032, + "learning_rate": 2.7322633180367286e-05, + "loss": 3.1652366638183596, + "step": 8480 + }, + { + "epoch": 1.6301633045148896, + "grad_norm": 0.24528291821479797, + "learning_rate": 2.7185682992802215e-05, + "loss": 3.165533447265625, + "step": 8485 + }, + { + "epoch": 1.6311239193083573, + "grad_norm": 0.24676673114299774, + "learning_rate": 2.7049042678358157e-05, + "loss": 3.162363624572754, + "step": 8490 + }, + { + "epoch": 1.6320845341018253, + "grad_norm": 0.28905969858169556, + "learning_rate": 2.6912712581792684e-05, + "loss": 3.165178680419922, + "step": 8495 + }, + { + "epoch": 1.633045148895293, + "grad_norm": 0.25018957257270813, + "learning_rate": 2.6776693047080546e-05, + "loss": 3.163674736022949, + "step": 8500 + }, + { + "epoch": 1.6340057636887608, + "grad_norm": 0.24032709002494812, + "learning_rate": 2.6640984417412996e-05, + "loss": 3.1674957275390625, + "step": 8505 + }, + { + "epoch": 1.6349663784822286, + "grad_norm": 0.2825995981693268, + "learning_rate": 2.6505587035196862e-05, + "loss": 3.164845085144043, + "step": 8510 + }, + { + "epoch": 1.6359269932756964, + "grad_norm": 0.3082216680049896, + "learning_rate": 2.6370501242053655e-05, + "loss": 3.164422607421875, + "step": 8515 + }, + { + "epoch": 1.6368876080691641, + "grad_norm": 0.2581336796283722, + "learning_rate": 2.6235727378818617e-05, + "loss": 3.1643039703369142, + "step": 8520 + }, + { + "epoch": 1.637848222862632, + "grad_norm": 0.2364456057548523, + "learning_rate": 2.6101265785540054e-05, + "loss": 3.1647056579589843, + "step": 8525 + }, + { + "epoch": 1.6388088376560999, + "grad_norm": 0.2648424804210663, + "learning_rate": 2.596711680147837e-05, + "loss": 3.1666595458984377, + "step": 8530 + }, + { + "epoch": 1.6397694524495678, + "grad_norm": 0.21114224195480347, + "learning_rate": 2.5833280765105218e-05, + "loss": 3.1652767181396486, + "step": 8535 + }, + { + "epoch": 1.6407300672430356, + "grad_norm": 0.29811277985572815, + "learning_rate": 2.5699758014102627e-05, + "loss": 3.1638210296630858, + "step": 8540 + }, + { + "epoch": 1.6416906820365034, + "grad_norm": 0.22697743773460388, + "learning_rate": 2.5566548885362233e-05, + "loss": 3.1660308837890625, + "step": 8545 + }, + { + "epoch": 1.6426512968299711, + "grad_norm": 0.2113056480884552, + "learning_rate": 2.543365371498434e-05, + "loss": 3.1632022857666016, + "step": 8550 + }, + { + "epoch": 1.643611911623439, + "grad_norm": 0.24464410543441772, + "learning_rate": 2.5301072838277026e-05, + "loss": 3.1658231735229494, + "step": 8555 + }, + { + "epoch": 1.6445725264169067, + "grad_norm": 0.23001989722251892, + "learning_rate": 2.5168806589755497e-05, + "loss": 3.163945960998535, + "step": 8560 + }, + { + "epoch": 1.6455331412103746, + "grad_norm": 0.20726712048053741, + "learning_rate": 2.503685530314109e-05, + "loss": 3.162770080566406, + "step": 8565 + }, + { + "epoch": 1.6464937560038426, + "grad_norm": 0.23915457725524902, + "learning_rate": 2.490521931136036e-05, + "loss": 3.1642734527587892, + "step": 8570 + }, + { + "epoch": 1.6474543707973104, + "grad_norm": 0.2551625370979309, + "learning_rate": 2.4773898946544473e-05, + "loss": 3.1642024993896483, + "step": 8575 + }, + { + "epoch": 1.6484149855907781, + "grad_norm": 0.2408856302499771, + "learning_rate": 2.4642894540028164e-05, + "loss": 3.1599807739257812, + "step": 8580 + }, + { + "epoch": 1.649375600384246, + "grad_norm": 0.2513860762119293, + "learning_rate": 2.4512206422349024e-05, + "loss": 3.163848876953125, + "step": 8585 + }, + { + "epoch": 1.6503362151777137, + "grad_norm": 0.21447288990020752, + "learning_rate": 2.438183492324654e-05, + "loss": 3.1619583129882813, + "step": 8590 + }, + { + "epoch": 1.6512968299711814, + "grad_norm": 0.1881914734840393, + "learning_rate": 2.4251780371661373e-05, + "loss": 3.1636112213134764, + "step": 8595 + }, + { + "epoch": 1.6522574447646494, + "grad_norm": 0.19982366263866425, + "learning_rate": 2.4122043095734518e-05, + "loss": 3.1659828186035157, + "step": 8600 + }, + { + "epoch": 1.6532180595581172, + "grad_norm": 0.2119777649641037, + "learning_rate": 2.3992623422806444e-05, + "loss": 3.161235809326172, + "step": 8605 + }, + { + "epoch": 1.6541786743515852, + "grad_norm": 0.2393619865179062, + "learning_rate": 2.3863521679416237e-05, + "loss": 3.1633071899414062, + "step": 8610 + }, + { + "epoch": 1.655139289145053, + "grad_norm": 0.24367739260196686, + "learning_rate": 2.3734738191300862e-05, + "loss": 3.16229133605957, + "step": 8615 + }, + { + "epoch": 1.6560999039385207, + "grad_norm": 0.29672330617904663, + "learning_rate": 2.36062732833943e-05, + "loss": 3.1618356704711914, + "step": 8620 + }, + { + "epoch": 1.6570605187319885, + "grad_norm": 0.24469807744026184, + "learning_rate": 2.347812727982661e-05, + "loss": 3.161136245727539, + "step": 8625 + }, + { + "epoch": 1.6580211335254562, + "grad_norm": 0.21228572726249695, + "learning_rate": 2.3350300503923352e-05, + "loss": 3.1614845275878904, + "step": 8630 + }, + { + "epoch": 1.658981748318924, + "grad_norm": 0.20836694538593292, + "learning_rate": 2.3222793278204626e-05, + "loss": 3.1641204833984373, + "step": 8635 + }, + { + "epoch": 1.659942363112392, + "grad_norm": 0.2582319974899292, + "learning_rate": 2.309560592438417e-05, + "loss": 3.1627979278564453, + "step": 8640 + }, + { + "epoch": 1.6609029779058597, + "grad_norm": 0.23718759417533875, + "learning_rate": 2.2968738763368765e-05, + "loss": 3.1604846954345702, + "step": 8645 + }, + { + "epoch": 1.6618635926993277, + "grad_norm": 0.19198426604270935, + "learning_rate": 2.2842192115257295e-05, + "loss": 3.1641172409057616, + "step": 8650 + }, + { + "epoch": 1.6628242074927955, + "grad_norm": 0.258806049823761, + "learning_rate": 2.271596629933992e-05, + "loss": 3.16113224029541, + "step": 8655 + }, + { + "epoch": 1.6637848222862632, + "grad_norm": 0.21029417216777802, + "learning_rate": 2.25900616340973e-05, + "loss": 3.1632869720458983, + "step": 8660 + }, + { + "epoch": 1.664745437079731, + "grad_norm": 0.21884505450725555, + "learning_rate": 2.2464478437199808e-05, + "loss": 3.163255310058594, + "step": 8665 + }, + { + "epoch": 1.6657060518731988, + "grad_norm": 0.22930586338043213, + "learning_rate": 2.2339217025506813e-05, + "loss": 3.163426399230957, + "step": 8670 + }, + { + "epoch": 1.6666666666666665, + "grad_norm": 0.2358139008283615, + "learning_rate": 2.2214277715065636e-05, + "loss": 3.1608362197875977, + "step": 8675 + }, + { + "epoch": 1.6676272814601345, + "grad_norm": 0.22611092031002045, + "learning_rate": 2.2089660821110988e-05, + "loss": 3.1634220123291015, + "step": 8680 + }, + { + "epoch": 1.6685878962536023, + "grad_norm": 0.19509707391262054, + "learning_rate": 2.1965366658064086e-05, + "loss": 3.158903884887695, + "step": 8685 + }, + { + "epoch": 1.6695485110470702, + "grad_norm": 0.19261038303375244, + "learning_rate": 2.1841395539531893e-05, + "loss": 3.161344528198242, + "step": 8690 + }, + { + "epoch": 1.670509125840538, + "grad_norm": 0.1918260157108307, + "learning_rate": 2.1717747778306177e-05, + "loss": 3.1661489486694334, + "step": 8695 + }, + { + "epoch": 1.6714697406340058, + "grad_norm": 0.20543381571769714, + "learning_rate": 2.159442368636297e-05, + "loss": 3.1635005950927733, + "step": 8700 + }, + { + "epoch": 1.6724303554274735, + "grad_norm": 0.2433311641216278, + "learning_rate": 2.147142357486164e-05, + "loss": 3.163587951660156, + "step": 8705 + }, + { + "epoch": 1.6733909702209413, + "grad_norm": 0.21309548616409302, + "learning_rate": 2.1348747754144004e-05, + "loss": 3.1650224685668946, + "step": 8710 + }, + { + "epoch": 1.674351585014409, + "grad_norm": 0.2069951891899109, + "learning_rate": 2.1226396533733796e-05, + "loss": 3.164658546447754, + "step": 8715 + }, + { + "epoch": 1.675312199807877, + "grad_norm": 0.24692699313163757, + "learning_rate": 2.1104370222335688e-05, + "loss": 3.1658843994140624, + "step": 8720 + }, + { + "epoch": 1.6762728146013448, + "grad_norm": 0.23039954900741577, + "learning_rate": 2.0982669127834622e-05, + "loss": 3.1638277053833006, + "step": 8725 + }, + { + "epoch": 1.6772334293948128, + "grad_norm": 0.2396543025970459, + "learning_rate": 2.0861293557294862e-05, + "loss": 3.162588119506836, + "step": 8730 + }, + { + "epoch": 1.6781940441882806, + "grad_norm": 0.2332344651222229, + "learning_rate": 2.0740243816959452e-05, + "loss": 3.163272476196289, + "step": 8735 + }, + { + "epoch": 1.6791546589817483, + "grad_norm": 0.21478639543056488, + "learning_rate": 2.061952021224938e-05, + "loss": 3.162495803833008, + "step": 8740 + }, + { + "epoch": 1.680115273775216, + "grad_norm": 0.19731956720352173, + "learning_rate": 2.0499123047762576e-05, + "loss": 3.162389945983887, + "step": 8745 + }, + { + "epoch": 1.6810758885686838, + "grad_norm": 0.19572487473487854, + "learning_rate": 2.03790526272735e-05, + "loss": 3.1616092681884767, + "step": 8750 + }, + { + "epoch": 1.6820365033621518, + "grad_norm": 0.184941828250885, + "learning_rate": 2.025930925373213e-05, + "loss": 3.1640493392944338, + "step": 8755 + }, + { + "epoch": 1.6829971181556196, + "grad_norm": 0.1868741363286972, + "learning_rate": 2.013989322926331e-05, + "loss": 3.164847183227539, + "step": 8760 + }, + { + "epoch": 1.6839577329490876, + "grad_norm": 0.19907650351524353, + "learning_rate": 2.0020804855165857e-05, + "loss": 3.164986801147461, + "step": 8765 + }, + { + "epoch": 1.6849183477425553, + "grad_norm": 0.19891948997974396, + "learning_rate": 1.9902044431912e-05, + "loss": 3.16265869140625, + "step": 8770 + }, + { + "epoch": 1.685878962536023, + "grad_norm": 0.23641318082809448, + "learning_rate": 1.9783612259146485e-05, + "loss": 3.164572525024414, + "step": 8775 + }, + { + "epoch": 1.6868395773294909, + "grad_norm": 0.2543870210647583, + "learning_rate": 1.9665508635685767e-05, + "loss": 3.1619945526123048, + "step": 8780 + }, + { + "epoch": 1.6878001921229586, + "grad_norm": 0.1898890733718872, + "learning_rate": 1.9547733859517443e-05, + "loss": 3.161852264404297, + "step": 8785 + }, + { + "epoch": 1.6887608069164264, + "grad_norm": 0.18471160531044006, + "learning_rate": 1.9430288227799367e-05, + "loss": 3.167014312744141, + "step": 8790 + }, + { + "epoch": 1.6897214217098944, + "grad_norm": 0.1837644726037979, + "learning_rate": 1.931317203685893e-05, + "loss": 3.1675960540771486, + "step": 8795 + }, + { + "epoch": 1.6906820365033621, + "grad_norm": 0.16458193957805634, + "learning_rate": 1.9196385582192247e-05, + "loss": 3.161328125, + "step": 8800 + }, + { + "epoch": 1.6916426512968301, + "grad_norm": 0.18657121062278748, + "learning_rate": 1.9079929158463526e-05, + "loss": 3.163207244873047, + "step": 8805 + }, + { + "epoch": 1.6926032660902979, + "grad_norm": 0.2095474749803543, + "learning_rate": 1.896380305950434e-05, + "loss": 3.1646053314208986, + "step": 8810 + }, + { + "epoch": 1.6935638808837656, + "grad_norm": 0.1963958740234375, + "learning_rate": 1.8848007578312686e-05, + "loss": 3.164917755126953, + "step": 8815 + }, + { + "epoch": 1.6945244956772334, + "grad_norm": 0.18170610070228577, + "learning_rate": 1.8732543007052452e-05, + "loss": 3.1576461791992188, + "step": 8820 + }, + { + "epoch": 1.6954851104707012, + "grad_norm": 0.2065751552581787, + "learning_rate": 1.8617409637052606e-05, + "loss": 3.166869354248047, + "step": 8825 + }, + { + "epoch": 1.696445725264169, + "grad_norm": 0.18570922315120697, + "learning_rate": 1.8502607758806487e-05, + "loss": 3.162985992431641, + "step": 8830 + }, + { + "epoch": 1.697406340057637, + "grad_norm": 0.22870124876499176, + "learning_rate": 1.8388137661970948e-05, + "loss": 3.165153694152832, + "step": 8835 + }, + { + "epoch": 1.6983669548511047, + "grad_norm": 0.18405385315418243, + "learning_rate": 1.8273999635365806e-05, + "loss": 3.162175750732422, + "step": 8840 + }, + { + "epoch": 1.6993275696445727, + "grad_norm": 0.1959793120622635, + "learning_rate": 1.8160193966973047e-05, + "loss": 3.161460113525391, + "step": 8845 + }, + { + "epoch": 1.7002881844380404, + "grad_norm": 0.20207500457763672, + "learning_rate": 1.8046720943936e-05, + "loss": 3.1632549285888674, + "step": 8850 + }, + { + "epoch": 1.7012487992315082, + "grad_norm": 0.20162688195705414, + "learning_rate": 1.7933580852558742e-05, + "loss": 3.1618804931640625, + "step": 8855 + }, + { + "epoch": 1.702209414024976, + "grad_norm": 0.18540622293949127, + "learning_rate": 1.7820773978305365e-05, + "loss": 3.161595916748047, + "step": 8860 + }, + { + "epoch": 1.7031700288184437, + "grad_norm": 0.19820909202098846, + "learning_rate": 1.7708300605799202e-05, + "loss": 3.1606048583984374, + "step": 8865 + }, + { + "epoch": 1.7041306436119115, + "grad_norm": 0.2218032330274582, + "learning_rate": 1.7596161018822007e-05, + "loss": 3.1620101928710938, + "step": 8870 + }, + { + "epoch": 1.7050912584053795, + "grad_norm": 0.1894785612821579, + "learning_rate": 1.7484355500313568e-05, + "loss": 3.161252975463867, + "step": 8875 + }, + { + "epoch": 1.7060518731988472, + "grad_norm": 0.20882797241210938, + "learning_rate": 1.7372884332370677e-05, + "loss": 3.162582015991211, + "step": 8880 + }, + { + "epoch": 1.7070124879923152, + "grad_norm": 0.2097301483154297, + "learning_rate": 1.7261747796246478e-05, + "loss": 3.162227249145508, + "step": 8885 + }, + { + "epoch": 1.707973102785783, + "grad_norm": 0.18620242178440094, + "learning_rate": 1.7150946172349844e-05, + "loss": 3.163176345825195, + "step": 8890 + }, + { + "epoch": 1.7089337175792507, + "grad_norm": 0.1744890809059143, + "learning_rate": 1.7040479740244684e-05, + "loss": 3.1600067138671877, + "step": 8895 + }, + { + "epoch": 1.7098943323727185, + "grad_norm": 0.20441320538520813, + "learning_rate": 1.693034877864913e-05, + "loss": 3.161859130859375, + "step": 8900 + }, + { + "epoch": 1.7108549471661862, + "grad_norm": 0.18480157852172852, + "learning_rate": 1.682055356543487e-05, + "loss": 3.1619277954101563, + "step": 8905 + }, + { + "epoch": 1.7118155619596542, + "grad_norm": 0.19677984714508057, + "learning_rate": 1.6711094377626495e-05, + "loss": 3.1647516250610352, + "step": 8910 + }, + { + "epoch": 1.712776176753122, + "grad_norm": 0.19380095601081848, + "learning_rate": 1.66019714914008e-05, + "loss": 3.1602848052978514, + "step": 8915 + }, + { + "epoch": 1.71373679154659, + "grad_norm": 0.17474035918712616, + "learning_rate": 1.6493185182085967e-05, + "loss": 3.164753532409668, + "step": 8920 + }, + { + "epoch": 1.7146974063400577, + "grad_norm": 0.18277983367443085, + "learning_rate": 1.6384735724161045e-05, + "loss": 3.162521743774414, + "step": 8925 + }, + { + "epoch": 1.7156580211335255, + "grad_norm": 0.17841650545597076, + "learning_rate": 1.6276623391255146e-05, + "loss": 3.1669170379638674, + "step": 8930 + }, + { + "epoch": 1.7166186359269933, + "grad_norm": 0.18692415952682495, + "learning_rate": 1.6168848456146793e-05, + "loss": 3.1644006729125977, + "step": 8935 + }, + { + "epoch": 1.717579250720461, + "grad_norm": 0.22790934145450592, + "learning_rate": 1.606141119076314e-05, + "loss": 3.164959716796875, + "step": 8940 + }, + { + "epoch": 1.7185398655139288, + "grad_norm": 0.19892485439777374, + "learning_rate": 1.595431186617948e-05, + "loss": 3.1620708465576173, + "step": 8945 + }, + { + "epoch": 1.7195004803073968, + "grad_norm": 0.19590741395950317, + "learning_rate": 1.5847550752618427e-05, + "loss": 3.158066177368164, + "step": 8950 + }, + { + "epoch": 1.7204610951008645, + "grad_norm": 0.2052772343158722, + "learning_rate": 1.5741128119449153e-05, + "loss": 3.161151885986328, + "step": 8955 + }, + { + "epoch": 1.7214217098943325, + "grad_norm": 0.18683621287345886, + "learning_rate": 1.5635044235186906e-05, + "loss": 3.1654678344726563, + "step": 8960 + }, + { + "epoch": 1.7223823246878003, + "grad_norm": 0.19518622756004333, + "learning_rate": 1.5529299367492208e-05, + "loss": 3.160481262207031, + "step": 8965 + }, + { + "epoch": 1.723342939481268, + "grad_norm": 0.1885916143655777, + "learning_rate": 1.5423893783170262e-05, + "loss": 3.162572479248047, + "step": 8970 + }, + { + "epoch": 1.7243035542747358, + "grad_norm": 0.18152864277362823, + "learning_rate": 1.531882774817007e-05, + "loss": 3.1627574920654298, + "step": 8975 + }, + { + "epoch": 1.7252641690682036, + "grad_norm": 0.17760129272937775, + "learning_rate": 1.5214101527584071e-05, + "loss": 3.1614215850830076, + "step": 8980 + }, + { + "epoch": 1.7262247838616713, + "grad_norm": 0.20811888575553894, + "learning_rate": 1.5109715385647297e-05, + "loss": 3.163728141784668, + "step": 8985 + }, + { + "epoch": 1.7271853986551393, + "grad_norm": 0.18515148758888245, + "learning_rate": 1.5005669585736618e-05, + "loss": 3.161965179443359, + "step": 8990 + }, + { + "epoch": 1.728146013448607, + "grad_norm": 0.1933458298444748, + "learning_rate": 1.4901964390370308e-05, + "loss": 3.163407325744629, + "step": 8995 + }, + { + "epoch": 1.729106628242075, + "grad_norm": 0.20849090814590454, + "learning_rate": 1.4798600061207195e-05, + "loss": 3.163631057739258, + "step": 9000 + }, + { + "epoch": 1.7300672430355428, + "grad_norm": 0.18622250854969025, + "learning_rate": 1.4695576859046127e-05, + "loss": 3.159561538696289, + "step": 9005 + }, + { + "epoch": 1.7310278578290106, + "grad_norm": 0.1835029423236847, + "learning_rate": 1.4592895043825126e-05, + "loss": 3.163999557495117, + "step": 9010 + }, + { + "epoch": 1.7319884726224783, + "grad_norm": 0.1808691918849945, + "learning_rate": 1.449055487462102e-05, + "loss": 3.1594869613647463, + "step": 9015 + }, + { + "epoch": 1.732949087415946, + "grad_norm": 0.16524961590766907, + "learning_rate": 1.4388556609648572e-05, + "loss": 3.163185691833496, + "step": 9020 + }, + { + "epoch": 1.7339097022094139, + "grad_norm": 0.16896623373031616, + "learning_rate": 1.428690050625979e-05, + "loss": 3.1632381439208985, + "step": 9025 + }, + { + "epoch": 1.7348703170028819, + "grad_norm": 0.1609111875295639, + "learning_rate": 1.4185586820943506e-05, + "loss": 3.1643299102783202, + "step": 9030 + }, + { + "epoch": 1.7358309317963496, + "grad_norm": 0.16877403855323792, + "learning_rate": 1.4084615809324523e-05, + "loss": 3.164197540283203, + "step": 9035 + }, + { + "epoch": 1.7367915465898176, + "grad_norm": 0.17374147474765778, + "learning_rate": 1.3983987726163087e-05, + "loss": 3.1597652435302734, + "step": 9040 + }, + { + "epoch": 1.7377521613832854, + "grad_norm": 0.1869489699602127, + "learning_rate": 1.3883702825354138e-05, + "loss": 3.160333251953125, + "step": 9045 + }, + { + "epoch": 1.7387127761767531, + "grad_norm": 0.19941848516464233, + "learning_rate": 1.3783761359926771e-05, + "loss": 3.1596229553222654, + "step": 9050 + }, + { + "epoch": 1.739673390970221, + "grad_norm": 0.16694855690002441, + "learning_rate": 1.3684163582043595e-05, + "loss": 3.159475898742676, + "step": 9055 + }, + { + "epoch": 1.7406340057636887, + "grad_norm": 0.1921820491552353, + "learning_rate": 1.3584909742999978e-05, + "loss": 3.162578010559082, + "step": 9060 + }, + { + "epoch": 1.7415946205571564, + "grad_norm": 0.17421427369117737, + "learning_rate": 1.3486000093223565e-05, + "loss": 3.1633106231689454, + "step": 9065 + }, + { + "epoch": 1.7425552353506244, + "grad_norm": 0.16822272539138794, + "learning_rate": 1.3387434882273529e-05, + "loss": 3.1628774642944335, + "step": 9070 + }, + { + "epoch": 1.7435158501440924, + "grad_norm": 0.2147689312696457, + "learning_rate": 1.3289214358840022e-05, + "loss": 3.164061737060547, + "step": 9075 + }, + { + "epoch": 1.7444764649375601, + "grad_norm": 0.19877059757709503, + "learning_rate": 1.3191338770743493e-05, + "loss": 3.164179039001465, + "step": 9080 + }, + { + "epoch": 1.745437079731028, + "grad_norm": 0.17470191419124603, + "learning_rate": 1.30938083649341e-05, + "loss": 3.162880706787109, + "step": 9085 + }, + { + "epoch": 1.7463976945244957, + "grad_norm": 0.17637498676776886, + "learning_rate": 1.2996623387491085e-05, + "loss": 3.163307952880859, + "step": 9090 + }, + { + "epoch": 1.7473583093179634, + "grad_norm": 0.17012238502502441, + "learning_rate": 1.2899784083622067e-05, + "loss": 3.1613311767578125, + "step": 9095 + }, + { + "epoch": 1.7483189241114312, + "grad_norm": 0.15726859867572784, + "learning_rate": 1.2803290697662566e-05, + "loss": 3.161721038818359, + "step": 9100 + }, + { + "epoch": 1.7492795389048992, + "grad_norm": 0.1671910136938095, + "learning_rate": 1.2707143473075299e-05, + "loss": 3.1644134521484375, + "step": 9105 + }, + { + "epoch": 1.750240153698367, + "grad_norm": 0.17212195694446564, + "learning_rate": 1.2611342652449597e-05, + "loss": 3.1598962783813476, + "step": 9110 + }, + { + "epoch": 1.751200768491835, + "grad_norm": 0.17955084145069122, + "learning_rate": 1.2515888477500708e-05, + "loss": 3.161183166503906, + "step": 9115 + }, + { + "epoch": 1.7521613832853027, + "grad_norm": 0.15037564933300018, + "learning_rate": 1.2420781189069346e-05, + "loss": 3.164091873168945, + "step": 9120 + }, + { + "epoch": 1.7531219980787704, + "grad_norm": 0.16244594752788544, + "learning_rate": 1.2326021027120958e-05, + "loss": 3.162504196166992, + "step": 9125 + }, + { + "epoch": 1.7540826128722382, + "grad_norm": 0.1728479117155075, + "learning_rate": 1.2231608230745128e-05, + "loss": 3.159666633605957, + "step": 9130 + }, + { + "epoch": 1.755043227665706, + "grad_norm": 0.17677444219589233, + "learning_rate": 1.2137543038155034e-05, + "loss": 3.16094970703125, + "step": 9135 + }, + { + "epoch": 1.7560038424591737, + "grad_norm": 0.16001811623573303, + "learning_rate": 1.2043825686686798e-05, + "loss": 3.1602264404296876, + "step": 9140 + }, + { + "epoch": 1.7569644572526417, + "grad_norm": 0.1979171186685562, + "learning_rate": 1.1950456412798954e-05, + "loss": 3.160991096496582, + "step": 9145 + }, + { + "epoch": 1.7579250720461095, + "grad_norm": 0.18909943103790283, + "learning_rate": 1.1857435452071706e-05, + "loss": 3.164293098449707, + "step": 9150 + }, + { + "epoch": 1.7588856868395775, + "grad_norm": 0.1578613966703415, + "learning_rate": 1.1764763039206516e-05, + "loss": 3.1651378631591798, + "step": 9155 + }, + { + "epoch": 1.7598463016330452, + "grad_norm": 0.17254026234149933, + "learning_rate": 1.1672439408025409e-05, + "loss": 3.1591054916381838, + "step": 9160 + }, + { + "epoch": 1.760806916426513, + "grad_norm": 0.1548478901386261, + "learning_rate": 1.1580464791470317e-05, + "loss": 3.162530517578125, + "step": 9165 + }, + { + "epoch": 1.7617675312199808, + "grad_norm": 0.16993309557437897, + "learning_rate": 1.1488839421602663e-05, + "loss": 3.1617578506469726, + "step": 9170 + }, + { + "epoch": 1.7627281460134485, + "grad_norm": 0.18223033845424652, + "learning_rate": 1.1397563529602654e-05, + "loss": 3.1608226776123045, + "step": 9175 + }, + { + "epoch": 1.7636887608069163, + "grad_norm": 0.15401305258274078, + "learning_rate": 1.130663734576877e-05, + "loss": 3.1587234497070313, + "step": 9180 + }, + { + "epoch": 1.7646493756003843, + "grad_norm": 0.19235272705554962, + "learning_rate": 1.1216061099517016e-05, + "loss": 3.158829116821289, + "step": 9185 + }, + { + "epoch": 1.765609990393852, + "grad_norm": 0.1655486822128296, + "learning_rate": 1.1125835019380614e-05, + "loss": 3.160569190979004, + "step": 9190 + }, + { + "epoch": 1.76657060518732, + "grad_norm": 0.17558416724205017, + "learning_rate": 1.1035959333009231e-05, + "loss": 3.1613176345825194, + "step": 9195 + }, + { + "epoch": 1.7675312199807878, + "grad_norm": 0.16277319192886353, + "learning_rate": 1.0946434267168375e-05, + "loss": 3.1619541168212892, + "step": 9200 + }, + { + "epoch": 1.7684918347742555, + "grad_norm": 0.15174464881420135, + "learning_rate": 1.0857260047739025e-05, + "loss": 3.1618457794189454, + "step": 9205 + }, + { + "epoch": 1.7694524495677233, + "grad_norm": 0.14863047003746033, + "learning_rate": 1.076843689971687e-05, + "loss": 3.1616443634033202, + "step": 9210 + }, + { + "epoch": 1.770413064361191, + "grad_norm": 0.16424240171909332, + "learning_rate": 1.0679965047211841e-05, + "loss": 3.1608753204345703, + "step": 9215 + }, + { + "epoch": 1.7713736791546588, + "grad_norm": 0.15550029277801514, + "learning_rate": 1.0591844713447479e-05, + "loss": 3.15771427154541, + "step": 9220 + }, + { + "epoch": 1.7723342939481268, + "grad_norm": 0.15241944789886475, + "learning_rate": 1.0504076120760413e-05, + "loss": 3.1600357055664063, + "step": 9225 + }, + { + "epoch": 1.7732949087415946, + "grad_norm": 0.15490137040615082, + "learning_rate": 1.0416659490599871e-05, + "loss": 3.163216400146484, + "step": 9230 + }, + { + "epoch": 1.7742555235350626, + "grad_norm": 0.17097872495651245, + "learning_rate": 1.0329595043526905e-05, + "loss": 3.164168930053711, + "step": 9235 + }, + { + "epoch": 1.7752161383285303, + "grad_norm": 0.19601082801818848, + "learning_rate": 1.024288299921408e-05, + "loss": 3.1567230224609375, + "step": 9240 + }, + { + "epoch": 1.776176753121998, + "grad_norm": 0.1483290195465088, + "learning_rate": 1.015652357644477e-05, + "loss": 3.1643510818481446, + "step": 9245 + }, + { + "epoch": 1.7771373679154658, + "grad_norm": 0.1415608525276184, + "learning_rate": 1.0070516993112676e-05, + "loss": 3.157514953613281, + "step": 9250 + }, + { + "epoch": 1.7780979827089336, + "grad_norm": 0.14610423147678375, + "learning_rate": 9.984863466221199e-06, + "loss": 3.1627222061157227, + "step": 9255 + }, + { + "epoch": 1.7790585975024016, + "grad_norm": 0.17516079545021057, + "learning_rate": 9.899563211883e-06, + "loss": 3.1590530395507814, + "step": 9260 + }, + { + "epoch": 1.7800192122958693, + "grad_norm": 0.1540340632200241, + "learning_rate": 9.814616445319384e-06, + "loss": 3.1593671798706056, + "step": 9265 + }, + { + "epoch": 1.7809798270893373, + "grad_norm": 0.18057873845100403, + "learning_rate": 9.730023380859725e-06, + "loss": 3.162105751037598, + "step": 9270 + }, + { + "epoch": 1.781940441882805, + "grad_norm": 0.17139260470867157, + "learning_rate": 9.645784231941005e-06, + "loss": 3.1630107879638674, + "step": 9275 + }, + { + "epoch": 1.7829010566762729, + "grad_norm": 0.16848276555538177, + "learning_rate": 9.561899211107244e-06, + "loss": 3.1598880767822264, + "step": 9280 + }, + { + "epoch": 1.7838616714697406, + "grad_norm": 0.16549763083457947, + "learning_rate": 9.478368530008967e-06, + "loss": 3.160906219482422, + "step": 9285 + }, + { + "epoch": 1.7848222862632084, + "grad_norm": 0.15802277624607086, + "learning_rate": 9.39519239940264e-06, + "loss": 3.161147689819336, + "step": 9290 + }, + { + "epoch": 1.7857829010566761, + "grad_norm": 0.1761743128299713, + "learning_rate": 9.31237102915019e-06, + "loss": 3.158751678466797, + "step": 9295 + }, + { + "epoch": 1.7867435158501441, + "grad_norm": 0.1671990603208542, + "learning_rate": 9.229904628218427e-06, + "loss": 3.1624494552612306, + "step": 9300 + }, + { + "epoch": 1.7877041306436119, + "grad_norm": 0.19958704710006714, + "learning_rate": 9.14779340467851e-06, + "loss": 3.1618110656738283, + "step": 9305 + }, + { + "epoch": 1.7886647454370799, + "grad_norm": 0.16844980418682098, + "learning_rate": 9.06603756570552e-06, + "loss": 3.156951141357422, + "step": 9310 + }, + { + "epoch": 1.7896253602305476, + "grad_norm": 0.14970090985298157, + "learning_rate": 8.984637317577782e-06, + "loss": 3.1624423980712892, + "step": 9315 + }, + { + "epoch": 1.7905859750240154, + "grad_norm": 0.15226224064826965, + "learning_rate": 8.90359286567654e-06, + "loss": 3.1624835968017577, + "step": 9320 + }, + { + "epoch": 1.7915465898174832, + "grad_norm": 0.1699807345867157, + "learning_rate": 8.822904414485194e-06, + "loss": 3.1621606826782225, + "step": 9325 + }, + { + "epoch": 1.792507204610951, + "grad_norm": 0.15113097429275513, + "learning_rate": 8.742572167589008e-06, + "loss": 3.16312255859375, + "step": 9330 + }, + { + "epoch": 1.7934678194044187, + "grad_norm": 0.16513291001319885, + "learning_rate": 8.662596327674499e-06, + "loss": 3.1624629974365233, + "step": 9335 + }, + { + "epoch": 1.7944284341978867, + "grad_norm": 0.154737189412117, + "learning_rate": 8.582977096528887e-06, + "loss": 3.1638771057128907, + "step": 9340 + }, + { + "epoch": 1.7953890489913544, + "grad_norm": 0.18483127653598785, + "learning_rate": 8.503714675039663e-06, + "loss": 3.16489315032959, + "step": 9345 + }, + { + "epoch": 1.7963496637848224, + "grad_norm": 0.16609624028205872, + "learning_rate": 8.424809263194054e-06, + "loss": 3.1592342376708986, + "step": 9350 + }, + { + "epoch": 1.7973102785782902, + "grad_norm": 0.13648296892642975, + "learning_rate": 8.346261060078524e-06, + "loss": 3.163716506958008, + "step": 9355 + }, + { + "epoch": 1.798270893371758, + "grad_norm": 0.16231390833854675, + "learning_rate": 8.268070263878223e-06, + "loss": 3.1625925064086915, + "step": 9360 + }, + { + "epoch": 1.7992315081652257, + "grad_norm": 0.15581952035427094, + "learning_rate": 8.190237071876576e-06, + "loss": 3.161655235290527, + "step": 9365 + }, + { + "epoch": 1.8001921229586935, + "grad_norm": 0.15568803250789642, + "learning_rate": 8.11276168045471e-06, + "loss": 3.1620370864868166, + "step": 9370 + }, + { + "epoch": 1.8011527377521612, + "grad_norm": 0.1662697196006775, + "learning_rate": 8.035644285090958e-06, + "loss": 3.1623306274414062, + "step": 9375 + }, + { + "epoch": 1.8021133525456292, + "grad_norm": 0.16496489942073822, + "learning_rate": 7.958885080360445e-06, + "loss": 3.1613933563232424, + "step": 9380 + }, + { + "epoch": 1.803073967339097, + "grad_norm": 0.15053115785121918, + "learning_rate": 7.882484259934497e-06, + "loss": 3.1627952575683596, + "step": 9385 + }, + { + "epoch": 1.804034582132565, + "grad_norm": 0.14695799350738525, + "learning_rate": 7.806442016580267e-06, + "loss": 3.160305404663086, + "step": 9390 + }, + { + "epoch": 1.8049951969260327, + "grad_norm": 0.15005071461200714, + "learning_rate": 7.730758542160064e-06, + "loss": 3.1611255645751952, + "step": 9395 + }, + { + "epoch": 1.8059558117195005, + "grad_norm": 0.15226683020591736, + "learning_rate": 7.655434027631124e-06, + "loss": 3.1629497528076174, + "step": 9400 + }, + { + "epoch": 1.8069164265129682, + "grad_norm": 0.15546758472919464, + "learning_rate": 7.580468663044903e-06, + "loss": 3.1586400985717775, + "step": 9405 + }, + { + "epoch": 1.807877041306436, + "grad_norm": 0.1452452838420868, + "learning_rate": 7.505862637546684e-06, + "loss": 3.160877227783203, + "step": 9410 + }, + { + "epoch": 1.808837656099904, + "grad_norm": 0.13006171584129333, + "learning_rate": 7.43161613937514e-06, + "loss": 3.1579397201538084, + "step": 9415 + }, + { + "epoch": 1.8097982708933718, + "grad_norm": 0.15625065565109253, + "learning_rate": 7.3577293558618075e-06, + "loss": 3.1598045349121096, + "step": 9420 + }, + { + "epoch": 1.8107588856868397, + "grad_norm": 0.16364313662052155, + "learning_rate": 7.284202473430612e-06, + "loss": 3.1616336822509767, + "step": 9425 + }, + { + "epoch": 1.8117195004803075, + "grad_norm": 0.15080556273460388, + "learning_rate": 7.211035677597421e-06, + "loss": 3.1628395080566407, + "step": 9430 + }, + { + "epoch": 1.8126801152737753, + "grad_norm": 0.1453072428703308, + "learning_rate": 7.138229152969582e-06, + "loss": 3.161026382446289, + "step": 9435 + }, + { + "epoch": 1.813640730067243, + "grad_norm": 0.1453263759613037, + "learning_rate": 7.065783083245435e-06, + "loss": 3.16085262298584, + "step": 9440 + }, + { + "epoch": 1.8146013448607108, + "grad_norm": 0.13254213333129883, + "learning_rate": 6.993697651213798e-06, + "loss": 3.1604400634765626, + "step": 9445 + }, + { + "epoch": 1.8155619596541785, + "grad_norm": 0.1310938000679016, + "learning_rate": 6.921973038753647e-06, + "loss": 3.158857727050781, + "step": 9450 + }, + { + "epoch": 1.8165225744476465, + "grad_norm": 0.13702644407749176, + "learning_rate": 6.8506094268335255e-06, + "loss": 3.16072998046875, + "step": 9455 + }, + { + "epoch": 1.8174831892411143, + "grad_norm": 0.15652552247047424, + "learning_rate": 6.779606995511183e-06, + "loss": 3.161703109741211, + "step": 9460 + }, + { + "epoch": 1.8184438040345823, + "grad_norm": 0.1748208999633789, + "learning_rate": 6.7089659239329685e-06, + "loss": 3.162480926513672, + "step": 9465 + }, + { + "epoch": 1.81940441882805, + "grad_norm": 0.15371812880039215, + "learning_rate": 6.638686390333592e-06, + "loss": 3.1619606018066406, + "step": 9470 + }, + { + "epoch": 1.8203650336215178, + "grad_norm": 0.15862195193767548, + "learning_rate": 6.56876857203556e-06, + "loss": 3.1612228393554687, + "step": 9475 + }, + { + "epoch": 1.8213256484149856, + "grad_norm": 0.14958879351615906, + "learning_rate": 6.499212645448626e-06, + "loss": 3.161460113525391, + "step": 9480 + }, + { + "epoch": 1.8222862632084533, + "grad_norm": 0.14320014417171478, + "learning_rate": 6.430018786069574e-06, + "loss": 3.1611602783203123, + "step": 9485 + }, + { + "epoch": 1.823246878001921, + "grad_norm": 0.14477957785129547, + "learning_rate": 6.361187168481618e-06, + "loss": 3.162501907348633, + "step": 9490 + }, + { + "epoch": 1.824207492795389, + "grad_norm": 0.14906539022922516, + "learning_rate": 6.292717966353955e-06, + "loss": 3.1564708709716798, + "step": 9495 + }, + { + "epoch": 1.8251681075888568, + "grad_norm": 0.1561676263809204, + "learning_rate": 6.224611352441444e-06, + "loss": 3.164084053039551, + "step": 9500 + }, + { + "epoch": 1.8261287223823248, + "grad_norm": 0.15524768829345703, + "learning_rate": 6.156867498584028e-06, + "loss": 3.164693832397461, + "step": 9505 + }, + { + "epoch": 1.8270893371757926, + "grad_norm": 0.13482025265693665, + "learning_rate": 6.089486575706431e-06, + "loss": 3.161751556396484, + "step": 9510 + }, + { + "epoch": 1.8280499519692603, + "grad_norm": 0.1538052260875702, + "learning_rate": 6.022468753817611e-06, + "loss": 3.162434196472168, + "step": 9515 + }, + { + "epoch": 1.829010566762728, + "grad_norm": 0.13271762430667877, + "learning_rate": 5.955814202010406e-06, + "loss": 3.1618423461914062, + "step": 9520 + }, + { + "epoch": 1.8299711815561959, + "grad_norm": 0.14370056986808777, + "learning_rate": 5.88952308846109e-06, + "loss": 3.1616491317749023, + "step": 9525 + }, + { + "epoch": 1.8309317963496636, + "grad_norm": 0.13553740084171295, + "learning_rate": 5.823595580428969e-06, + "loss": 3.1615522384643553, + "step": 9530 + }, + { + "epoch": 1.8318924111431316, + "grad_norm": 0.14558495581150055, + "learning_rate": 5.758031844255884e-06, + "loss": 3.163438415527344, + "step": 9535 + }, + { + "epoch": 1.8328530259365994, + "grad_norm": 0.1292106956243515, + "learning_rate": 5.692832045365858e-06, + "loss": 3.1632017135620116, + "step": 9540 + }, + { + "epoch": 1.8338136407300674, + "grad_norm": 0.13839663565158844, + "learning_rate": 5.627996348264685e-06, + "loss": 3.1595331192016602, + "step": 9545 + }, + { + "epoch": 1.8347742555235351, + "grad_norm": 0.13135622441768646, + "learning_rate": 5.563524916539508e-06, + "loss": 3.161990165710449, + "step": 9550 + }, + { + "epoch": 1.8357348703170029, + "grad_norm": 0.13140968978405, + "learning_rate": 5.499417912858289e-06, + "loss": 3.159449005126953, + "step": 9555 + }, + { + "epoch": 1.8366954851104706, + "grad_norm": 0.14151158928871155, + "learning_rate": 5.435675498969661e-06, + "loss": 3.1604162216186524, + "step": 9560 + }, + { + "epoch": 1.8376560999039384, + "grad_norm": 0.1417761594057083, + "learning_rate": 5.372297835702222e-06, + "loss": 3.1592178344726562, + "step": 9565 + }, + { + "epoch": 1.8386167146974062, + "grad_norm": 0.13075558841228485, + "learning_rate": 5.309285082964343e-06, + "loss": 3.159617233276367, + "step": 9570 + }, + { + "epoch": 1.8395773294908742, + "grad_norm": 0.13188965618610382, + "learning_rate": 5.24663739974363e-06, + "loss": 3.1601367950439454, + "step": 9575 + }, + { + "epoch": 1.8405379442843421, + "grad_norm": 0.13191638886928558, + "learning_rate": 5.184354944106661e-06, + "loss": 3.16494197845459, + "step": 9580 + }, + { + "epoch": 1.84149855907781, + "grad_norm": 0.16160184144973755, + "learning_rate": 5.122437873198415e-06, + "loss": 3.159535217285156, + "step": 9585 + }, + { + "epoch": 1.8424591738712777, + "grad_norm": 0.13549911975860596, + "learning_rate": 5.060886343242043e-06, + "loss": 3.1587257385253906, + "step": 9590 + }, + { + "epoch": 1.8434197886647454, + "grad_norm": 0.13970741629600525, + "learning_rate": 4.999700509538368e-06, + "loss": 3.1649681091308595, + "step": 9595 + }, + { + "epoch": 1.8443804034582132, + "grad_norm": 0.134404718875885, + "learning_rate": 4.938880526465516e-06, + "loss": 3.1567209243774412, + "step": 9600 + }, + { + "epoch": 1.845341018251681, + "grad_norm": 0.14060495793819427, + "learning_rate": 4.878426547478537e-06, + "loss": 3.161220741271973, + "step": 9605 + }, + { + "epoch": 1.846301633045149, + "grad_norm": 0.14048533141613007, + "learning_rate": 4.818338725109033e-06, + "loss": 3.1618812561035154, + "step": 9610 + }, + { + "epoch": 1.8472622478386167, + "grad_norm": 0.13674895465373993, + "learning_rate": 4.758617210964749e-06, + "loss": 3.1565229415893556, + "step": 9615 + }, + { + "epoch": 1.8482228626320847, + "grad_norm": 0.13577382266521454, + "learning_rate": 4.6992621557292e-06, + "loss": 3.158913230895996, + "step": 9620 + }, + { + "epoch": 1.8491834774255524, + "grad_norm": 0.13337448239326477, + "learning_rate": 4.640273709161224e-06, + "loss": 3.161669158935547, + "step": 9625 + }, + { + "epoch": 1.8501440922190202, + "grad_norm": 0.134219229221344, + "learning_rate": 4.5816520200948005e-06, + "loss": 3.160819435119629, + "step": 9630 + }, + { + "epoch": 1.851104707012488, + "grad_norm": 0.13151714205741882, + "learning_rate": 4.5233972364383975e-06, + "loss": 3.1615375518798827, + "step": 9635 + }, + { + "epoch": 1.8520653218059557, + "grad_norm": 0.11864073574542999, + "learning_rate": 4.465509505174858e-06, + "loss": 3.1609018325805662, + "step": 9640 + }, + { + "epoch": 1.8530259365994235, + "grad_norm": 0.14346528053283691, + "learning_rate": 4.407988972360849e-06, + "loss": 3.161561393737793, + "step": 9645 + }, + { + "epoch": 1.8539865513928915, + "grad_norm": 0.13390202820301056, + "learning_rate": 4.350835783126594e-06, + "loss": 3.1598655700683596, + "step": 9650 + }, + { + "epoch": 1.8549471661863592, + "grad_norm": 0.12858052551746368, + "learning_rate": 4.2940500816754094e-06, + "loss": 3.160798263549805, + "step": 9655 + }, + { + "epoch": 1.8559077809798272, + "grad_norm": 0.13176590204238892, + "learning_rate": 4.237632011283471e-06, + "loss": 3.158690643310547, + "step": 9660 + }, + { + "epoch": 1.856868395773295, + "grad_norm": 0.12532593309879303, + "learning_rate": 4.181581714299359e-06, + "loss": 3.159891128540039, + "step": 9665 + }, + { + "epoch": 1.8578290105667628, + "grad_norm": 0.11781672388315201, + "learning_rate": 4.12589933214375e-06, + "loss": 3.161539077758789, + "step": 9670 + }, + { + "epoch": 1.8587896253602305, + "grad_norm": 0.1212988868355751, + "learning_rate": 4.070585005308946e-06, + "loss": 3.16253662109375, + "step": 9675 + }, + { + "epoch": 1.8597502401536983, + "grad_norm": 0.11885207146406174, + "learning_rate": 4.015638873358707e-06, + "loss": 3.162197303771973, + "step": 9680 + }, + { + "epoch": 1.860710854947166, + "grad_norm": 0.12261178344488144, + "learning_rate": 3.961061074927752e-06, + "loss": 3.1611923217773437, + "step": 9685 + }, + { + "epoch": 1.861671469740634, + "grad_norm": 0.13136336207389832, + "learning_rate": 3.906851747721495e-06, + "loss": 3.160177993774414, + "step": 9690 + }, + { + "epoch": 1.8626320845341018, + "grad_norm": 0.13114266097545624, + "learning_rate": 3.853011028515563e-06, + "loss": 3.15836124420166, + "step": 9695 + }, + { + "epoch": 1.8635926993275698, + "grad_norm": 0.13357418775558472, + "learning_rate": 3.79953905315567e-06, + "loss": 3.161663627624512, + "step": 9700 + }, + { + "epoch": 1.8645533141210375, + "grad_norm": 0.11809264123439789, + "learning_rate": 3.746435956557081e-06, + "loss": 3.1624309539794924, + "step": 9705 + }, + { + "epoch": 1.8655139289145053, + "grad_norm": 0.1218031793832779, + "learning_rate": 3.6937018727043523e-06, + "loss": 3.1634284973144533, + "step": 9710 + }, + { + "epoch": 1.866474543707973, + "grad_norm": 0.12914207577705383, + "learning_rate": 3.6413369346509848e-06, + "loss": 3.1623727798461916, + "step": 9715 + }, + { + "epoch": 1.8674351585014408, + "grad_norm": 0.12263166159391403, + "learning_rate": 3.5893412745191085e-06, + "loss": 3.1578353881835937, + "step": 9720 + }, + { + "epoch": 1.8683957732949086, + "grad_norm": 0.11959100514650345, + "learning_rate": 3.5377150234990824e-06, + "loss": 3.1602123260498045, + "step": 9725 + }, + { + "epoch": 1.8693563880883766, + "grad_norm": 0.132080078125, + "learning_rate": 3.4864583118492438e-06, + "loss": 3.1630222320556642, + "step": 9730 + }, + { + "epoch": 1.8703170028818443, + "grad_norm": 0.13228577375411987, + "learning_rate": 3.435571268895526e-06, + "loss": 3.155730438232422, + "step": 9735 + }, + { + "epoch": 1.8712776176753123, + "grad_norm": 0.12898947298526764, + "learning_rate": 3.3850540230311918e-06, + "loss": 3.1614667892456056, + "step": 9740 + }, + { + "epoch": 1.87223823246878, + "grad_norm": 0.1316165179014206, + "learning_rate": 3.3349067017163833e-06, + "loss": 3.162592315673828, + "step": 9745 + }, + { + "epoch": 1.8731988472622478, + "grad_norm": 0.12914599478244781, + "learning_rate": 3.28512943147799e-06, + "loss": 3.160664749145508, + "step": 9750 + }, + { + "epoch": 1.8741594620557156, + "grad_norm": 0.11813253164291382, + "learning_rate": 3.23572233790913e-06, + "loss": 3.1575859069824217, + "step": 9755 + }, + { + "epoch": 1.8751200768491834, + "grad_norm": 0.1261204481124878, + "learning_rate": 3.1866855456689878e-06, + "loss": 3.1581016540527345, + "step": 9760 + }, + { + "epoch": 1.8760806916426513, + "grad_norm": 0.120733842253685, + "learning_rate": 3.138019178482409e-06, + "loss": 3.1588390350341795, + "step": 9765 + }, + { + "epoch": 1.877041306436119, + "grad_norm": 0.13560213148593903, + "learning_rate": 3.0897233591396562e-06, + "loss": 3.1578319549560545, + "step": 9770 + }, + { + "epoch": 1.878001921229587, + "grad_norm": 0.11367765814065933, + "learning_rate": 3.0417982094960212e-06, + "loss": 3.1566787719726563, + "step": 9775 + }, + { + "epoch": 1.8789625360230549, + "grad_norm": 0.11928807944059372, + "learning_rate": 2.994243850471545e-06, + "loss": 3.162850570678711, + "step": 9780 + }, + { + "epoch": 1.8799231508165226, + "grad_norm": 0.11847756057977676, + "learning_rate": 2.9470604020507994e-06, + "loss": 3.1612308502197264, + "step": 9785 + }, + { + "epoch": 1.8808837656099904, + "grad_norm": 0.1250849962234497, + "learning_rate": 2.9002479832824553e-06, + "loss": 3.1637001037597656, + "step": 9790 + }, + { + "epoch": 1.8818443804034581, + "grad_norm": 0.12311087548732758, + "learning_rate": 2.8538067122790164e-06, + "loss": 3.1589778900146483, + "step": 9795 + }, + { + "epoch": 1.882804995196926, + "grad_norm": 0.11985550820827484, + "learning_rate": 2.8077367062166013e-06, + "loss": 3.1609506607055664, + "step": 9800 + }, + { + "epoch": 1.8837656099903939, + "grad_norm": 0.12335570156574249, + "learning_rate": 2.762038081334561e-06, + "loss": 3.1584909439086912, + "step": 9805 + }, + { + "epoch": 1.8847262247838616, + "grad_norm": 0.12690205872058868, + "learning_rate": 2.716710952935214e-06, + "loss": 3.160411071777344, + "step": 9810 + }, + { + "epoch": 1.8856868395773296, + "grad_norm": 0.12958942353725433, + "learning_rate": 2.6717554353835435e-06, + "loss": 3.16156005859375, + "step": 9815 + }, + { + "epoch": 1.8866474543707974, + "grad_norm": 0.1530323475599289, + "learning_rate": 2.6271716421069344e-06, + "loss": 3.1630611419677734, + "step": 9820 + }, + { + "epoch": 1.8876080691642652, + "grad_norm": 0.11445850133895874, + "learning_rate": 2.5829596855948876e-06, + "loss": 3.160757064819336, + "step": 9825 + }, + { + "epoch": 1.888568683957733, + "grad_norm": 0.12335135042667389, + "learning_rate": 2.539119677398671e-06, + "loss": 3.159043884277344, + "step": 9830 + }, + { + "epoch": 1.8895292987512007, + "grad_norm": 0.12193870544433594, + "learning_rate": 2.495651728131154e-06, + "loss": 3.1591602325439454, + "step": 9835 + }, + { + "epoch": 1.8904899135446684, + "grad_norm": 0.11826537549495697, + "learning_rate": 2.452555947466439e-06, + "loss": 3.161988067626953, + "step": 9840 + }, + { + "epoch": 1.8914505283381364, + "grad_norm": 0.12225425243377686, + "learning_rate": 2.4098324441395645e-06, + "loss": 3.164390754699707, + "step": 9845 + }, + { + "epoch": 1.8924111431316042, + "grad_norm": 0.11919926851987839, + "learning_rate": 2.367481325946352e-06, + "loss": 3.160402297973633, + "step": 9850 + }, + { + "epoch": 1.8933717579250722, + "grad_norm": 0.11494944989681244, + "learning_rate": 2.325502699742976e-06, + "loss": 3.159711456298828, + "step": 9855 + }, + { + "epoch": 1.89433237271854, + "grad_norm": 0.11843105405569077, + "learning_rate": 2.283896671445862e-06, + "loss": 3.159711456298828, + "step": 9860 + }, + { + "epoch": 1.8952929875120077, + "grad_norm": 0.12161135673522949, + "learning_rate": 2.2426633460312373e-06, + "loss": 3.1607955932617187, + "step": 9865 + }, + { + "epoch": 1.8962536023054755, + "grad_norm": 0.114869125187397, + "learning_rate": 2.2018028275350652e-06, + "loss": 3.161262321472168, + "step": 9870 + }, + { + "epoch": 1.8972142170989432, + "grad_norm": 0.129729226231575, + "learning_rate": 2.1613152190525785e-06, + "loss": 3.161954879760742, + "step": 9875 + }, + { + "epoch": 1.898174831892411, + "grad_norm": 0.12527115643024445, + "learning_rate": 2.1212006227382117e-06, + "loss": 3.159313201904297, + "step": 9880 + }, + { + "epoch": 1.899135446685879, + "grad_norm": 0.1106216311454773, + "learning_rate": 2.0814591398051527e-06, + "loss": 3.160965347290039, + "step": 9885 + }, + { + "epoch": 1.9000960614793467, + "grad_norm": 0.11833595484495163, + "learning_rate": 2.0420908705252926e-06, + "loss": 3.160472106933594, + "step": 9890 + }, + { + "epoch": 1.9010566762728147, + "grad_norm": 0.10949314385652542, + "learning_rate": 2.003095914228775e-06, + "loss": 3.163528060913086, + "step": 9895 + }, + { + "epoch": 1.9020172910662825, + "grad_norm": 0.1126236766576767, + "learning_rate": 1.9644743693038977e-06, + "loss": 3.1650867462158203, + "step": 9900 + }, + { + "epoch": 1.9029779058597502, + "grad_norm": 0.1123654693365097, + "learning_rate": 1.9262263331967785e-06, + "loss": 3.1654218673706054, + "step": 9905 + }, + { + "epoch": 1.903938520653218, + "grad_norm": 0.12411278486251831, + "learning_rate": 1.8883519024111227e-06, + "loss": 3.156511688232422, + "step": 9910 + }, + { + "epoch": 1.9048991354466858, + "grad_norm": 0.1198534443974495, + "learning_rate": 1.8508511725080388e-06, + "loss": 3.160737991333008, + "step": 9915 + }, + { + "epoch": 1.9058597502401537, + "grad_norm": 0.11347641795873642, + "learning_rate": 1.8137242381056571e-06, + "loss": 3.1631191253662108, + "step": 9920 + }, + { + "epoch": 1.9068203650336215, + "grad_norm": 0.11204273998737335, + "learning_rate": 1.7769711928790953e-06, + "loss": 3.158831787109375, + "step": 9925 + }, + { + "epoch": 1.9077809798270895, + "grad_norm": 0.12282492965459824, + "learning_rate": 1.7405921295600421e-06, + "loss": 3.1593994140625, + "step": 9930 + }, + { + "epoch": 1.9087415946205573, + "grad_norm": 0.11055224388837814, + "learning_rate": 1.7045871399365918e-06, + "loss": 3.159510040283203, + "step": 9935 + }, + { + "epoch": 1.909702209414025, + "grad_norm": 0.11742359399795532, + "learning_rate": 1.668956314853026e-06, + "loss": 3.161764144897461, + "step": 9940 + }, + { + "epoch": 1.9106628242074928, + "grad_norm": 0.10971947759389877, + "learning_rate": 1.6336997442095823e-06, + "loss": 3.1589191436767576, + "step": 9945 + }, + { + "epoch": 1.9116234390009605, + "grad_norm": 0.11144320666790009, + "learning_rate": 1.5988175169622197e-06, + "loss": 3.1609643936157226, + "step": 9950 + }, + { + "epoch": 1.9125840537944283, + "grad_norm": 0.11729393154382706, + "learning_rate": 1.5643097211223533e-06, + "loss": 3.1646373748779295, + "step": 9955 + }, + { + "epoch": 1.9135446685878963, + "grad_norm": 0.11545123904943466, + "learning_rate": 1.5301764437567032e-06, + "loss": 3.1612579345703127, + "step": 9960 + }, + { + "epoch": 1.914505283381364, + "grad_norm": 0.1158137395977974, + "learning_rate": 1.4964177709870128e-06, + "loss": 3.1609870910644533, + "step": 9965 + }, + { + "epoch": 1.915465898174832, + "grad_norm": 0.1098775789141655, + "learning_rate": 1.4630337879899146e-06, + "loss": 3.1606449127197265, + "step": 9970 + }, + { + "epoch": 1.9164265129682998, + "grad_norm": 0.11874253302812576, + "learning_rate": 1.4300245789965803e-06, + "loss": 3.1594825744628907, + "step": 9975 + }, + { + "epoch": 1.9173871277617676, + "grad_norm": 0.11549612134695053, + "learning_rate": 1.3973902272926384e-06, + "loss": 3.160371780395508, + "step": 9980 + }, + { + "epoch": 1.9183477425552353, + "grad_norm": 0.10717111080884933, + "learning_rate": 1.3651308152179407e-06, + "loss": 3.1619895935058593, + "step": 9985 + }, + { + "epoch": 1.919308357348703, + "grad_norm": 0.11006828397512436, + "learning_rate": 1.3332464241662622e-06, + "loss": 3.1590812683105467, + "step": 9990 + }, + { + "epoch": 1.9202689721421708, + "grad_norm": 0.11138733476400375, + "learning_rate": 1.3017371345851846e-06, + "loss": 3.158272933959961, + "step": 9995 + }, + { + "epoch": 1.9212295869356388, + "grad_norm": 0.1066170483827591, + "learning_rate": 1.2706030259759137e-06, + "loss": 3.158770942687988, + "step": 10000 + }, + { + "epoch": 1.9221902017291066, + "grad_norm": 0.11191676557064056, + "learning_rate": 1.2398441768929624e-06, + "loss": 3.1578929901123045, + "step": 10005 + }, + { + "epoch": 1.9231508165225746, + "grad_norm": 0.11436916887760162, + "learning_rate": 1.2094606649440841e-06, + "loss": 3.1585205078125, + "step": 10010 + }, + { + "epoch": 1.9241114313160423, + "grad_norm": 0.11085876077413559, + "learning_rate": 1.1794525667899901e-06, + "loss": 3.1594661712646483, + "step": 10015 + }, + { + "epoch": 1.92507204610951, + "grad_norm": 0.11293391138315201, + "learning_rate": 1.149819958144199e-06, + "loss": 3.1591583251953126, + "step": 10020 + }, + { + "epoch": 1.9260326609029779, + "grad_norm": 0.10681528598070145, + "learning_rate": 1.1205629137728044e-06, + "loss": 3.160277557373047, + "step": 10025 + }, + { + "epoch": 1.9269932756964456, + "grad_norm": 0.1138753667473793, + "learning_rate": 1.0916815074943242e-06, + "loss": 3.15789909362793, + "step": 10030 + }, + { + "epoch": 1.9279538904899134, + "grad_norm": 0.10544762760400772, + "learning_rate": 1.0631758121795508e-06, + "loss": 3.1582752227783204, + "step": 10035 + }, + { + "epoch": 1.9289145052833814, + "grad_norm": 0.11651872098445892, + "learning_rate": 1.0350458997512356e-06, + "loss": 3.1609378814697267, + "step": 10040 + }, + { + "epoch": 1.9298751200768491, + "grad_norm": 0.11160679161548615, + "learning_rate": 1.0072918411840548e-06, + "loss": 3.161101722717285, + "step": 10045 + }, + { + "epoch": 1.9308357348703171, + "grad_norm": 0.11564800143241882, + "learning_rate": 9.799137065043428e-07, + "loss": 3.1580835342407227, + "step": 10050 + }, + { + "epoch": 1.9317963496637849, + "grad_norm": 0.12115105986595154, + "learning_rate": 9.529115647899599e-07, + "loss": 3.160675811767578, + "step": 10055 + }, + { + "epoch": 1.9327569644572526, + "grad_norm": 0.11305255442857742, + "learning_rate": 9.262854841700418e-07, + "loss": 3.1640155792236326, + "step": 10060 + }, + { + "epoch": 1.9337175792507204, + "grad_norm": 0.11431337147951126, + "learning_rate": 9.00035531824933e-07, + "loss": 3.162336730957031, + "step": 10065 + }, + { + "epoch": 1.9346781940441882, + "grad_norm": 0.10593244433403015, + "learning_rate": 8.741617739859708e-07, + "loss": 3.1616371154785154, + "step": 10070 + }, + { + "epoch": 1.9356388088376562, + "grad_norm": 0.10147248208522797, + "learning_rate": 8.486642759353013e-07, + "loss": 3.1602996826171874, + "step": 10075 + }, + { + "epoch": 1.936599423631124, + "grad_norm": 0.11662466824054718, + "learning_rate": 8.23543102005697e-07, + "loss": 3.1570274353027346, + "step": 10080 + }, + { + "epoch": 1.937560038424592, + "grad_norm": 0.10574644058942795, + "learning_rate": 7.9879831558049e-07, + "loss": 3.156863975524902, + "step": 10085 + }, + { + "epoch": 1.9385206532180597, + "grad_norm": 0.10977693647146225, + "learning_rate": 7.744299790933217e-07, + "loss": 3.1590835571289064, + "step": 10090 + }, + { + "epoch": 1.9394812680115274, + "grad_norm": 0.10733279585838318, + "learning_rate": 7.504381540279603e-07, + "loss": 3.1599714279174806, + "step": 10095 + }, + { + "epoch": 1.9404418828049952, + "grad_norm": 0.1001046821475029, + "learning_rate": 7.268229009182836e-07, + "loss": 3.158583068847656, + "step": 10100 + }, + { + "epoch": 1.941402497598463, + "grad_norm": 0.10647624731063843, + "learning_rate": 7.035842793479795e-07, + "loss": 3.158983612060547, + "step": 10105 + }, + { + "epoch": 1.9423631123919307, + "grad_norm": 0.10598830133676529, + "learning_rate": 6.807223479504631e-07, + "loss": 3.1631649017333983, + "step": 10110 + }, + { + "epoch": 1.9433237271853987, + "grad_norm": 0.10469937324523926, + "learning_rate": 6.582371644087592e-07, + "loss": 3.1588485717773436, + "step": 10115 + }, + { + "epoch": 1.9442843419788665, + "grad_norm": 0.10287413746118546, + "learning_rate": 6.361287854552865e-07, + "loss": 3.161878967285156, + "step": 10120 + }, + { + "epoch": 1.9452449567723344, + "grad_norm": 0.10505318641662598, + "learning_rate": 6.143972668717411e-07, + "loss": 3.1627128601074217, + "step": 10125 + }, + { + "epoch": 1.9462055715658022, + "grad_norm": 0.1040089949965477, + "learning_rate": 5.930426634889796e-07, + "loss": 3.1591156005859373, + "step": 10130 + }, + { + "epoch": 1.94716618635927, + "grad_norm": 0.10172892361879349, + "learning_rate": 5.720650291868523e-07, + "loss": 3.1622541427612303, + "step": 10135 + }, + { + "epoch": 1.9481268011527377, + "grad_norm": 0.10863189399242401, + "learning_rate": 5.514644168941041e-07, + "loss": 3.161451721191406, + "step": 10140 + }, + { + "epoch": 1.9490874159462055, + "grad_norm": 0.10447549819946289, + "learning_rate": 5.312408785881573e-07, + "loss": 3.1583093643188476, + "step": 10145 + }, + { + "epoch": 1.9500480307396733, + "grad_norm": 0.10685736685991287, + "learning_rate": 5.113944652951118e-07, + "loss": 3.1573591232299805, + "step": 10150 + }, + { + "epoch": 1.9510086455331412, + "grad_norm": 0.10184746235609055, + "learning_rate": 4.919252270894625e-07, + "loss": 3.155690383911133, + "step": 10155 + }, + { + "epoch": 1.951969260326609, + "grad_norm": 0.10571934282779694, + "learning_rate": 4.7283321309413145e-07, + "loss": 3.159967613220215, + "step": 10160 + }, + { + "epoch": 1.952929875120077, + "grad_norm": 0.1015496551990509, + "learning_rate": 4.541184714802193e-07, + "loss": 3.1597368240356447, + "step": 10165 + }, + { + "epoch": 1.9538904899135447, + "grad_norm": 0.1008259505033493, + "learning_rate": 4.3578104946692114e-07, + "loss": 3.156237030029297, + "step": 10170 + }, + { + "epoch": 1.9548511047070125, + "grad_norm": 0.10270283371210098, + "learning_rate": 4.1782099332146046e-07, + "loss": 3.162929916381836, + "step": 10175 + }, + { + "epoch": 1.9558117195004803, + "grad_norm": 0.10830524563789368, + "learning_rate": 4.002383483588723e-07, + "loss": 3.157078170776367, + "step": 10180 + }, + { + "epoch": 1.956772334293948, + "grad_norm": 0.09829416126012802, + "learning_rate": 3.8303315894200326e-07, + "loss": 3.1587419509887695, + "step": 10185 + }, + { + "epoch": 1.9577329490874158, + "grad_norm": 0.10378840565681458, + "learning_rate": 3.662054684812954e-07, + "loss": 3.1602014541625976, + "step": 10190 + }, + { + "epoch": 1.9586935638808838, + "grad_norm": 0.10607574880123138, + "learning_rate": 3.497553194347358e-07, + "loss": 3.1587203979492187, + "step": 10195 + }, + { + "epoch": 1.9596541786743515, + "grad_norm": 0.10446880012750626, + "learning_rate": 3.3368275330775686e-07, + "loss": 3.160129928588867, + "step": 10200 + }, + { + "epoch": 1.9606147934678195, + "grad_norm": 0.10566939413547516, + "learning_rate": 3.1798781065305314e-07, + "loss": 3.162166213989258, + "step": 10205 + }, + { + "epoch": 1.9615754082612873, + "grad_norm": 0.10455495119094849, + "learning_rate": 3.0267053107061456e-07, + "loss": 3.1622539520263673, + "step": 10210 + }, + { + "epoch": 1.962536023054755, + "grad_norm": 0.10329185426235199, + "learning_rate": 2.877309532074934e-07, + "loss": 3.155780029296875, + "step": 10215 + }, + { + "epoch": 1.9634966378482228, + "grad_norm": 0.09786950796842575, + "learning_rate": 2.7316911475777083e-07, + "loss": 3.160215377807617, + "step": 10220 + }, + { + "epoch": 1.9644572526416906, + "grad_norm": 0.10430426150560379, + "learning_rate": 2.589850524624737e-07, + "loss": 3.158669662475586, + "step": 10225 + }, + { + "epoch": 1.9654178674351583, + "grad_norm": 0.10500769317150116, + "learning_rate": 2.451788021094414e-07, + "loss": 3.1637641906738283, + "step": 10230 + }, + { + "epoch": 1.9663784822286263, + "grad_norm": 0.09861162304878235, + "learning_rate": 2.317503985332425e-07, + "loss": 3.1611236572265624, + "step": 10235 + }, + { + "epoch": 1.967339097022094, + "grad_norm": 0.09918565303087234, + "learning_rate": 2.1869987561514147e-07, + "loss": 3.15871639251709, + "step": 10240 + }, + { + "epoch": 1.968299711815562, + "grad_norm": 0.1011766865849495, + "learning_rate": 2.060272662829321e-07, + "loss": 3.16119499206543, + "step": 10245 + }, + { + "epoch": 1.9692603266090298, + "grad_norm": 0.11001937091350555, + "learning_rate": 1.9373260251092092e-07, + "loss": 3.1602630615234375, + "step": 10250 + }, + { + "epoch": 1.9702209414024976, + "grad_norm": 0.10449342429637909, + "learning_rate": 1.8181591531977736e-07, + "loss": 3.160749816894531, + "step": 10255 + }, + { + "epoch": 1.9711815561959654, + "grad_norm": 0.09981942176818848, + "learning_rate": 1.7027723477656686e-07, + "loss": 3.159575653076172, + "step": 10260 + }, + { + "epoch": 1.9721421709894331, + "grad_norm": 0.10080744326114655, + "learning_rate": 1.5911658999453458e-07, + "loss": 3.16530704498291, + "step": 10265 + }, + { + "epoch": 1.973102785782901, + "grad_norm": 0.09806544333696365, + "learning_rate": 1.4833400913313864e-07, + "loss": 3.1608566284179687, + "step": 10270 + }, + { + "epoch": 1.9740634005763689, + "grad_norm": 0.10161542147397995, + "learning_rate": 1.379295193979335e-07, + "loss": 3.1610599517822267, + "step": 10275 + }, + { + "epoch": 1.9750240153698368, + "grad_norm": 0.10026436299085617, + "learning_rate": 1.2790314704052008e-07, + "loss": 3.1599550247192383, + "step": 10280 + }, + { + "epoch": 1.9759846301633046, + "grad_norm": 0.09924504905939102, + "learning_rate": 1.1825491735846237e-07, + "loss": 3.1612056732177733, + "step": 10285 + }, + { + "epoch": 1.9769452449567724, + "grad_norm": 0.10245411098003387, + "learning_rate": 1.0898485469523766e-07, + "loss": 3.1604587554931642, + "step": 10290 + }, + { + "epoch": 1.9779058597502401, + "grad_norm": 0.09821955114603043, + "learning_rate": 1.0009298244016972e-07, + "loss": 3.1620269775390626, + "step": 10295 + }, + { + "epoch": 1.978866474543708, + "grad_norm": 0.09943889081478119, + "learning_rate": 9.15793230283457e-08, + "loss": 3.160098648071289, + "step": 10300 + }, + { + "epoch": 1.9798270893371757, + "grad_norm": 0.09800975024700165, + "learning_rate": 8.3443897940616e-08, + "loss": 3.1579994201660155, + "step": 10305 + }, + { + "epoch": 1.9807877041306436, + "grad_norm": 0.0975160002708435, + "learning_rate": 7.568672770349449e-08, + "loss": 3.160677909851074, + "step": 10310 + }, + { + "epoch": 1.9817483189241114, + "grad_norm": 0.09866613149642944, + "learning_rate": 6.830783188910838e-08, + "loss": 3.15645809173584, + "step": 10315 + }, + { + "epoch": 1.9827089337175794, + "grad_norm": 0.09947178512811661, + "learning_rate": 6.130722911516505e-08, + "loss": 3.1583110809326174, + "step": 10320 + }, + { + "epoch": 1.9836695485110472, + "grad_norm": 0.09944622218608856, + "learning_rate": 5.468493704491872e-08, + "loss": 3.1639808654785155, + "step": 10325 + }, + { + "epoch": 1.984630163304515, + "grad_norm": 0.09704981744289398, + "learning_rate": 4.844097238708711e-08, + "loss": 3.1581130981445313, + "step": 10330 + }, + { + "epoch": 1.9855907780979827, + "grad_norm": 0.09650828689336777, + "learning_rate": 4.257535089581821e-08, + "loss": 3.156038284301758, + "step": 10335 + }, + { + "epoch": 1.9865513928914504, + "grad_norm": 0.09726572781801224, + "learning_rate": 3.70880873707069e-08, + "loss": 3.1570310592651367, + "step": 10340 + }, + { + "epoch": 1.9875120076849182, + "grad_norm": 0.09816308319568634, + "learning_rate": 3.1979195656645084e-08, + "loss": 3.160786819458008, + "step": 10345 + }, + { + "epoch": 1.9884726224783862, + "grad_norm": 0.09846552461385727, + "learning_rate": 2.72486886439216e-08, + "loss": 3.1611724853515626, + "step": 10350 + }, + { + "epoch": 1.989433237271854, + "grad_norm": 0.09757398068904877, + "learning_rate": 2.289657826807234e-08, + "loss": 3.1581047058105467, + "step": 10355 + }, + { + "epoch": 1.990393852065322, + "grad_norm": 0.10126172751188278, + "learning_rate": 1.8922875509930212e-08, + "loss": 3.160317611694336, + "step": 10360 + }, + { + "epoch": 1.9913544668587897, + "grad_norm": 0.09841560572385788, + "learning_rate": 1.532759039554188e-08, + "loss": 3.158650207519531, + "step": 10365 + }, + { + "epoch": 1.9923150816522575, + "grad_norm": 0.09537240862846375, + "learning_rate": 1.2110731996201051e-08, + "loss": 3.1590465545654296, + "step": 10370 + }, + { + "epoch": 1.9932756964457252, + "grad_norm": 0.09531023353338242, + "learning_rate": 9.272308428348586e-09, + "loss": 3.159567642211914, + "step": 10375 + }, + { + "epoch": 1.994236311239193, + "grad_norm": 0.10250640660524368, + "learning_rate": 6.812326853639083e-09, + "loss": 3.1602352142333983, + "step": 10380 + }, + { + "epoch": 1.9951969260326607, + "grad_norm": 0.09563779830932617, + "learning_rate": 4.730793478840977e-09, + "loss": 3.157382583618164, + "step": 10385 + }, + { + "epoch": 1.9961575408261287, + "grad_norm": 0.09357914328575134, + "learning_rate": 3.0277135558864906e-09, + "loss": 3.1614322662353516, + "step": 10390 + }, + { + "epoch": 1.9971181556195965, + "grad_norm": 0.09247033298015594, + "learning_rate": 1.7030913818050217e-09, + "loss": 3.161212921142578, + "step": 10395 + }, + { + "epoch": 1.9980787704130645, + "grad_norm": 0.09770449995994568, + "learning_rate": 7.569302987897597e-10, + "loss": 3.1566755294799806, + "step": 10400 + }, + { + "epoch": 1.9990393852065322, + "grad_norm": 0.0940788984298706, + "learning_rate": 1.8923269406445618e-10, + "loss": 3.157200050354004, + "step": 10405 + }, + { + "epoch": 2.0, + "grad_norm": 0.10163447260856628, + "learning_rate": 0.0, + "loss": 3.162412643432617, + "step": 10410 + } + ], + "logging_steps": 5, + "max_steps": 10410, + "num_input_tokens_seen": 0, + "num_train_epochs": 2, + "save_steps": 1000, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 1.9541356371946714e+20, + "train_batch_size": 64, + "trial_name": null, + "trial_params": null +}