diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,7474 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.22003725936659077, + "eval_steps": 500, + "global_step": 1063, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00020699648105982198, + "grad_norm": 0.3105248212814331, + "learning_rate": 1e-05, + "loss": 0.3097, + "step": 1 + }, + { + "epoch": 0.00041399296211964395, + "grad_norm": 0.35843613743782043, + "learning_rate": 2e-05, + "loss": 0.3333, + "step": 2 + }, + { + "epoch": 0.000620989443179466, + "grad_norm": 0.31169694662094116, + "learning_rate": 3e-05, + "loss": 0.3367, + "step": 3 + }, + { + "epoch": 0.0008279859242392879, + "grad_norm": 0.4025513529777527, + "learning_rate": 4e-05, + "loss": 0.3568, + "step": 4 + }, + { + "epoch": 0.0010349824052991099, + "grad_norm": 0.38164222240448, + "learning_rate": 5e-05, + "loss": 0.375, + "step": 5 + }, + { + "epoch": 0.001241978886358932, + "grad_norm": 0.41811275482177734, + "learning_rate": 6e-05, + "loss": 0.3421, + "step": 6 + }, + { + "epoch": 0.001448975367418754, + "grad_norm": 0.40563058853149414, + "learning_rate": 7e-05, + "loss": 0.3046, + "step": 7 + }, + { + "epoch": 0.0016559718484785758, + "grad_norm": 0.32561907172203064, + "learning_rate": 8e-05, + "loss": 0.3298, + "step": 8 + }, + { + "epoch": 0.0018629683295383979, + "grad_norm": 0.2336910218000412, + "learning_rate": 9e-05, + "loss": 0.2331, + "step": 9 + }, + { + "epoch": 0.0020699648105982197, + "grad_norm": 0.2214404195547104, + "learning_rate": 0.0001, + "loss": 0.2747, + "step": 10 + }, + { + "epoch": 0.002276961291658042, + "grad_norm": 0.1755189299583435, + "learning_rate": 0.00011000000000000002, + "loss": 0.215, + "step": 11 + }, + { + "epoch": 0.002483957772717864, + "grad_norm": 0.13464422523975372, + "learning_rate": 0.00012, + "loss": 0.1854, + "step": 12 + }, + { + "epoch": 0.0026909542537776857, + "grad_norm": 0.1568724364042282, + "learning_rate": 0.00013000000000000002, + "loss": 0.1954, + "step": 13 + }, + { + "epoch": 0.002897950734837508, + "grad_norm": 0.1855469048023224, + "learning_rate": 0.00014, + "loss": 0.1788, + "step": 14 + }, + { + "epoch": 0.00310494721589733, + "grad_norm": 0.16363286972045898, + "learning_rate": 0.00015000000000000001, + "loss": 0.1816, + "step": 15 + }, + { + "epoch": 0.0033119436969571516, + "grad_norm": 0.14675703644752502, + "learning_rate": 0.00016, + "loss": 0.144, + "step": 16 + }, + { + "epoch": 0.003518940178016974, + "grad_norm": 0.13585928082466125, + "learning_rate": 0.00017, + "loss": 0.1265, + "step": 17 + }, + { + "epoch": 0.0037259366590767957, + "grad_norm": 0.13742923736572266, + "learning_rate": 0.00018, + "loss": 0.1373, + "step": 18 + }, + { + "epoch": 0.003932933140136618, + "grad_norm": 0.09649409353733063, + "learning_rate": 0.00019, + "loss": 0.1161, + "step": 19 + }, + { + "epoch": 0.004139929621196439, + "grad_norm": 0.08537352085113525, + "learning_rate": 0.0002, + "loss": 0.1086, + "step": 20 + }, + { + "epoch": 0.004346926102256262, + "grad_norm": 0.07582477480173111, + "learning_rate": 0.00019995842860112245, + "loss": 0.1047, + "step": 21 + }, + { + "epoch": 0.004553922583316084, + "grad_norm": 0.07321921736001968, + "learning_rate": 0.00019991685720224486, + "loss": 0.1095, + "step": 22 + }, + { + "epoch": 0.004760919064375905, + "grad_norm": 0.07746334373950958, + "learning_rate": 0.0001998752858033673, + "loss": 0.08, + "step": 23 + }, + { + "epoch": 0.004967915545435728, + "grad_norm": 0.06693358719348907, + "learning_rate": 0.00019983371440448973, + "loss": 0.0932, + "step": 24 + }, + { + "epoch": 0.00517491202649555, + "grad_norm": 0.09097249805927277, + "learning_rate": 0.00019979214300561214, + "loss": 0.095, + "step": 25 + }, + { + "epoch": 0.005381908507555371, + "grad_norm": 0.06965727359056473, + "learning_rate": 0.00019975057160673458, + "loss": 0.0847, + "step": 26 + }, + { + "epoch": 0.005588904988615194, + "grad_norm": 0.060345325618982315, + "learning_rate": 0.00019970900020785701, + "loss": 0.0609, + "step": 27 + }, + { + "epoch": 0.005795901469675016, + "grad_norm": 0.06169256567955017, + "learning_rate": 0.00019966742880897945, + "loss": 0.0651, + "step": 28 + }, + { + "epoch": 0.006002897950734837, + "grad_norm": 0.07536791265010834, + "learning_rate": 0.00019962585741010186, + "loss": 0.0614, + "step": 29 + }, + { + "epoch": 0.00620989443179466, + "grad_norm": 0.06803280860185623, + "learning_rate": 0.0001995842860112243, + "loss": 0.0589, + "step": 30 + }, + { + "epoch": 0.006416890912854482, + "grad_norm": 0.09668745845556259, + "learning_rate": 0.0001995427146123467, + "loss": 0.0511, + "step": 31 + }, + { + "epoch": 0.006623887393914303, + "grad_norm": 0.060853298753499985, + "learning_rate": 0.00019950114321346915, + "loss": 0.0428, + "step": 32 + }, + { + "epoch": 0.0068308838749741255, + "grad_norm": 0.04822453856468201, + "learning_rate": 0.00019945957181459156, + "loss": 0.0442, + "step": 33 + }, + { + "epoch": 0.007037880356033948, + "grad_norm": 0.07689573615789413, + "learning_rate": 0.000199418000415714, + "loss": 0.0448, + "step": 34 + }, + { + "epoch": 0.007244876837093769, + "grad_norm": 0.07483747601509094, + "learning_rate": 0.00019937642901683643, + "loss": 0.0449, + "step": 35 + }, + { + "epoch": 0.0074518733181535915, + "grad_norm": 0.06677654385566711, + "learning_rate": 0.00019933485761795884, + "loss": 0.0368, + "step": 36 + }, + { + "epoch": 0.007658869799213414, + "grad_norm": 0.06132747232913971, + "learning_rate": 0.00019929328621908128, + "loss": 0.0365, + "step": 37 + }, + { + "epoch": 0.007865866280273236, + "grad_norm": 0.04215759411454201, + "learning_rate": 0.00019925171482020371, + "loss": 0.0307, + "step": 38 + }, + { + "epoch": 0.008072862761333057, + "grad_norm": 0.05545351654291153, + "learning_rate": 0.00019921014342132612, + "loss": 0.0308, + "step": 39 + }, + { + "epoch": 0.008279859242392879, + "grad_norm": 0.06014477089047432, + "learning_rate": 0.00019916857202244856, + "loss": 0.0335, + "step": 40 + }, + { + "epoch": 0.008486855723452702, + "grad_norm": 0.06840485334396362, + "learning_rate": 0.000199127000623571, + "loss": 0.0287, + "step": 41 + }, + { + "epoch": 0.008693852204512523, + "grad_norm": 0.0705009400844574, + "learning_rate": 0.0001990854292246934, + "loss": 0.0271, + "step": 42 + }, + { + "epoch": 0.008900848685572345, + "grad_norm": 0.05818328261375427, + "learning_rate": 0.00019904385782581585, + "loss": 0.0253, + "step": 43 + }, + { + "epoch": 0.009107845166632168, + "grad_norm": 0.03936947509646416, + "learning_rate": 0.00019900228642693828, + "loss": 0.0216, + "step": 44 + }, + { + "epoch": 0.00931484164769199, + "grad_norm": 0.044559165835380554, + "learning_rate": 0.00019896071502806072, + "loss": 0.0212, + "step": 45 + }, + { + "epoch": 0.00952183812875181, + "grad_norm": 0.04012398421764374, + "learning_rate": 0.00019891914362918313, + "loss": 0.014, + "step": 46 + }, + { + "epoch": 0.009728834609811634, + "grad_norm": 0.048441193997859955, + "learning_rate": 0.00019887757223030557, + "loss": 0.0144, + "step": 47 + }, + { + "epoch": 0.009935831090871455, + "grad_norm": 0.046538762748241425, + "learning_rate": 0.000198836000831428, + "loss": 0.0111, + "step": 48 + }, + { + "epoch": 0.010142827571931277, + "grad_norm": 0.04020772501826286, + "learning_rate": 0.0001987944294325504, + "loss": 0.0103, + "step": 49 + }, + { + "epoch": 0.0103498240529911, + "grad_norm": 0.030860010534524918, + "learning_rate": 0.00019875285803367285, + "loss": 0.0071, + "step": 50 + }, + { + "epoch": 0.010556820534050921, + "grad_norm": 0.0534852109849453, + "learning_rate": 0.0001987112866347953, + "loss": 0.0082, + "step": 51 + }, + { + "epoch": 0.010763817015110743, + "grad_norm": 0.061904191970825195, + "learning_rate": 0.0001986697152359177, + "loss": 0.0092, + "step": 52 + }, + { + "epoch": 0.010970813496170566, + "grad_norm": 0.06135628744959831, + "learning_rate": 0.0001986281438370401, + "loss": 0.0091, + "step": 53 + }, + { + "epoch": 0.011177809977230387, + "grad_norm": 0.060930103063583374, + "learning_rate": 0.00019858657243816254, + "loss": 0.0103, + "step": 54 + }, + { + "epoch": 0.011384806458290209, + "grad_norm": 0.038847871124744415, + "learning_rate": 0.00019854500103928498, + "loss": 0.0077, + "step": 55 + }, + { + "epoch": 0.011591802939350032, + "grad_norm": 0.01643364317715168, + "learning_rate": 0.0001985034296404074, + "loss": 0.0065, + "step": 56 + }, + { + "epoch": 0.011798799420409853, + "grad_norm": 0.03226076811552048, + "learning_rate": 0.00019846185824152983, + "loss": 0.0059, + "step": 57 + }, + { + "epoch": 0.012005795901469675, + "grad_norm": 0.045181699097156525, + "learning_rate": 0.00019842028684265227, + "loss": 0.0075, + "step": 58 + }, + { + "epoch": 0.012212792382529498, + "grad_norm": 0.0442410409450531, + "learning_rate": 0.00019837871544377468, + "loss": 0.007, + "step": 59 + }, + { + "epoch": 0.01241978886358932, + "grad_norm": 0.04646408557891846, + "learning_rate": 0.0001983371440448971, + "loss": 0.009, + "step": 60 + }, + { + "epoch": 0.01262678534464914, + "grad_norm": 0.037814535200595856, + "learning_rate": 0.00019829557264601955, + "loss": 0.0062, + "step": 61 + }, + { + "epoch": 0.012833781825708964, + "grad_norm": 0.029840698465704918, + "learning_rate": 0.000198254001247142, + "loss": 0.0075, + "step": 62 + }, + { + "epoch": 0.013040778306768785, + "grad_norm": 0.009363808669149876, + "learning_rate": 0.0001982124298482644, + "loss": 0.0054, + "step": 63 + }, + { + "epoch": 0.013247774787828607, + "grad_norm": 0.02504296600818634, + "learning_rate": 0.00019817085844938683, + "loss": 0.0051, + "step": 64 + }, + { + "epoch": 0.01345477126888843, + "grad_norm": 0.03861517831683159, + "learning_rate": 0.00019812928705050927, + "loss": 0.0061, + "step": 65 + }, + { + "epoch": 0.013661767749948251, + "grad_norm": 0.04222877696156502, + "learning_rate": 0.00019808771565163168, + "loss": 0.0074, + "step": 66 + }, + { + "epoch": 0.013868764231008072, + "grad_norm": 0.03613612800836563, + "learning_rate": 0.00019804614425275412, + "loss": 0.0063, + "step": 67 + }, + { + "epoch": 0.014075760712067896, + "grad_norm": 0.02817763015627861, + "learning_rate": 0.00019800457285387655, + "loss": 0.0074, + "step": 68 + }, + { + "epoch": 0.014282757193127717, + "grad_norm": 0.01215298566967249, + "learning_rate": 0.00019796300145499896, + "loss": 0.0057, + "step": 69 + }, + { + "epoch": 0.014489753674187538, + "grad_norm": 0.01781376637518406, + "learning_rate": 0.0001979214300561214, + "loss": 0.0067, + "step": 70 + }, + { + "epoch": 0.014696750155247362, + "grad_norm": 0.02770097553730011, + "learning_rate": 0.00019787985865724384, + "loss": 0.0057, + "step": 71 + }, + { + "epoch": 0.014903746636307183, + "grad_norm": 0.029749557375907898, + "learning_rate": 0.00019783828725836625, + "loss": 0.0063, + "step": 72 + }, + { + "epoch": 0.015110743117367004, + "grad_norm": 0.02843872457742691, + "learning_rate": 0.00019779671585948869, + "loss": 0.0076, + "step": 73 + }, + { + "epoch": 0.015317739598426828, + "grad_norm": 0.019167358055710793, + "learning_rate": 0.00019775514446061112, + "loss": 0.0059, + "step": 74 + }, + { + "epoch": 0.015524736079486649, + "grad_norm": 0.01691405475139618, + "learning_rate": 0.00019771357306173353, + "loss": 0.0072, + "step": 75 + }, + { + "epoch": 0.015731732560546472, + "grad_norm": 0.00585334375500679, + "learning_rate": 0.00019767200166285594, + "loss": 0.0047, + "step": 76 + }, + { + "epoch": 0.015938729041606294, + "grad_norm": 0.019127612933516502, + "learning_rate": 0.00019763043026397838, + "loss": 0.0047, + "step": 77 + }, + { + "epoch": 0.016145725522666115, + "grad_norm": 0.026558954268693924, + "learning_rate": 0.00019758885886510082, + "loss": 0.0064, + "step": 78 + }, + { + "epoch": 0.016352722003725936, + "grad_norm": 0.0275382362306118, + "learning_rate": 0.00019754728746622325, + "loss": 0.006, + "step": 79 + }, + { + "epoch": 0.016559718484785758, + "grad_norm": 0.016686394810676575, + "learning_rate": 0.00019750571606734566, + "loss": 0.0069, + "step": 80 + }, + { + "epoch": 0.01676671496584558, + "grad_norm": 0.011664893478155136, + "learning_rate": 0.0001974641446684681, + "loss": 0.0055, + "step": 81 + }, + { + "epoch": 0.016973711446905404, + "grad_norm": 0.010350242257118225, + "learning_rate": 0.00019742257326959054, + "loss": 0.0045, + "step": 82 + }, + { + "epoch": 0.017180707927965225, + "grad_norm": 0.018541481345891953, + "learning_rate": 0.00019738100187071295, + "loss": 0.0056, + "step": 83 + }, + { + "epoch": 0.017387704409025047, + "grad_norm": 0.015899088233709335, + "learning_rate": 0.00019733943047183539, + "loss": 0.0047, + "step": 84 + }, + { + "epoch": 0.017594700890084868, + "grad_norm": 0.01706838794052601, + "learning_rate": 0.00019729785907295782, + "loss": 0.0059, + "step": 85 + }, + { + "epoch": 0.01780169737114469, + "grad_norm": 0.01610150933265686, + "learning_rate": 0.00019725628767408023, + "loss": 0.0065, + "step": 86 + }, + { + "epoch": 0.01800869385220451, + "grad_norm": 0.011388594284653664, + "learning_rate": 0.00019721471627520267, + "loss": 0.0054, + "step": 87 + }, + { + "epoch": 0.018215690333264336, + "grad_norm": 0.010438695549964905, + "learning_rate": 0.0001971731448763251, + "loss": 0.005, + "step": 88 + }, + { + "epoch": 0.018422686814324157, + "grad_norm": 0.01277916319668293, + "learning_rate": 0.00019713157347744752, + "loss": 0.0045, + "step": 89 + }, + { + "epoch": 0.01862968329538398, + "grad_norm": 0.016964582726359367, + "learning_rate": 0.00019709000207856995, + "loss": 0.0064, + "step": 90 + }, + { + "epoch": 0.0188366797764438, + "grad_norm": 0.015511998906731606, + "learning_rate": 0.0001970484306796924, + "loss": 0.0046, + "step": 91 + }, + { + "epoch": 0.01904367625750362, + "grad_norm": 0.014610686339437962, + "learning_rate": 0.00019700685928081483, + "loss": 0.0068, + "step": 92 + }, + { + "epoch": 0.019250672738563443, + "grad_norm": 0.011182552203536034, + "learning_rate": 0.00019696528788193724, + "loss": 0.005, + "step": 93 + }, + { + "epoch": 0.019457669219623268, + "grad_norm": 0.00904427282512188, + "learning_rate": 0.00019692371648305967, + "loss": 0.0052, + "step": 94 + }, + { + "epoch": 0.01966466570068309, + "grad_norm": 0.013450189493596554, + "learning_rate": 0.0001968821450841821, + "loss": 0.0053, + "step": 95 + }, + { + "epoch": 0.01987166218174291, + "grad_norm": 0.011684760451316833, + "learning_rate": 0.00019684057368530452, + "loss": 0.0051, + "step": 96 + }, + { + "epoch": 0.020078658662802732, + "grad_norm": 0.009463542141020298, + "learning_rate": 0.00019679900228642693, + "loss": 0.0055, + "step": 97 + }, + { + "epoch": 0.020285655143862553, + "grad_norm": 0.007411513477563858, + "learning_rate": 0.00019675743088754937, + "loss": 0.005, + "step": 98 + }, + { + "epoch": 0.020492651624922375, + "grad_norm": 0.013031812384724617, + "learning_rate": 0.0001967158594886718, + "loss": 0.0063, + "step": 99 + }, + { + "epoch": 0.0206996481059822, + "grad_norm": 0.006821679417043924, + "learning_rate": 0.00019667428808979422, + "loss": 0.0058, + "step": 100 + }, + { + "epoch": 0.02090664458704202, + "grad_norm": 0.009770995937287807, + "learning_rate": 0.00019663271669091665, + "loss": 0.0045, + "step": 101 + }, + { + "epoch": 0.021113641068101843, + "grad_norm": 0.007675915956497192, + "learning_rate": 0.0001965911452920391, + "loss": 0.0049, + "step": 102 + }, + { + "epoch": 0.021320637549161664, + "grad_norm": 0.009076464921236038, + "learning_rate": 0.0001965495738931615, + "loss": 0.0048, + "step": 103 + }, + { + "epoch": 0.021527634030221485, + "grad_norm": 0.010678350925445557, + "learning_rate": 0.00019650800249428394, + "loss": 0.0056, + "step": 104 + }, + { + "epoch": 0.021734630511281307, + "grad_norm": 0.029563505202531815, + "learning_rate": 0.00019646643109540637, + "loss": 0.0049, + "step": 105 + }, + { + "epoch": 0.02194162699234113, + "grad_norm": 0.0072585404850542545, + "learning_rate": 0.00019642485969652878, + "loss": 0.005, + "step": 106 + }, + { + "epoch": 0.022148623473400953, + "grad_norm": 0.009124008938670158, + "learning_rate": 0.00019638328829765122, + "loss": 0.0045, + "step": 107 + }, + { + "epoch": 0.022355619954460774, + "grad_norm": 0.005743277724832296, + "learning_rate": 0.00019634171689877366, + "loss": 0.0051, + "step": 108 + }, + { + "epoch": 0.022562616435520596, + "grad_norm": 0.01300257071852684, + "learning_rate": 0.0001963001454998961, + "loss": 0.0049, + "step": 109 + }, + { + "epoch": 0.022769612916580417, + "grad_norm": 0.02877631224691868, + "learning_rate": 0.0001962585741010185, + "loss": 0.0053, + "step": 110 + }, + { + "epoch": 0.02297660939764024, + "grad_norm": 0.010237788781523705, + "learning_rate": 0.00019621700270214094, + "loss": 0.0046, + "step": 111 + }, + { + "epoch": 0.023183605878700064, + "grad_norm": 0.010189997963607311, + "learning_rate": 0.00019617543130326338, + "loss": 0.0056, + "step": 112 + }, + { + "epoch": 0.023390602359759885, + "grad_norm": 0.010226712562143803, + "learning_rate": 0.0001961338599043858, + "loss": 0.005, + "step": 113 + }, + { + "epoch": 0.023597598840819706, + "grad_norm": 0.011194237507879734, + "learning_rate": 0.00019609228850550823, + "loss": 0.0054, + "step": 114 + }, + { + "epoch": 0.023804595321879528, + "grad_norm": 0.0065891253761947155, + "learning_rate": 0.00019605071710663066, + "loss": 0.0057, + "step": 115 + }, + { + "epoch": 0.02401159180293935, + "grad_norm": 0.008131214417517185, + "learning_rate": 0.00019600914570775307, + "loss": 0.0056, + "step": 116 + }, + { + "epoch": 0.02421858828399917, + "grad_norm": 0.013916265219449997, + "learning_rate": 0.0001959675743088755, + "loss": 0.0051, + "step": 117 + }, + { + "epoch": 0.024425584765058996, + "grad_norm": 0.00658258656039834, + "learning_rate": 0.00019592600290999795, + "loss": 0.0048, + "step": 118 + }, + { + "epoch": 0.024632581246118817, + "grad_norm": 0.005407229065895081, + "learning_rate": 0.00019588443151112036, + "loss": 0.0053, + "step": 119 + }, + { + "epoch": 0.02483957772717864, + "grad_norm": 0.009581067599356174, + "learning_rate": 0.00019584286011224277, + "loss": 0.0048, + "step": 120 + }, + { + "epoch": 0.02504657420823846, + "grad_norm": 0.011583163402974606, + "learning_rate": 0.0001958012887133652, + "loss": 0.0049, + "step": 121 + }, + { + "epoch": 0.02525357068929828, + "grad_norm": 0.017572317272424698, + "learning_rate": 0.00019575971731448764, + "loss": 0.0049, + "step": 122 + }, + { + "epoch": 0.025460567170358103, + "grad_norm": 0.018644949421286583, + "learning_rate": 0.00019571814591561005, + "loss": 0.0068, + "step": 123 + }, + { + "epoch": 0.025667563651417927, + "grad_norm": 0.014408939518034458, + "learning_rate": 0.0001956765745167325, + "loss": 0.0056, + "step": 124 + }, + { + "epoch": 0.02587456013247775, + "grad_norm": 0.0033774918410927057, + "learning_rate": 0.00019563500311785493, + "loss": 0.0053, + "step": 125 + }, + { + "epoch": 0.02608155661353757, + "grad_norm": 0.005626993719488382, + "learning_rate": 0.00019559343171897736, + "loss": 0.0046, + "step": 126 + }, + { + "epoch": 0.02628855309459739, + "grad_norm": 0.018701711669564247, + "learning_rate": 0.00019555186032009977, + "loss": 0.0077, + "step": 127 + }, + { + "epoch": 0.026495549575657213, + "grad_norm": 0.016636714339256287, + "learning_rate": 0.0001955102889212222, + "loss": 0.005, + "step": 128 + }, + { + "epoch": 0.026702546056717034, + "grad_norm": 0.013526069931685925, + "learning_rate": 0.00019546871752234465, + "loss": 0.0047, + "step": 129 + }, + { + "epoch": 0.02690954253777686, + "grad_norm": 0.024728018790483475, + "learning_rate": 0.00019542714612346706, + "loss": 0.0045, + "step": 130 + }, + { + "epoch": 0.02711653901883668, + "grad_norm": 0.02217916212975979, + "learning_rate": 0.0001953855747245895, + "loss": 0.0045, + "step": 131 + }, + { + "epoch": 0.027323535499896502, + "grad_norm": 0.010518092662096024, + "learning_rate": 0.00019534400332571193, + "loss": 0.0065, + "step": 132 + }, + { + "epoch": 0.027530531980956324, + "grad_norm": 0.008342115208506584, + "learning_rate": 0.00019530243192683434, + "loss": 0.0063, + "step": 133 + }, + { + "epoch": 0.027737528462016145, + "grad_norm": 0.008312125690281391, + "learning_rate": 0.00019526086052795678, + "loss": 0.0047, + "step": 134 + }, + { + "epoch": 0.027944524943075966, + "grad_norm": 0.006928388494998217, + "learning_rate": 0.00019521928912907921, + "loss": 0.005, + "step": 135 + }, + { + "epoch": 0.02815152142413579, + "grad_norm": 0.0073064775206148624, + "learning_rate": 0.00019517771773020162, + "loss": 0.0061, + "step": 136 + }, + { + "epoch": 0.028358517905195613, + "grad_norm": 0.007849021814763546, + "learning_rate": 0.00019513614633132406, + "loss": 0.0052, + "step": 137 + }, + { + "epoch": 0.028565514386255434, + "grad_norm": 0.006330843083560467, + "learning_rate": 0.0001950945749324465, + "loss": 0.006, + "step": 138 + }, + { + "epoch": 0.028772510867315255, + "grad_norm": 0.002727820537984371, + "learning_rate": 0.00019505300353356894, + "loss": 0.0047, + "step": 139 + }, + { + "epoch": 0.028979507348375077, + "grad_norm": 0.006755925714969635, + "learning_rate": 0.00019501143213469135, + "loss": 0.0043, + "step": 140 + }, + { + "epoch": 0.0291865038294349, + "grad_norm": 0.007393544539809227, + "learning_rate": 0.00019496986073581376, + "loss": 0.0047, + "step": 141 + }, + { + "epoch": 0.029393500310494723, + "grad_norm": 0.007699685171246529, + "learning_rate": 0.0001949282893369362, + "loss": 0.0053, + "step": 142 + }, + { + "epoch": 0.029600496791554545, + "grad_norm": 0.003382055787369609, + "learning_rate": 0.0001948867179380586, + "loss": 0.0042, + "step": 143 + }, + { + "epoch": 0.029807493272614366, + "grad_norm": 0.01596757024526596, + "learning_rate": 0.00019484514653918104, + "loss": 0.0078, + "step": 144 + }, + { + "epoch": 0.030014489753674187, + "grad_norm": 0.00668082432821393, + "learning_rate": 0.00019480357514030348, + "loss": 0.0052, + "step": 145 + }, + { + "epoch": 0.03022148623473401, + "grad_norm": 0.007241010665893555, + "learning_rate": 0.00019476200374142591, + "loss": 0.0054, + "step": 146 + }, + { + "epoch": 0.03042848271579383, + "grad_norm": 0.009339329786598682, + "learning_rate": 0.00019472043234254832, + "loss": 0.0066, + "step": 147 + }, + { + "epoch": 0.030635479196853655, + "grad_norm": 0.007252382580190897, + "learning_rate": 0.00019467886094367076, + "loss": 0.0046, + "step": 148 + }, + { + "epoch": 0.030842475677913476, + "grad_norm": 0.004665658809244633, + "learning_rate": 0.0001946372895447932, + "loss": 0.0044, + "step": 149 + }, + { + "epoch": 0.031049472158973298, + "grad_norm": 0.006148173939436674, + "learning_rate": 0.0001945957181459156, + "loss": 0.0049, + "step": 150 + }, + { + "epoch": 0.03125646864003312, + "grad_norm": 0.004451967775821686, + "learning_rate": 0.00019455414674703805, + "loss": 0.0061, + "step": 151 + }, + { + "epoch": 0.031463465121092944, + "grad_norm": 0.004053746350109577, + "learning_rate": 0.00019451257534816048, + "loss": 0.0047, + "step": 152 + }, + { + "epoch": 0.03167046160215276, + "grad_norm": 0.009219355881214142, + "learning_rate": 0.0001944710039492829, + "loss": 0.0048, + "step": 153 + }, + { + "epoch": 0.03187745808321259, + "grad_norm": 0.020168175920844078, + "learning_rate": 0.00019442943255040533, + "loss": 0.0063, + "step": 154 + }, + { + "epoch": 0.032084454564272405, + "grad_norm": 0.0030542612075805664, + "learning_rate": 0.00019438786115152777, + "loss": 0.0048, + "step": 155 + }, + { + "epoch": 0.03229145104533223, + "grad_norm": 0.00561768002808094, + "learning_rate": 0.0001943462897526502, + "loss": 0.0047, + "step": 156 + }, + { + "epoch": 0.032498447526392055, + "grad_norm": 0.004265373572707176, + "learning_rate": 0.0001943047183537726, + "loss": 0.0062, + "step": 157 + }, + { + "epoch": 0.03270544400745187, + "grad_norm": 0.005523406434804201, + "learning_rate": 0.00019426314695489505, + "loss": 0.0044, + "step": 158 + }, + { + "epoch": 0.0329124404885117, + "grad_norm": 0.009324849583208561, + "learning_rate": 0.0001942215755560175, + "loss": 0.0055, + "step": 159 + }, + { + "epoch": 0.033119436969571515, + "grad_norm": 0.012468270026147366, + "learning_rate": 0.0001941800041571399, + "loss": 0.0045, + "step": 160 + }, + { + "epoch": 0.03332643345063134, + "grad_norm": 0.0018690524157136679, + "learning_rate": 0.00019413843275826233, + "loss": 0.0047, + "step": 161 + }, + { + "epoch": 0.03353342993169116, + "grad_norm": 0.009851769544184208, + "learning_rate": 0.00019409686135938477, + "loss": 0.0052, + "step": 162 + }, + { + "epoch": 0.03374042641275098, + "grad_norm": 0.0014802832156419754, + "learning_rate": 0.00019405528996050718, + "loss": 0.0045, + "step": 163 + }, + { + "epoch": 0.03394742289381081, + "grad_norm": 0.00177583540789783, + "learning_rate": 0.0001940137185616296, + "loss": 0.0046, + "step": 164 + }, + { + "epoch": 0.034154419374870626, + "grad_norm": 0.0017039773520082235, + "learning_rate": 0.00019397214716275203, + "loss": 0.0044, + "step": 165 + }, + { + "epoch": 0.03436141585593045, + "grad_norm": 0.010193880647420883, + "learning_rate": 0.00019393057576387447, + "loss": 0.0054, + "step": 166 + }, + { + "epoch": 0.03456841233699027, + "grad_norm": 0.011056206189095974, + "learning_rate": 0.00019388900436499688, + "loss": 0.0051, + "step": 167 + }, + { + "epoch": 0.034775408818050094, + "grad_norm": 0.004655253142118454, + "learning_rate": 0.0001938474329661193, + "loss": 0.0046, + "step": 168 + }, + { + "epoch": 0.03498240529910992, + "grad_norm": 0.0035964485723525286, + "learning_rate": 0.00019380586156724175, + "loss": 0.0045, + "step": 169 + }, + { + "epoch": 0.035189401780169736, + "grad_norm": 0.0026267431676387787, + "learning_rate": 0.00019376429016836416, + "loss": 0.0047, + "step": 170 + }, + { + "epoch": 0.03539639826122956, + "grad_norm": 0.008452721871435642, + "learning_rate": 0.0001937227187694866, + "loss": 0.0059, + "step": 171 + }, + { + "epoch": 0.03560339474228938, + "grad_norm": 0.006845233030617237, + "learning_rate": 0.00019368114737060903, + "loss": 0.0051, + "step": 172 + }, + { + "epoch": 0.035810391223349204, + "grad_norm": 0.005468891002237797, + "learning_rate": 0.00019363957597173144, + "loss": 0.0059, + "step": 173 + }, + { + "epoch": 0.03601738770440902, + "grad_norm": 0.007444227579981089, + "learning_rate": 0.00019359800457285388, + "loss": 0.0044, + "step": 174 + }, + { + "epoch": 0.03622438418546885, + "grad_norm": 0.011544904671609402, + "learning_rate": 0.00019355643317397632, + "loss": 0.0053, + "step": 175 + }, + { + "epoch": 0.03643138066652867, + "grad_norm": 0.04777868092060089, + "learning_rate": 0.00019351486177509875, + "loss": 0.005, + "step": 176 + }, + { + "epoch": 0.03663837714758849, + "grad_norm": 0.004177747759968042, + "learning_rate": 0.00019347329037622116, + "loss": 0.0048, + "step": 177 + }, + { + "epoch": 0.036845373628648315, + "grad_norm": 0.0020709133241325617, + "learning_rate": 0.0001934317189773436, + "loss": 0.0045, + "step": 178 + }, + { + "epoch": 0.03705237010970813, + "grad_norm": 0.012106567621231079, + "learning_rate": 0.00019339014757846604, + "loss": 0.0043, + "step": 179 + }, + { + "epoch": 0.03725936659076796, + "grad_norm": 0.005318734794855118, + "learning_rate": 0.00019334857617958845, + "loss": 0.005, + "step": 180 + }, + { + "epoch": 0.03746636307182778, + "grad_norm": 0.04815113916993141, + "learning_rate": 0.00019330700478071089, + "loss": 0.0056, + "step": 181 + }, + { + "epoch": 0.0376733595528876, + "grad_norm": 0.029821882024407387, + "learning_rate": 0.00019326543338183332, + "loss": 0.0053, + "step": 182 + }, + { + "epoch": 0.037880356033947425, + "grad_norm": 0.010615061037242413, + "learning_rate": 0.00019322386198295573, + "loss": 0.0049, + "step": 183 + }, + { + "epoch": 0.03808735251500724, + "grad_norm": 0.00788772851228714, + "learning_rate": 0.00019318229058407817, + "loss": 0.0043, + "step": 184 + }, + { + "epoch": 0.03829434899606707, + "grad_norm": 0.0045401486568152905, + "learning_rate": 0.00019314071918520058, + "loss": 0.0056, + "step": 185 + }, + { + "epoch": 0.038501345477126886, + "grad_norm": 0.004325198009610176, + "learning_rate": 0.00019309914778632302, + "loss": 0.0059, + "step": 186 + }, + { + "epoch": 0.03870834195818671, + "grad_norm": 0.012164851650595665, + "learning_rate": 0.00019305757638744543, + "loss": 0.0046, + "step": 187 + }, + { + "epoch": 0.038915338439246536, + "grad_norm": 0.02496037259697914, + "learning_rate": 0.00019301600498856786, + "loss": 0.0068, + "step": 188 + }, + { + "epoch": 0.039122334920306354, + "grad_norm": 0.011729522608220577, + "learning_rate": 0.0001929744335896903, + "loss": 0.0047, + "step": 189 + }, + { + "epoch": 0.03932933140136618, + "grad_norm": 0.003992550540715456, + "learning_rate": 0.0001929328621908127, + "loss": 0.0047, + "step": 190 + }, + { + "epoch": 0.039536327882425996, + "grad_norm": 0.012220533564686775, + "learning_rate": 0.00019289129079193515, + "loss": 0.0059, + "step": 191 + }, + { + "epoch": 0.03974332436348582, + "grad_norm": 0.0037318835966289043, + "learning_rate": 0.00019284971939305759, + "loss": 0.0049, + "step": 192 + }, + { + "epoch": 0.039950320844545646, + "grad_norm": 0.008259239606559277, + "learning_rate": 0.00019280814799418002, + "loss": 0.0047, + "step": 193 + }, + { + "epoch": 0.040157317325605464, + "grad_norm": 0.0029760177712887526, + "learning_rate": 0.00019276657659530243, + "loss": 0.0046, + "step": 194 + }, + { + "epoch": 0.04036431380666529, + "grad_norm": 0.007835526019334793, + "learning_rate": 0.00019272500519642487, + "loss": 0.0045, + "step": 195 + }, + { + "epoch": 0.04057131028772511, + "grad_norm": 0.019788436591625214, + "learning_rate": 0.0001926834337975473, + "loss": 0.0056, + "step": 196 + }, + { + "epoch": 0.04077830676878493, + "grad_norm": 0.0054263221099972725, + "learning_rate": 0.00019264186239866972, + "loss": 0.0047, + "step": 197 + }, + { + "epoch": 0.04098530324984475, + "grad_norm": 0.006738686002790928, + "learning_rate": 0.00019260029099979215, + "loss": 0.0053, + "step": 198 + }, + { + "epoch": 0.041192299730904575, + "grad_norm": 0.0019211384933441877, + "learning_rate": 0.0001925587196009146, + "loss": 0.0046, + "step": 199 + }, + { + "epoch": 0.0413992962119644, + "grad_norm": 0.0064520747400820255, + "learning_rate": 0.000192517148202037, + "loss": 0.0049, + "step": 200 + }, + { + "epoch": 0.04160629269302422, + "grad_norm": 0.0034363584127277136, + "learning_rate": 0.00019247557680315944, + "loss": 0.0048, + "step": 201 + }, + { + "epoch": 0.04181328917408404, + "grad_norm": 0.021868525072932243, + "learning_rate": 0.00019243400540428187, + "loss": 0.0055, + "step": 202 + }, + { + "epoch": 0.04202028565514386, + "grad_norm": 0.011735360138118267, + "learning_rate": 0.0001923924340054043, + "loss": 0.0059, + "step": 203 + }, + { + "epoch": 0.042227282136203685, + "grad_norm": 0.008518829010426998, + "learning_rate": 0.00019235086260652672, + "loss": 0.0062, + "step": 204 + }, + { + "epoch": 0.0424342786172635, + "grad_norm": 0.007598051335662603, + "learning_rate": 0.00019230929120764916, + "loss": 0.005, + "step": 205 + }, + { + "epoch": 0.04264127509832333, + "grad_norm": 0.004902805667370558, + "learning_rate": 0.0001922677198087716, + "loss": 0.0047, + "step": 206 + }, + { + "epoch": 0.04284827157938315, + "grad_norm": 0.01092604547739029, + "learning_rate": 0.00019222614840989398, + "loss": 0.0062, + "step": 207 + }, + { + "epoch": 0.04305526806044297, + "grad_norm": 0.017829621210694313, + "learning_rate": 0.00019218457701101642, + "loss": 0.0069, + "step": 208 + }, + { + "epoch": 0.043262264541502796, + "grad_norm": 0.0056928307749331, + "learning_rate": 0.00019214300561213885, + "loss": 0.0052, + "step": 209 + }, + { + "epoch": 0.043469261022562614, + "grad_norm": 0.0032917701173573732, + "learning_rate": 0.0001921014342132613, + "loss": 0.0049, + "step": 210 + }, + { + "epoch": 0.04367625750362244, + "grad_norm": 0.00267777475528419, + "learning_rate": 0.0001920598628143837, + "loss": 0.0047, + "step": 211 + }, + { + "epoch": 0.04388325398468226, + "grad_norm": 0.005063153337687254, + "learning_rate": 0.00019201829141550614, + "loss": 0.0065, + "step": 212 + }, + { + "epoch": 0.04409025046574208, + "grad_norm": 0.007821328938007355, + "learning_rate": 0.00019197672001662857, + "loss": 0.0073, + "step": 213 + }, + { + "epoch": 0.044297246946801906, + "grad_norm": 0.0022040351759642363, + "learning_rate": 0.00019193514861775098, + "loss": 0.0042, + "step": 214 + }, + { + "epoch": 0.044504243427861724, + "grad_norm": 0.010549996979534626, + "learning_rate": 0.00019189357721887342, + "loss": 0.0048, + "step": 215 + }, + { + "epoch": 0.04471123990892155, + "grad_norm": 0.003572909627109766, + "learning_rate": 0.00019185200581999586, + "loss": 0.0048, + "step": 216 + }, + { + "epoch": 0.04491823638998137, + "grad_norm": 0.0047572036273777485, + "learning_rate": 0.00019181043442111827, + "loss": 0.0046, + "step": 217 + }, + { + "epoch": 0.04512523287104119, + "grad_norm": 0.003330536652356386, + "learning_rate": 0.0001917688630222407, + "loss": 0.0043, + "step": 218 + }, + { + "epoch": 0.04533222935210102, + "grad_norm": 0.003489327384158969, + "learning_rate": 0.00019172729162336314, + "loss": 0.0044, + "step": 219 + }, + { + "epoch": 0.045539225833160835, + "grad_norm": 0.006631118711084127, + "learning_rate": 0.00019168572022448555, + "loss": 0.0046, + "step": 220 + }, + { + "epoch": 0.04574622231422066, + "grad_norm": 0.0016155489720404148, + "learning_rate": 0.000191644148825608, + "loss": 0.0045, + "step": 221 + }, + { + "epoch": 0.04595321879528048, + "grad_norm": 0.00484581058844924, + "learning_rate": 0.00019160257742673043, + "loss": 0.0049, + "step": 222 + }, + { + "epoch": 0.0461602152763403, + "grad_norm": 0.016010191291570663, + "learning_rate": 0.00019156100602785286, + "loss": 0.0052, + "step": 223 + }, + { + "epoch": 0.04636721175740013, + "grad_norm": 0.0038696257397532463, + "learning_rate": 0.00019151943462897527, + "loss": 0.0046, + "step": 224 + }, + { + "epoch": 0.046574208238459945, + "grad_norm": 0.003132582874968648, + "learning_rate": 0.0001914778632300977, + "loss": 0.0055, + "step": 225 + }, + { + "epoch": 0.04678120471951977, + "grad_norm": 0.003856977680698037, + "learning_rate": 0.00019143629183122015, + "loss": 0.0048, + "step": 226 + }, + { + "epoch": 0.04698820120057959, + "grad_norm": 0.0084042027592659, + "learning_rate": 0.00019139472043234256, + "loss": 0.005, + "step": 227 + }, + { + "epoch": 0.04719519768163941, + "grad_norm": 0.004033537581562996, + "learning_rate": 0.000191353149033465, + "loss": 0.0053, + "step": 228 + }, + { + "epoch": 0.04740219416269923, + "grad_norm": 0.002745938254520297, + "learning_rate": 0.0001913115776345874, + "loss": 0.005, + "step": 229 + }, + { + "epoch": 0.047609190643759056, + "grad_norm": 0.0033783107064664364, + "learning_rate": 0.00019127000623570984, + "loss": 0.0045, + "step": 230 + }, + { + "epoch": 0.04781618712481888, + "grad_norm": 0.011202207766473293, + "learning_rate": 0.00019122843483683225, + "loss": 0.0054, + "step": 231 + }, + { + "epoch": 0.0480231836058787, + "grad_norm": 0.006426738575100899, + "learning_rate": 0.0001911868634379547, + "loss": 0.0049, + "step": 232 + }, + { + "epoch": 0.04823018008693852, + "grad_norm": 0.0026592197827994823, + "learning_rate": 0.00019114529203907713, + "loss": 0.0045, + "step": 233 + }, + { + "epoch": 0.04843717656799834, + "grad_norm": 0.0044938791543245316, + "learning_rate": 0.00019110372064019954, + "loss": 0.0048, + "step": 234 + }, + { + "epoch": 0.048644173049058166, + "grad_norm": 0.004537639208137989, + "learning_rate": 0.00019106214924132197, + "loss": 0.0043, + "step": 235 + }, + { + "epoch": 0.04885116953011799, + "grad_norm": 0.007899290882050991, + "learning_rate": 0.0001910205778424444, + "loss": 0.0047, + "step": 236 + }, + { + "epoch": 0.04905816601117781, + "grad_norm": 0.003346335142850876, + "learning_rate": 0.00019097900644356682, + "loss": 0.0046, + "step": 237 + }, + { + "epoch": 0.049265162492237634, + "grad_norm": 0.012666060589253902, + "learning_rate": 0.00019093743504468926, + "loss": 0.0051, + "step": 238 + }, + { + "epoch": 0.04947215897329745, + "grad_norm": 0.005454343743622303, + "learning_rate": 0.0001908958636458117, + "loss": 0.0052, + "step": 239 + }, + { + "epoch": 0.04967915545435728, + "grad_norm": 0.0051568858325481415, + "learning_rate": 0.00019085429224693413, + "loss": 0.0054, + "step": 240 + }, + { + "epoch": 0.049886151935417095, + "grad_norm": 0.005354621913284063, + "learning_rate": 0.00019081272084805654, + "loss": 0.0045, + "step": 241 + }, + { + "epoch": 0.05009314841647692, + "grad_norm": 0.00728578818961978, + "learning_rate": 0.00019077114944917898, + "loss": 0.0056, + "step": 242 + }, + { + "epoch": 0.050300144897536744, + "grad_norm": 0.012164080515503883, + "learning_rate": 0.00019072957805030141, + "loss": 0.0055, + "step": 243 + }, + { + "epoch": 0.05050714137859656, + "grad_norm": 0.006451157853007317, + "learning_rate": 0.00019068800665142382, + "loss": 0.0045, + "step": 244 + }, + { + "epoch": 0.05071413785965639, + "grad_norm": 0.04119205102324486, + "learning_rate": 0.00019064643525254626, + "loss": 0.0048, + "step": 245 + }, + { + "epoch": 0.050921134340716205, + "grad_norm": 0.006683278828859329, + "learning_rate": 0.0001906048638536687, + "loss": 0.0054, + "step": 246 + }, + { + "epoch": 0.05112813082177603, + "grad_norm": 0.004711155779659748, + "learning_rate": 0.0001905632924547911, + "loss": 0.0052, + "step": 247 + }, + { + "epoch": 0.051335127302835855, + "grad_norm": 0.022338053211569786, + "learning_rate": 0.00019052172105591355, + "loss": 0.0051, + "step": 248 + }, + { + "epoch": 0.05154212378389567, + "grad_norm": 0.0033723730593919754, + "learning_rate": 0.00019048014965703598, + "loss": 0.0041, + "step": 249 + }, + { + "epoch": 0.0517491202649555, + "grad_norm": 0.013185818679630756, + "learning_rate": 0.00019043857825815842, + "loss": 0.0071, + "step": 250 + }, + { + "epoch": 0.051956116746015316, + "grad_norm": 0.01329563558101654, + "learning_rate": 0.0001903970068592808, + "loss": 0.0047, + "step": 251 + }, + { + "epoch": 0.05216311322707514, + "grad_norm": 0.0038430816493928432, + "learning_rate": 0.00019035543546040324, + "loss": 0.0049, + "step": 252 + }, + { + "epoch": 0.05237010970813496, + "grad_norm": 0.006868777330964804, + "learning_rate": 0.00019031386406152568, + "loss": 0.0045, + "step": 253 + }, + { + "epoch": 0.05257710618919478, + "grad_norm": 0.002608188660815358, + "learning_rate": 0.0001902722926626481, + "loss": 0.005, + "step": 254 + }, + { + "epoch": 0.05278410267025461, + "grad_norm": 0.005354885943233967, + "learning_rate": 0.00019023072126377052, + "loss": 0.0058, + "step": 255 + }, + { + "epoch": 0.052991099151314426, + "grad_norm": 0.006729793269187212, + "learning_rate": 0.00019018914986489296, + "loss": 0.0056, + "step": 256 + }, + { + "epoch": 0.05319809563237425, + "grad_norm": 0.0042056431993842125, + "learning_rate": 0.0001901475784660154, + "loss": 0.0042, + "step": 257 + }, + { + "epoch": 0.05340509211343407, + "grad_norm": 0.006845582276582718, + "learning_rate": 0.0001901060070671378, + "loss": 0.0048, + "step": 258 + }, + { + "epoch": 0.053612088594493894, + "grad_norm": 0.004010177683085203, + "learning_rate": 0.00019006443566826025, + "loss": 0.005, + "step": 259 + }, + { + "epoch": 0.05381908507555372, + "grad_norm": 0.0019074059091508389, + "learning_rate": 0.00019002286426938268, + "loss": 0.005, + "step": 260 + }, + { + "epoch": 0.05402608155661354, + "grad_norm": 0.002619614126160741, + "learning_rate": 0.0001899812928705051, + "loss": 0.0043, + "step": 261 + }, + { + "epoch": 0.05423307803767336, + "grad_norm": 0.002715028589591384, + "learning_rate": 0.00018993972147162753, + "loss": 0.0046, + "step": 262 + }, + { + "epoch": 0.05444007451873318, + "grad_norm": 0.004556506406515837, + "learning_rate": 0.00018989815007274997, + "loss": 0.0046, + "step": 263 + }, + { + "epoch": 0.054647070999793004, + "grad_norm": 0.010189310647547245, + "learning_rate": 0.00018985657867387238, + "loss": 0.0047, + "step": 264 + }, + { + "epoch": 0.05485406748085282, + "grad_norm": 0.002120416844263673, + "learning_rate": 0.0001898150072749948, + "loss": 0.0051, + "step": 265 + }, + { + "epoch": 0.05506106396191265, + "grad_norm": 0.008936331607401371, + "learning_rate": 0.00018977343587611725, + "loss": 0.005, + "step": 266 + }, + { + "epoch": 0.05526806044297247, + "grad_norm": 0.0026866530533879995, + "learning_rate": 0.00018973186447723966, + "loss": 0.0047, + "step": 267 + }, + { + "epoch": 0.05547505692403229, + "grad_norm": 0.009859035722911358, + "learning_rate": 0.0001896902930783621, + "loss": 0.0052, + "step": 268 + }, + { + "epoch": 0.055682053405092115, + "grad_norm": 0.0024370807223021984, + "learning_rate": 0.00018964872167948453, + "loss": 0.0053, + "step": 269 + }, + { + "epoch": 0.05588904988615193, + "grad_norm": 0.006978074554353952, + "learning_rate": 0.00018960715028060697, + "loss": 0.0062, + "step": 270 + }, + { + "epoch": 0.05609604636721176, + "grad_norm": 0.0037223445251584053, + "learning_rate": 0.00018956557888172938, + "loss": 0.0047, + "step": 271 + }, + { + "epoch": 0.05630304284827158, + "grad_norm": 0.018946697935461998, + "learning_rate": 0.00018952400748285182, + "loss": 0.0054, + "step": 272 + }, + { + "epoch": 0.0565100393293314, + "grad_norm": 0.009405361488461494, + "learning_rate": 0.00018948243608397423, + "loss": 0.0043, + "step": 273 + }, + { + "epoch": 0.056717035810391225, + "grad_norm": 0.0008082574931904674, + "learning_rate": 0.00018944086468509667, + "loss": 0.0045, + "step": 274 + }, + { + "epoch": 0.05692403229145104, + "grad_norm": 0.004950170870870352, + "learning_rate": 0.00018939929328621908, + "loss": 0.0044, + "step": 275 + }, + { + "epoch": 0.05713102877251087, + "grad_norm": 0.010343370027840137, + "learning_rate": 0.0001893577218873415, + "loss": 0.0046, + "step": 276 + }, + { + "epoch": 0.057338025253570686, + "grad_norm": 0.005096866749227047, + "learning_rate": 0.00018931615048846395, + "loss": 0.005, + "step": 277 + }, + { + "epoch": 0.05754502173463051, + "grad_norm": 0.0033187547232955694, + "learning_rate": 0.00018927457908958636, + "loss": 0.0041, + "step": 278 + }, + { + "epoch": 0.057752018215690336, + "grad_norm": 0.0031861995812505484, + "learning_rate": 0.0001892330076907088, + "loss": 0.0059, + "step": 279 + }, + { + "epoch": 0.057959014696750154, + "grad_norm": 0.0053593809716403484, + "learning_rate": 0.00018919143629183123, + "loss": 0.0047, + "step": 280 + }, + { + "epoch": 0.05816601117780998, + "grad_norm": 0.005950809922069311, + "learning_rate": 0.00018914986489295364, + "loss": 0.0046, + "step": 281 + }, + { + "epoch": 0.0583730076588698, + "grad_norm": 0.006816082634031773, + "learning_rate": 0.00018910829349407608, + "loss": 0.0048, + "step": 282 + }, + { + "epoch": 0.05858000413992962, + "grad_norm": 0.005741049535572529, + "learning_rate": 0.00018906672209519852, + "loss": 0.0048, + "step": 283 + }, + { + "epoch": 0.058787000620989446, + "grad_norm": 0.004367714747786522, + "learning_rate": 0.00018902515069632093, + "loss": 0.0044, + "step": 284 + }, + { + "epoch": 0.058993997102049264, + "grad_norm": 0.0051370360888540745, + "learning_rate": 0.00018898357929744336, + "loss": 0.0049, + "step": 285 + }, + { + "epoch": 0.05920099358310909, + "grad_norm": 0.00557302962988615, + "learning_rate": 0.0001889420078985658, + "loss": 0.005, + "step": 286 + }, + { + "epoch": 0.05940799006416891, + "grad_norm": 0.004271439276635647, + "learning_rate": 0.00018890043649968824, + "loss": 0.0042, + "step": 287 + }, + { + "epoch": 0.05961498654522873, + "grad_norm": 0.00942255649715662, + "learning_rate": 0.00018885886510081065, + "loss": 0.005, + "step": 288 + }, + { + "epoch": 0.05982198302628855, + "grad_norm": 0.001850366359576583, + "learning_rate": 0.00018881729370193309, + "loss": 0.0049, + "step": 289 + }, + { + "epoch": 0.060028979507348375, + "grad_norm": 0.005905089899897575, + "learning_rate": 0.00018877572230305552, + "loss": 0.0049, + "step": 290 + }, + { + "epoch": 0.0602359759884082, + "grad_norm": 0.006148039363324642, + "learning_rate": 0.00018873415090417793, + "loss": 0.0058, + "step": 291 + }, + { + "epoch": 0.06044297246946802, + "grad_norm": 0.004467264749109745, + "learning_rate": 0.00018869257950530037, + "loss": 0.0046, + "step": 292 + }, + { + "epoch": 0.06064996895052784, + "grad_norm": 0.0026016500778496265, + "learning_rate": 0.0001886510081064228, + "loss": 0.0046, + "step": 293 + }, + { + "epoch": 0.06085696543158766, + "grad_norm": 0.00536306481808424, + "learning_rate": 0.00018860943670754522, + "loss": 0.0046, + "step": 294 + }, + { + "epoch": 0.061063961912647485, + "grad_norm": 0.00043015365372411907, + "learning_rate": 0.00018856786530866763, + "loss": 0.0051, + "step": 295 + }, + { + "epoch": 0.06127095839370731, + "grad_norm": 0.002035475103184581, + "learning_rate": 0.00018852629390979006, + "loss": 0.0044, + "step": 296 + }, + { + "epoch": 0.06147795487476713, + "grad_norm": 0.004757678601890802, + "learning_rate": 0.0001884847225109125, + "loss": 0.0055, + "step": 297 + }, + { + "epoch": 0.06168495135582695, + "grad_norm": 0.004153924528509378, + "learning_rate": 0.0001884431511120349, + "loss": 0.0047, + "step": 298 + }, + { + "epoch": 0.06189194783688677, + "grad_norm": 0.004693943541496992, + "learning_rate": 0.00018840157971315735, + "loss": 0.006, + "step": 299 + }, + { + "epoch": 0.062098944317946596, + "grad_norm": 0.005683131981641054, + "learning_rate": 0.00018836000831427979, + "loss": 0.0054, + "step": 300 + }, + { + "epoch": 0.062305940799006414, + "grad_norm": 0.01867171749472618, + "learning_rate": 0.0001883184369154022, + "loss": 0.0047, + "step": 301 + }, + { + "epoch": 0.06251293728006624, + "grad_norm": 0.010631178505718708, + "learning_rate": 0.00018827686551652463, + "loss": 0.0052, + "step": 302 + }, + { + "epoch": 0.06271993376112606, + "grad_norm": 0.0025544106028974056, + "learning_rate": 0.00018823529411764707, + "loss": 0.0044, + "step": 303 + }, + { + "epoch": 0.06292693024218589, + "grad_norm": 0.005773225799202919, + "learning_rate": 0.0001881937227187695, + "loss": 0.0047, + "step": 304 + }, + { + "epoch": 0.0631339267232457, + "grad_norm": 0.008623667992651463, + "learning_rate": 0.00018815215131989192, + "loss": 0.0041, + "step": 305 + }, + { + "epoch": 0.06334092320430552, + "grad_norm": 0.0023567613679915667, + "learning_rate": 0.00018811057992101435, + "loss": 0.0044, + "step": 306 + }, + { + "epoch": 0.06354791968536534, + "grad_norm": 0.005568039603531361, + "learning_rate": 0.0001880690085221368, + "loss": 0.0048, + "step": 307 + }, + { + "epoch": 0.06375491616642517, + "grad_norm": 0.006429716479033232, + "learning_rate": 0.0001880274371232592, + "loss": 0.0049, + "step": 308 + }, + { + "epoch": 0.06396191264748499, + "grad_norm": 0.0012258924543857574, + "learning_rate": 0.00018798586572438164, + "loss": 0.0046, + "step": 309 + }, + { + "epoch": 0.06416890912854481, + "grad_norm": 0.013900945894420147, + "learning_rate": 0.00018794429432550407, + "loss": 0.0054, + "step": 310 + }, + { + "epoch": 0.06437590560960464, + "grad_norm": 0.0164532121270895, + "learning_rate": 0.00018790272292662648, + "loss": 0.0052, + "step": 311 + }, + { + "epoch": 0.06458290209066446, + "grad_norm": 0.0039049319457262754, + "learning_rate": 0.00018786115152774892, + "loss": 0.0047, + "step": 312 + }, + { + "epoch": 0.06478989857172428, + "grad_norm": 0.0025435080751776695, + "learning_rate": 0.00018781958012887136, + "loss": 0.0045, + "step": 313 + }, + { + "epoch": 0.06499689505278411, + "grad_norm": 0.0066621373407542706, + "learning_rate": 0.00018777800872999377, + "loss": 0.0057, + "step": 314 + }, + { + "epoch": 0.06520389153384393, + "grad_norm": 0.0011938404059037566, + "learning_rate": 0.0001877364373311162, + "loss": 0.0045, + "step": 315 + }, + { + "epoch": 0.06541088801490375, + "grad_norm": 0.005898007657378912, + "learning_rate": 0.00018769486593223864, + "loss": 0.0047, + "step": 316 + }, + { + "epoch": 0.06561788449596356, + "grad_norm": 0.004224811680614948, + "learning_rate": 0.00018765329453336105, + "loss": 0.0059, + "step": 317 + }, + { + "epoch": 0.0658248809770234, + "grad_norm": 0.02106441557407379, + "learning_rate": 0.00018761172313448346, + "loss": 0.0051, + "step": 318 + }, + { + "epoch": 0.06603187745808321, + "grad_norm": 0.0010845439974218607, + "learning_rate": 0.0001875701517356059, + "loss": 0.0042, + "step": 319 + }, + { + "epoch": 0.06623887393914303, + "grad_norm": 0.007267239037901163, + "learning_rate": 0.00018752858033672834, + "loss": 0.0048, + "step": 320 + }, + { + "epoch": 0.06644587042020286, + "grad_norm": 0.0066713071428239346, + "learning_rate": 0.00018748700893785077, + "loss": 0.0049, + "step": 321 + }, + { + "epoch": 0.06665286690126268, + "grad_norm": 0.007623916491866112, + "learning_rate": 0.00018744543753897318, + "loss": 0.0052, + "step": 322 + }, + { + "epoch": 0.0668598633823225, + "grad_norm": 0.003484464716166258, + "learning_rate": 0.00018740386614009562, + "loss": 0.0045, + "step": 323 + }, + { + "epoch": 0.06706685986338232, + "grad_norm": 0.013743946328759193, + "learning_rate": 0.00018736229474121806, + "loss": 0.0049, + "step": 324 + }, + { + "epoch": 0.06727385634444215, + "grad_norm": 0.0030819710809737444, + "learning_rate": 0.00018732072334234047, + "loss": 0.0052, + "step": 325 + }, + { + "epoch": 0.06748085282550197, + "grad_norm": 0.014786194078624249, + "learning_rate": 0.0001872791519434629, + "loss": 0.0058, + "step": 326 + }, + { + "epoch": 0.06768784930656178, + "grad_norm": 0.007248689886182547, + "learning_rate": 0.00018723758054458534, + "loss": 0.0046, + "step": 327 + }, + { + "epoch": 0.06789484578762162, + "grad_norm": 0.010181601159274578, + "learning_rate": 0.00018719600914570775, + "loss": 0.0053, + "step": 328 + }, + { + "epoch": 0.06810184226868143, + "grad_norm": 0.004160667769610882, + "learning_rate": 0.0001871544377468302, + "loss": 0.0056, + "step": 329 + }, + { + "epoch": 0.06830883874974125, + "grad_norm": 0.0021253142040222883, + "learning_rate": 0.00018711286634795263, + "loss": 0.0045, + "step": 330 + }, + { + "epoch": 0.06851583523080107, + "grad_norm": 0.0058175004087388515, + "learning_rate": 0.00018707129494907504, + "loss": 0.0058, + "step": 331 + }, + { + "epoch": 0.0687228317118609, + "grad_norm": 0.005257429089397192, + "learning_rate": 0.00018702972355019747, + "loss": 0.0049, + "step": 332 + }, + { + "epoch": 0.06892982819292072, + "grad_norm": 0.003563474863767624, + "learning_rate": 0.0001869881521513199, + "loss": 0.0043, + "step": 333 + }, + { + "epoch": 0.06913682467398054, + "grad_norm": 0.002636804012581706, + "learning_rate": 0.00018694658075244235, + "loss": 0.0046, + "step": 334 + }, + { + "epoch": 0.06934382115504037, + "grad_norm": 0.009255892597138882, + "learning_rate": 0.00018690500935356476, + "loss": 0.0053, + "step": 335 + }, + { + "epoch": 0.06955081763610019, + "grad_norm": 0.0023986424785107374, + "learning_rate": 0.0001868634379546872, + "loss": 0.0045, + "step": 336 + }, + { + "epoch": 0.06975781411716, + "grad_norm": 0.006371774710714817, + "learning_rate": 0.00018682186655580963, + "loss": 0.0046, + "step": 337 + }, + { + "epoch": 0.06996481059821984, + "grad_norm": 0.009518152102828026, + "learning_rate": 0.00018678029515693204, + "loss": 0.0049, + "step": 338 + }, + { + "epoch": 0.07017180707927965, + "grad_norm": 0.003742037108168006, + "learning_rate": 0.00018673872375805445, + "loss": 0.0064, + "step": 339 + }, + { + "epoch": 0.07037880356033947, + "grad_norm": 0.009771923534572124, + "learning_rate": 0.0001866971523591769, + "loss": 0.0045, + "step": 340 + }, + { + "epoch": 0.07058580004139929, + "grad_norm": 0.01101437397301197, + "learning_rate": 0.00018665558096029933, + "loss": 0.0055, + "step": 341 + }, + { + "epoch": 0.07079279652245912, + "grad_norm": 0.008826150558888912, + "learning_rate": 0.00018661400956142174, + "loss": 0.005, + "step": 342 + }, + { + "epoch": 0.07099979300351894, + "grad_norm": 0.006243105512112379, + "learning_rate": 0.00018657243816254417, + "loss": 0.0048, + "step": 343 + }, + { + "epoch": 0.07120678948457876, + "grad_norm": 0.0014233957044780254, + "learning_rate": 0.0001865308667636666, + "loss": 0.0045, + "step": 344 + }, + { + "epoch": 0.07141378596563859, + "grad_norm": 0.002639338606968522, + "learning_rate": 0.00018648929536478902, + "loss": 0.0053, + "step": 345 + }, + { + "epoch": 0.07162078244669841, + "grad_norm": 0.003536937525495887, + "learning_rate": 0.00018644772396591146, + "loss": 0.0048, + "step": 346 + }, + { + "epoch": 0.07182777892775823, + "grad_norm": 0.0018274744506925344, + "learning_rate": 0.0001864061525670339, + "loss": 0.0043, + "step": 347 + }, + { + "epoch": 0.07203477540881804, + "grad_norm": 0.004306804854422808, + "learning_rate": 0.0001863645811681563, + "loss": 0.0058, + "step": 348 + }, + { + "epoch": 0.07224177188987788, + "grad_norm": 0.003877132898196578, + "learning_rate": 0.00018632300976927874, + "loss": 0.0058, + "step": 349 + }, + { + "epoch": 0.0724487683709377, + "grad_norm": 0.0018924670293927193, + "learning_rate": 0.00018628143837040118, + "loss": 0.0044, + "step": 350 + }, + { + "epoch": 0.07265576485199751, + "grad_norm": 0.005626944359391928, + "learning_rate": 0.00018623986697152361, + "loss": 0.0046, + "step": 351 + }, + { + "epoch": 0.07286276133305734, + "grad_norm": 0.006948824506253004, + "learning_rate": 0.00018619829557264602, + "loss": 0.0057, + "step": 352 + }, + { + "epoch": 0.07306975781411716, + "grad_norm": 0.002097270218655467, + "learning_rate": 0.00018615672417376846, + "loss": 0.0047, + "step": 353 + }, + { + "epoch": 0.07327675429517698, + "grad_norm": 0.0013399182353168726, + "learning_rate": 0.0001861151527748909, + "loss": 0.0047, + "step": 354 + }, + { + "epoch": 0.0734837507762368, + "grad_norm": 0.010953530669212341, + "learning_rate": 0.0001860735813760133, + "loss": 0.0061, + "step": 355 + }, + { + "epoch": 0.07369074725729663, + "grad_norm": 0.0024627153761684895, + "learning_rate": 0.00018603200997713575, + "loss": 0.0061, + "step": 356 + }, + { + "epoch": 0.07389774373835645, + "grad_norm": 0.002271963283419609, + "learning_rate": 0.00018599043857825818, + "loss": 0.0043, + "step": 357 + }, + { + "epoch": 0.07410474021941627, + "grad_norm": 0.0036786317359656096, + "learning_rate": 0.0001859488671793806, + "loss": 0.0044, + "step": 358 + }, + { + "epoch": 0.0743117367004761, + "grad_norm": 0.006179209798574448, + "learning_rate": 0.00018590729578050303, + "loss": 0.0049, + "step": 359 + }, + { + "epoch": 0.07451873318153591, + "grad_norm": 0.0028029060922563076, + "learning_rate": 0.00018586572438162547, + "loss": 0.0051, + "step": 360 + }, + { + "epoch": 0.07472572966259573, + "grad_norm": 0.003495444543659687, + "learning_rate": 0.00018582415298274788, + "loss": 0.0055, + "step": 361 + }, + { + "epoch": 0.07493272614365556, + "grad_norm": 0.005056194495409727, + "learning_rate": 0.0001857825815838703, + "loss": 0.0052, + "step": 362 + }, + { + "epoch": 0.07513972262471538, + "grad_norm": 0.005216763820499182, + "learning_rate": 0.00018574101018499272, + "loss": 0.0064, + "step": 363 + }, + { + "epoch": 0.0753467191057752, + "grad_norm": 0.006161578465253115, + "learning_rate": 0.00018569943878611516, + "loss": 0.0051, + "step": 364 + }, + { + "epoch": 0.07555371558683502, + "grad_norm": 0.0022837778087705374, + "learning_rate": 0.00018565786738723757, + "loss": 0.0046, + "step": 365 + }, + { + "epoch": 0.07576071206789485, + "grad_norm": 0.0017736790468916297, + "learning_rate": 0.00018561629598836, + "loss": 0.0043, + "step": 366 + }, + { + "epoch": 0.07596770854895467, + "grad_norm": 0.004538076464086771, + "learning_rate": 0.00018557472458948245, + "loss": 0.0048, + "step": 367 + }, + { + "epoch": 0.07617470503001449, + "grad_norm": 0.0018419534899294376, + "learning_rate": 0.00018553315319060488, + "loss": 0.0043, + "step": 368 + }, + { + "epoch": 0.07638170151107432, + "grad_norm": 0.004529993515461683, + "learning_rate": 0.0001854915817917273, + "loss": 0.0044, + "step": 369 + }, + { + "epoch": 0.07658869799213414, + "grad_norm": 0.0072746858932077885, + "learning_rate": 0.00018545001039284973, + "loss": 0.0045, + "step": 370 + }, + { + "epoch": 0.07679569447319395, + "grad_norm": 0.013997476547956467, + "learning_rate": 0.00018540843899397217, + "loss": 0.0048, + "step": 371 + }, + { + "epoch": 0.07700269095425377, + "grad_norm": 0.0026135060470551252, + "learning_rate": 0.00018536686759509458, + "loss": 0.0043, + "step": 372 + }, + { + "epoch": 0.0772096874353136, + "grad_norm": 0.008647504262626171, + "learning_rate": 0.000185325296196217, + "loss": 0.0043, + "step": 373 + }, + { + "epoch": 0.07741668391637342, + "grad_norm": 0.0028080667834728956, + "learning_rate": 0.00018528372479733945, + "loss": 0.0045, + "step": 374 + }, + { + "epoch": 0.07762368039743324, + "grad_norm": 0.00593935651704669, + "learning_rate": 0.00018524215339846186, + "loss": 0.0043, + "step": 375 + }, + { + "epoch": 0.07783067687849307, + "grad_norm": 0.005520394071936607, + "learning_rate": 0.0001852005819995843, + "loss": 0.0044, + "step": 376 + }, + { + "epoch": 0.07803767335955289, + "grad_norm": 0.004040780942887068, + "learning_rate": 0.00018515901060070673, + "loss": 0.0044, + "step": 377 + }, + { + "epoch": 0.07824466984061271, + "grad_norm": 0.006418270990252495, + "learning_rate": 0.00018511743920182914, + "loss": 0.0044, + "step": 378 + }, + { + "epoch": 0.07845166632167253, + "grad_norm": 0.01350860670208931, + "learning_rate": 0.00018507586780295158, + "loss": 0.0044, + "step": 379 + }, + { + "epoch": 0.07865866280273236, + "grad_norm": 0.011060641147196293, + "learning_rate": 0.00018503429640407402, + "loss": 0.004, + "step": 380 + }, + { + "epoch": 0.07886565928379217, + "grad_norm": 0.004036550410091877, + "learning_rate": 0.00018499272500519646, + "loss": 0.0038, + "step": 381 + }, + { + "epoch": 0.07907265576485199, + "grad_norm": 0.003959359601140022, + "learning_rate": 0.00018495115360631887, + "loss": 0.0038, + "step": 382 + }, + { + "epoch": 0.07927965224591182, + "grad_norm": 0.005686459131538868, + "learning_rate": 0.00018490958220744128, + "loss": 0.0047, + "step": 383 + }, + { + "epoch": 0.07948664872697164, + "grad_norm": 0.007630357053130865, + "learning_rate": 0.0001848680108085637, + "loss": 0.0042, + "step": 384 + }, + { + "epoch": 0.07969364520803146, + "grad_norm": 0.005852843634784222, + "learning_rate": 0.00018482643940968615, + "loss": 0.0036, + "step": 385 + }, + { + "epoch": 0.07990064168909129, + "grad_norm": 0.005719276610761881, + "learning_rate": 0.00018478486801080856, + "loss": 0.0033, + "step": 386 + }, + { + "epoch": 0.08010763817015111, + "grad_norm": 0.007002281956374645, + "learning_rate": 0.000184743296611931, + "loss": 0.0033, + "step": 387 + }, + { + "epoch": 0.08031463465121093, + "grad_norm": 0.007220590952783823, + "learning_rate": 0.00018470172521305343, + "loss": 0.0046, + "step": 388 + }, + { + "epoch": 0.08052163113227075, + "grad_norm": 0.005552125629037619, + "learning_rate": 0.00018466015381417584, + "loss": 0.0028, + "step": 389 + }, + { + "epoch": 0.08072862761333058, + "grad_norm": 0.011994168162345886, + "learning_rate": 0.00018461858241529828, + "loss": 0.0028, + "step": 390 + }, + { + "epoch": 0.0809356240943904, + "grad_norm": 0.011316240765154362, + "learning_rate": 0.00018457701101642072, + "loss": 0.0026, + "step": 391 + }, + { + "epoch": 0.08114262057545021, + "grad_norm": 0.006591492332518101, + "learning_rate": 0.00018453543961754313, + "loss": 0.0022, + "step": 392 + }, + { + "epoch": 0.08134961705651005, + "grad_norm": 0.007715560495853424, + "learning_rate": 0.00018449386821866556, + "loss": 0.0032, + "step": 393 + }, + { + "epoch": 0.08155661353756986, + "grad_norm": 0.009517872706055641, + "learning_rate": 0.000184452296819788, + "loss": 0.0016, + "step": 394 + }, + { + "epoch": 0.08176361001862968, + "grad_norm": 0.79290372133255, + "learning_rate": 0.0001844107254209104, + "loss": 0.0117, + "step": 395 + }, + { + "epoch": 0.0819706064996895, + "grad_norm": 0.10478183627128601, + "learning_rate": 0.00018436915402203285, + "loss": 0.0041, + "step": 396 + }, + { + "epoch": 0.08217760298074933, + "grad_norm": 0.11330251395702362, + "learning_rate": 0.00018432758262315529, + "loss": 0.0349, + "step": 397 + }, + { + "epoch": 0.08238459946180915, + "grad_norm": 0.022089608013629913, + "learning_rate": 0.00018428601122427772, + "loss": 0.0028, + "step": 398 + }, + { + "epoch": 0.08259159594286897, + "grad_norm": 0.045795392245054245, + "learning_rate": 0.00018424443982540013, + "loss": 0.0027, + "step": 399 + }, + { + "epoch": 0.0827985924239288, + "grad_norm": 0.052710726857185364, + "learning_rate": 0.00018420286842652257, + "loss": 0.0033, + "step": 400 + }, + { + "epoch": 0.08300558890498862, + "grad_norm": 0.0732388123869896, + "learning_rate": 0.000184161297027645, + "loss": 0.0031, + "step": 401 + }, + { + "epoch": 0.08321258538604843, + "grad_norm": 0.03560757264494896, + "learning_rate": 0.00018411972562876742, + "loss": 0.0025, + "step": 402 + }, + { + "epoch": 0.08341958186710825, + "grad_norm": 0.039032500237226486, + "learning_rate": 0.00018407815422988985, + "loss": 0.0027, + "step": 403 + }, + { + "epoch": 0.08362657834816808, + "grad_norm": 0.018673432990908623, + "learning_rate": 0.0001840365828310123, + "loss": 0.0018, + "step": 404 + }, + { + "epoch": 0.0838335748292279, + "grad_norm": 0.06421470642089844, + "learning_rate": 0.0001839950114321347, + "loss": 0.0293, + "step": 405 + }, + { + "epoch": 0.08404057131028772, + "grad_norm": 0.05356355383992195, + "learning_rate": 0.0001839534400332571, + "loss": 0.0234, + "step": 406 + }, + { + "epoch": 0.08424756779134755, + "grad_norm": 0.047022175043821335, + "learning_rate": 0.00018391186863437955, + "loss": 0.0198, + "step": 407 + }, + { + "epoch": 0.08445456427240737, + "grad_norm": 0.017866840586066246, + "learning_rate": 0.00018387029723550199, + "loss": 0.0017, + "step": 408 + }, + { + "epoch": 0.08466156075346719, + "grad_norm": 0.010899499990046024, + "learning_rate": 0.0001838287258366244, + "loss": 0.0013, + "step": 409 + }, + { + "epoch": 0.084868557234527, + "grad_norm": 0.0167918112128973, + "learning_rate": 0.00018378715443774683, + "loss": 0.0024, + "step": 410 + }, + { + "epoch": 0.08507555371558684, + "grad_norm": 0.02067534811794758, + "learning_rate": 0.00018374558303886927, + "loss": 0.0015, + "step": 411 + }, + { + "epoch": 0.08528255019664666, + "grad_norm": 0.01670040749013424, + "learning_rate": 0.00018370401163999168, + "loss": 0.0015, + "step": 412 + }, + { + "epoch": 0.08548954667770647, + "grad_norm": 0.008924894034862518, + "learning_rate": 0.00018366244024111412, + "loss": 0.0029, + "step": 413 + }, + { + "epoch": 0.0856965431587663, + "grad_norm": 0.013602840714156628, + "learning_rate": 0.00018362086884223655, + "loss": 0.0009, + "step": 414 + }, + { + "epoch": 0.08590353963982612, + "grad_norm": 0.013082594610750675, + "learning_rate": 0.000183579297443359, + "loss": 0.002, + "step": 415 + }, + { + "epoch": 0.08611053612088594, + "grad_norm": 0.012215960770845413, + "learning_rate": 0.0001835377260444814, + "loss": 0.0017, + "step": 416 + }, + { + "epoch": 0.08631753260194577, + "grad_norm": 0.16738812625408173, + "learning_rate": 0.00018349615464560384, + "loss": 0.0173, + "step": 417 + }, + { + "epoch": 0.08652452908300559, + "grad_norm": 0.006629611365497112, + "learning_rate": 0.00018345458324672627, + "loss": 0.0006, + "step": 418 + }, + { + "epoch": 0.08673152556406541, + "grad_norm": 0.006643650587648153, + "learning_rate": 0.00018341301184784868, + "loss": 0.0015, + "step": 419 + }, + { + "epoch": 0.08693852204512523, + "grad_norm": 0.0039656031876802444, + "learning_rate": 0.00018337144044897112, + "loss": 0.0003, + "step": 420 + }, + { + "epoch": 0.08714551852618506, + "grad_norm": 0.00564931146800518, + "learning_rate": 0.00018332986905009356, + "loss": 0.0004, + "step": 421 + }, + { + "epoch": 0.08735251500724488, + "grad_norm": 0.014363352209329605, + "learning_rate": 0.00018328829765121597, + "loss": 0.0006, + "step": 422 + }, + { + "epoch": 0.0875595114883047, + "grad_norm": 0.006862149108201265, + "learning_rate": 0.0001832467262523384, + "loss": 0.0004, + "step": 423 + }, + { + "epoch": 0.08776650796936453, + "grad_norm": 0.005224172957241535, + "learning_rate": 0.00018320515485346084, + "loss": 0.0003, + "step": 424 + }, + { + "epoch": 0.08797350445042434, + "grad_norm": 0.012813829816877842, + "learning_rate": 0.00018316358345458325, + "loss": 0.0005, + "step": 425 + }, + { + "epoch": 0.08818050093148416, + "grad_norm": 0.0045601376332342625, + "learning_rate": 0.0001831220120557057, + "loss": 0.0003, + "step": 426 + }, + { + "epoch": 0.08838749741254398, + "grad_norm": 0.002229505218565464, + "learning_rate": 0.0001830804406568281, + "loss": 0.0002, + "step": 427 + }, + { + "epoch": 0.08859449389360381, + "grad_norm": 0.005202361848205328, + "learning_rate": 0.00018303886925795054, + "loss": 0.0003, + "step": 428 + }, + { + "epoch": 0.08880149037466363, + "grad_norm": 0.010837195441126823, + "learning_rate": 0.00018299729785907295, + "loss": 0.0016, + "step": 429 + }, + { + "epoch": 0.08900848685572345, + "grad_norm": 0.006401981692761183, + "learning_rate": 0.00018295572646019538, + "loss": 0.0002, + "step": 430 + }, + { + "epoch": 0.08921548333678328, + "grad_norm": 0.0025153057649731636, + "learning_rate": 0.00018291415506131782, + "loss": 0.0001, + "step": 431 + }, + { + "epoch": 0.0894224798178431, + "grad_norm": 0.009693821892142296, + "learning_rate": 0.00018287258366244023, + "loss": 0.0004, + "step": 432 + }, + { + "epoch": 0.08962947629890292, + "grad_norm": 0.0013723783195018768, + "learning_rate": 0.00018283101226356267, + "loss": 0.0001, + "step": 433 + }, + { + "epoch": 0.08983647277996273, + "grad_norm": 0.008555575273931026, + "learning_rate": 0.0001827894408646851, + "loss": 0.0018, + "step": 434 + }, + { + "epoch": 0.09004346926102257, + "grad_norm": 0.0028277519159018993, + "learning_rate": 0.00018274786946580754, + "loss": 0.0002, + "step": 435 + }, + { + "epoch": 0.09025046574208238, + "grad_norm": 0.014325006864964962, + "learning_rate": 0.00018270629806692995, + "loss": 0.0009, + "step": 436 + }, + { + "epoch": 0.0904574622231422, + "grad_norm": 0.008406553417444229, + "learning_rate": 0.0001826647266680524, + "loss": 0.0012, + "step": 437 + }, + { + "epoch": 0.09066445870420203, + "grad_norm": 0.0018985685892403126, + "learning_rate": 0.00018262315526917483, + "loss": 0.0002, + "step": 438 + }, + { + "epoch": 0.09087145518526185, + "grad_norm": 0.008550492115318775, + "learning_rate": 0.00018258158387029724, + "loss": 0.0024, + "step": 439 + }, + { + "epoch": 0.09107845166632167, + "grad_norm": 0.0008987212786450982, + "learning_rate": 0.00018254001247141967, + "loss": 0.0001, + "step": 440 + }, + { + "epoch": 0.0912854481473815, + "grad_norm": 0.002059886697679758, + "learning_rate": 0.0001824984410725421, + "loss": 0.0001, + "step": 441 + }, + { + "epoch": 0.09149244462844132, + "grad_norm": 0.003429644973948598, + "learning_rate": 0.00018245686967366452, + "loss": 0.0001, + "step": 442 + }, + { + "epoch": 0.09169944110950114, + "grad_norm": 0.0026945085264742374, + "learning_rate": 0.00018241529827478696, + "loss": 0.0011, + "step": 443 + }, + { + "epoch": 0.09190643759056095, + "grad_norm": 0.00299448031000793, + "learning_rate": 0.0001823737268759094, + "loss": 0.0001, + "step": 444 + }, + { + "epoch": 0.09211343407162079, + "grad_norm": 0.004376763943582773, + "learning_rate": 0.00018233215547703183, + "loss": 0.0011, + "step": 445 + }, + { + "epoch": 0.0923204305526806, + "grad_norm": 0.005078963004052639, + "learning_rate": 0.00018229058407815424, + "loss": 0.0002, + "step": 446 + }, + { + "epoch": 0.09252742703374042, + "grad_norm": 0.009306194260716438, + "learning_rate": 0.00018224901267927668, + "loss": 0.0003, + "step": 447 + }, + { + "epoch": 0.09273442351480025, + "grad_norm": 0.0030481938738375902, + "learning_rate": 0.00018220744128039912, + "loss": 0.0013, + "step": 448 + }, + { + "epoch": 0.09294141999586007, + "grad_norm": 0.013268841430544853, + "learning_rate": 0.0001821658698815215, + "loss": 0.0017, + "step": 449 + }, + { + "epoch": 0.09314841647691989, + "grad_norm": 0.006063752807676792, + "learning_rate": 0.00018212429848264394, + "loss": 0.0002, + "step": 450 + }, + { + "epoch": 0.09335541295797971, + "grad_norm": 0.00182344822678715, + "learning_rate": 0.00018208272708376637, + "loss": 0.0001, + "step": 451 + }, + { + "epoch": 0.09356240943903954, + "grad_norm": 0.009372780099511147, + "learning_rate": 0.0001820411556848888, + "loss": 0.0005, + "step": 452 + }, + { + "epoch": 0.09376940592009936, + "grad_norm": 0.003136920742690563, + "learning_rate": 0.00018199958428601122, + "loss": 0.0001, + "step": 453 + }, + { + "epoch": 0.09397640240115918, + "grad_norm": 0.030062230303883553, + "learning_rate": 0.00018195801288713366, + "loss": 0.0006, + "step": 454 + }, + { + "epoch": 0.09418339888221901, + "grad_norm": 0.004309754353016615, + "learning_rate": 0.0001819164414882561, + "loss": 0.0007, + "step": 455 + }, + { + "epoch": 0.09439039536327883, + "grad_norm": 0.010606180876493454, + "learning_rate": 0.0001818748700893785, + "loss": 0.0002, + "step": 456 + }, + { + "epoch": 0.09459739184433864, + "grad_norm": 0.0054748812690377235, + "learning_rate": 0.00018183329869050094, + "loss": 0.0005, + "step": 457 + }, + { + "epoch": 0.09480438832539846, + "grad_norm": 0.001673020888119936, + "learning_rate": 0.00018179172729162338, + "loss": 0.0001, + "step": 458 + }, + { + "epoch": 0.0950113848064583, + "grad_norm": 0.004401384387165308, + "learning_rate": 0.0001817501558927458, + "loss": 0.0001, + "step": 459 + }, + { + "epoch": 0.09521838128751811, + "grad_norm": 0.005755012389272451, + "learning_rate": 0.00018170858449386822, + "loss": 0.0012, + "step": 460 + }, + { + "epoch": 0.09542537776857793, + "grad_norm": 0.004951901733875275, + "learning_rate": 0.00018166701309499066, + "loss": 0.0012, + "step": 461 + }, + { + "epoch": 0.09563237424963776, + "grad_norm": 0.0014370133867487311, + "learning_rate": 0.0001816254416961131, + "loss": 0.0001, + "step": 462 + }, + { + "epoch": 0.09583937073069758, + "grad_norm": 0.012078122235834599, + "learning_rate": 0.0001815838702972355, + "loss": 0.0013, + "step": 463 + }, + { + "epoch": 0.0960463672117574, + "grad_norm": 0.002765175886452198, + "learning_rate": 0.00018154229889835795, + "loss": 0.0009, + "step": 464 + }, + { + "epoch": 0.09625336369281723, + "grad_norm": 0.0017701378092169762, + "learning_rate": 0.00018150072749948038, + "loss": 0.0001, + "step": 465 + }, + { + "epoch": 0.09646036017387705, + "grad_norm": 0.003232579445466399, + "learning_rate": 0.0001814591561006028, + "loss": 0.0001, + "step": 466 + }, + { + "epoch": 0.09666735665493686, + "grad_norm": 0.004619232844561338, + "learning_rate": 0.00018141758470172523, + "loss": 0.0012, + "step": 467 + }, + { + "epoch": 0.09687435313599668, + "grad_norm": 0.008126890286803246, + "learning_rate": 0.00018137601330284767, + "loss": 0.0011, + "step": 468 + }, + { + "epoch": 0.09708134961705651, + "grad_norm": 0.004719397984445095, + "learning_rate": 0.00018133444190397008, + "loss": 0.0009, + "step": 469 + }, + { + "epoch": 0.09728834609811633, + "grad_norm": 0.005153202451765537, + "learning_rate": 0.00018129287050509251, + "loss": 0.0012, + "step": 470 + }, + { + "epoch": 0.09749534257917615, + "grad_norm": 0.01385215763002634, + "learning_rate": 0.00018125129910621492, + "loss": 0.0022, + "step": 471 + }, + { + "epoch": 0.09770233906023598, + "grad_norm": 0.004983994178473949, + "learning_rate": 0.00018120972770733736, + "loss": 0.0004, + "step": 472 + }, + { + "epoch": 0.0979093355412958, + "grad_norm": 0.007088206708431244, + "learning_rate": 0.00018116815630845977, + "loss": 0.0005, + "step": 473 + }, + { + "epoch": 0.09811633202235562, + "grad_norm": 0.004754175432026386, + "learning_rate": 0.0001811265849095822, + "loss": 0.0006, + "step": 474 + }, + { + "epoch": 0.09832332850341544, + "grad_norm": 0.004105637315660715, + "learning_rate": 0.00018108501351070465, + "loss": 0.0001, + "step": 475 + }, + { + "epoch": 0.09853032498447527, + "grad_norm": 0.002687152475118637, + "learning_rate": 0.00018104344211182706, + "loss": 0.0001, + "step": 476 + }, + { + "epoch": 0.09873732146553509, + "grad_norm": 0.0023124783765524626, + "learning_rate": 0.0001810018707129495, + "loss": 0.0001, + "step": 477 + }, + { + "epoch": 0.0989443179465949, + "grad_norm": 0.007885076105594635, + "learning_rate": 0.00018096029931407193, + "loss": 0.0006, + "step": 478 + }, + { + "epoch": 0.09915131442765474, + "grad_norm": 0.0014087413437664509, + "learning_rate": 0.00018091872791519434, + "loss": 0.0007, + "step": 479 + }, + { + "epoch": 0.09935831090871455, + "grad_norm": 0.0055119190365076065, + "learning_rate": 0.00018087715651631678, + "loss": 0.0002, + "step": 480 + }, + { + "epoch": 0.09956530738977437, + "grad_norm": 0.0003096537839155644, + "learning_rate": 0.0001808355851174392, + "loss": 0.0, + "step": 481 + }, + { + "epoch": 0.09977230387083419, + "grad_norm": 0.001899409806355834, + "learning_rate": 0.00018079401371856165, + "loss": 0.0002, + "step": 482 + }, + { + "epoch": 0.09997930035189402, + "grad_norm": 0.005635194014757872, + "learning_rate": 0.00018075244231968406, + "loss": 0.0001, + "step": 483 + }, + { + "epoch": 0.10018629683295384, + "grad_norm": 0.005856087896972895, + "learning_rate": 0.0001807108709208065, + "loss": 0.001, + "step": 484 + }, + { + "epoch": 0.10039329331401366, + "grad_norm": 0.005273948423564434, + "learning_rate": 0.00018066929952192893, + "loss": 0.0013, + "step": 485 + }, + { + "epoch": 0.10060028979507349, + "grad_norm": 0.001652201754041016, + "learning_rate": 0.00018062772812305134, + "loss": 0.0001, + "step": 486 + }, + { + "epoch": 0.1008072862761333, + "grad_norm": 0.006849886849522591, + "learning_rate": 0.00018058615672417378, + "loss": 0.0001, + "step": 487 + }, + { + "epoch": 0.10101428275719312, + "grad_norm": 0.009754250757396221, + "learning_rate": 0.00018054458532529622, + "loss": 0.0023, + "step": 488 + }, + { + "epoch": 0.10122127923825296, + "grad_norm": 0.0038455536123365164, + "learning_rate": 0.00018050301392641863, + "loss": 0.0016, + "step": 489 + }, + { + "epoch": 0.10142827571931277, + "grad_norm": 0.0006483698962256312, + "learning_rate": 0.00018046144252754107, + "loss": 0.0, + "step": 490 + }, + { + "epoch": 0.10163527220037259, + "grad_norm": 0.008700639940798283, + "learning_rate": 0.0001804198711286635, + "loss": 0.0007, + "step": 491 + }, + { + "epoch": 0.10184226868143241, + "grad_norm": 0.004151639994233847, + "learning_rate": 0.00018037829972978594, + "loss": 0.0006, + "step": 492 + }, + { + "epoch": 0.10204926516249224, + "grad_norm": 0.003242357401177287, + "learning_rate": 0.00018033672833090832, + "loss": 0.0001, + "step": 493 + }, + { + "epoch": 0.10225626164355206, + "grad_norm": 0.0038309101946651936, + "learning_rate": 0.00018029515693203076, + "loss": 0.0002, + "step": 494 + }, + { + "epoch": 0.10246325812461188, + "grad_norm": 0.0032492976170033216, + "learning_rate": 0.0001802535855331532, + "loss": 0.0001, + "step": 495 + }, + { + "epoch": 0.10267025460567171, + "grad_norm": 0.005621058400720358, + "learning_rate": 0.0001802120141342756, + "loss": 0.0002, + "step": 496 + }, + { + "epoch": 0.10287725108673153, + "grad_norm": 0.007613383699208498, + "learning_rate": 0.00018017044273539804, + "loss": 0.0013, + "step": 497 + }, + { + "epoch": 0.10308424756779135, + "grad_norm": 0.004469983279705048, + "learning_rate": 0.00018012887133652048, + "loss": 0.0001, + "step": 498 + }, + { + "epoch": 0.10329124404885116, + "grad_norm": 0.010518516413867474, + "learning_rate": 0.00018008729993764292, + "loss": 0.0002, + "step": 499 + }, + { + "epoch": 0.103498240529911, + "grad_norm": 0.0019439860479906201, + "learning_rate": 0.00018004572853876533, + "loss": 0.0001, + "step": 500 + }, + { + "epoch": 0.10370523701097081, + "grad_norm": 0.007315011695027351, + "learning_rate": 0.00018000415713988776, + "loss": 0.0002, + "step": 501 + }, + { + "epoch": 0.10391223349203063, + "grad_norm": 0.0027510204818099737, + "learning_rate": 0.0001799625857410102, + "loss": 0.0002, + "step": 502 + }, + { + "epoch": 0.10411922997309046, + "grad_norm": 0.007963057607412338, + "learning_rate": 0.0001799210143421326, + "loss": 0.0003, + "step": 503 + }, + { + "epoch": 0.10432622645415028, + "grad_norm": 0.004816776607185602, + "learning_rate": 0.00017987944294325505, + "loss": 0.001, + "step": 504 + }, + { + "epoch": 0.1045332229352101, + "grad_norm": 0.0010107432026416063, + "learning_rate": 0.00017983787154437749, + "loss": 0.0, + "step": 505 + }, + { + "epoch": 0.10474021941626992, + "grad_norm": 0.002849761163815856, + "learning_rate": 0.0001797963001454999, + "loss": 0.0001, + "step": 506 + }, + { + "epoch": 0.10494721589732975, + "grad_norm": 0.008465790189802647, + "learning_rate": 0.00017975472874662233, + "loss": 0.0017, + "step": 507 + }, + { + "epoch": 0.10515421237838957, + "grad_norm": 0.001012888620607555, + "learning_rate": 0.00017971315734774477, + "loss": 0.0, + "step": 508 + }, + { + "epoch": 0.10536120885944938, + "grad_norm": 0.005154603160917759, + "learning_rate": 0.0001796715859488672, + "loss": 0.0001, + "step": 509 + }, + { + "epoch": 0.10556820534050922, + "grad_norm": 0.012283824384212494, + "learning_rate": 0.00017963001454998962, + "loss": 0.0004, + "step": 510 + }, + { + "epoch": 0.10577520182156903, + "grad_norm": 0.001696570310741663, + "learning_rate": 0.00017958844315111205, + "loss": 0.0001, + "step": 511 + }, + { + "epoch": 0.10598219830262885, + "grad_norm": 0.0002515624219086021, + "learning_rate": 0.0001795468717522345, + "loss": 0.0, + "step": 512 + }, + { + "epoch": 0.10618919478368868, + "grad_norm": 0.0006335057551041245, + "learning_rate": 0.0001795053003533569, + "loss": 0.0, + "step": 513 + }, + { + "epoch": 0.1063961912647485, + "grad_norm": 0.0007280270801857114, + "learning_rate": 0.00017946372895447934, + "loss": 0.0, + "step": 514 + }, + { + "epoch": 0.10660318774580832, + "grad_norm": 0.0040188622660934925, + "learning_rate": 0.00017942215755560175, + "loss": 0.0002, + "step": 515 + }, + { + "epoch": 0.10681018422686814, + "grad_norm": 0.0054796175099909306, + "learning_rate": 0.00017938058615672419, + "loss": 0.0009, + "step": 516 + }, + { + "epoch": 0.10701718070792797, + "grad_norm": 0.0034792469814419746, + "learning_rate": 0.0001793390147578466, + "loss": 0.0001, + "step": 517 + }, + { + "epoch": 0.10722417718898779, + "grad_norm": 0.00432013813406229, + "learning_rate": 0.00017929744335896903, + "loss": 0.0002, + "step": 518 + }, + { + "epoch": 0.1074311736700476, + "grad_norm": 0.004862105939537287, + "learning_rate": 0.00017925587196009147, + "loss": 0.001, + "step": 519 + }, + { + "epoch": 0.10763817015110744, + "grad_norm": 0.002249139128252864, + "learning_rate": 0.00017921430056121388, + "loss": 0.0, + "step": 520 + }, + { + "epoch": 0.10784516663216726, + "grad_norm": 0.006818681955337524, + "learning_rate": 0.00017917272916233632, + "loss": 0.0004, + "step": 521 + }, + { + "epoch": 0.10805216311322707, + "grad_norm": 0.008624670095741749, + "learning_rate": 0.00017913115776345875, + "loss": 0.0004, + "step": 522 + }, + { + "epoch": 0.10825915959428689, + "grad_norm": 0.0007210278417915106, + "learning_rate": 0.00017908958636458116, + "loss": 0.0, + "step": 523 + }, + { + "epoch": 0.10846615607534672, + "grad_norm": 0.010820691473782063, + "learning_rate": 0.0001790480149657036, + "loss": 0.0003, + "step": 524 + }, + { + "epoch": 0.10867315255640654, + "grad_norm": 0.0023018312640488148, + "learning_rate": 0.00017900644356682604, + "loss": 0.0001, + "step": 525 + }, + { + "epoch": 0.10888014903746636, + "grad_norm": 0.00019024198991246521, + "learning_rate": 0.00017896487216794845, + "loss": 0.0, + "step": 526 + }, + { + "epoch": 0.10908714551852619, + "grad_norm": 0.005011410918086767, + "learning_rate": 0.00017892330076907088, + "loss": 0.0003, + "step": 527 + }, + { + "epoch": 0.10929414199958601, + "grad_norm": 0.007016469724476337, + "learning_rate": 0.00017888172937019332, + "loss": 0.0009, + "step": 528 + }, + { + "epoch": 0.10950113848064583, + "grad_norm": 0.008118787780404091, + "learning_rate": 0.00017884015797131576, + "loss": 0.0003, + "step": 529 + }, + { + "epoch": 0.10970813496170564, + "grad_norm": 0.005854643415659666, + "learning_rate": 0.00017879858657243817, + "loss": 0.0011, + "step": 530 + }, + { + "epoch": 0.10991513144276548, + "grad_norm": 0.0038967933505773544, + "learning_rate": 0.0001787570151735606, + "loss": 0.001, + "step": 531 + }, + { + "epoch": 0.1101221279238253, + "grad_norm": 0.003802061313763261, + "learning_rate": 0.00017871544377468304, + "loss": 0.0018, + "step": 532 + }, + { + "epoch": 0.11032912440488511, + "grad_norm": 0.004740913398563862, + "learning_rate": 0.00017867387237580545, + "loss": 0.0001, + "step": 533 + }, + { + "epoch": 0.11053612088594494, + "grad_norm": 0.008046228438615799, + "learning_rate": 0.0001786323009769279, + "loss": 0.0005, + "step": 534 + }, + { + "epoch": 0.11074311736700476, + "grad_norm": 0.0014560514828190207, + "learning_rate": 0.00017859072957805033, + "loss": 0.0, + "step": 535 + }, + { + "epoch": 0.11095011384806458, + "grad_norm": 0.009222283028066158, + "learning_rate": 0.00017854915817917274, + "loss": 0.0002, + "step": 536 + }, + { + "epoch": 0.11115711032912441, + "grad_norm": 0.005452610552310944, + "learning_rate": 0.00017850758678029515, + "loss": 0.0013, + "step": 537 + }, + { + "epoch": 0.11136410681018423, + "grad_norm": 0.0007964425021782517, + "learning_rate": 0.00017846601538141758, + "loss": 0.0, + "step": 538 + }, + { + "epoch": 0.11157110329124405, + "grad_norm": 0.007167865987867117, + "learning_rate": 0.00017842444398254002, + "loss": 0.0011, + "step": 539 + }, + { + "epoch": 0.11177809977230387, + "grad_norm": 0.01039041206240654, + "learning_rate": 0.00017838287258366243, + "loss": 0.0007, + "step": 540 + }, + { + "epoch": 0.1119850962533637, + "grad_norm": 0.0001594142959220335, + "learning_rate": 0.00017834130118478487, + "loss": 0.0, + "step": 541 + }, + { + "epoch": 0.11219209273442352, + "grad_norm": 0.0023100003600120544, + "learning_rate": 0.0001782997297859073, + "loss": 0.0007, + "step": 542 + }, + { + "epoch": 0.11239908921548333, + "grad_norm": 0.003656044602394104, + "learning_rate": 0.00017825815838702972, + "loss": 0.0014, + "step": 543 + }, + { + "epoch": 0.11260608569654317, + "grad_norm": 0.0014201959129422903, + "learning_rate": 0.00017821658698815215, + "loss": 0.0001, + "step": 544 + }, + { + "epoch": 0.11281308217760298, + "grad_norm": 0.015617001801729202, + "learning_rate": 0.0001781750155892746, + "loss": 0.0002, + "step": 545 + }, + { + "epoch": 0.1130200786586628, + "grad_norm": 0.008762934245169163, + "learning_rate": 0.00017813344419039703, + "loss": 0.0003, + "step": 546 + }, + { + "epoch": 0.11322707513972262, + "grad_norm": 0.002274678787216544, + "learning_rate": 0.00017809187279151944, + "loss": 0.0001, + "step": 547 + }, + { + "epoch": 0.11343407162078245, + "grad_norm": 0.003864066442474723, + "learning_rate": 0.00017805030139264187, + "loss": 0.0005, + "step": 548 + }, + { + "epoch": 0.11364106810184227, + "grad_norm": 0.006032771430909634, + "learning_rate": 0.0001780087299937643, + "loss": 0.0003, + "step": 549 + }, + { + "epoch": 0.11384806458290209, + "grad_norm": 0.0011935516959056258, + "learning_rate": 0.00017796715859488672, + "loss": 0.0, + "step": 550 + }, + { + "epoch": 0.11405506106396192, + "grad_norm": 0.013128140941262245, + "learning_rate": 0.00017792558719600916, + "loss": 0.0006, + "step": 551 + }, + { + "epoch": 0.11426205754502174, + "grad_norm": 0.000645163469016552, + "learning_rate": 0.0001778840157971316, + "loss": 0.0, + "step": 552 + }, + { + "epoch": 0.11446905402608155, + "grad_norm": 0.011457535438239574, + "learning_rate": 0.000177842444398254, + "loss": 0.0003, + "step": 553 + }, + { + "epoch": 0.11467605050714137, + "grad_norm": 0.003057427005842328, + "learning_rate": 0.00017780087299937644, + "loss": 0.0012, + "step": 554 + }, + { + "epoch": 0.1148830469882012, + "grad_norm": 0.0017485780408605933, + "learning_rate": 0.00017775930160049888, + "loss": 0.0006, + "step": 555 + }, + { + "epoch": 0.11509004346926102, + "grad_norm": 0.0005046813748776913, + "learning_rate": 0.0001777177302016213, + "loss": 0.0, + "step": 556 + }, + { + "epoch": 0.11529703995032084, + "grad_norm": 0.007802332751452923, + "learning_rate": 0.00017767615880274373, + "loss": 0.0009, + "step": 557 + }, + { + "epoch": 0.11550403643138067, + "grad_norm": 0.0074394443072378635, + "learning_rate": 0.00017763458740386616, + "loss": 0.0021, + "step": 558 + }, + { + "epoch": 0.11571103291244049, + "grad_norm": 0.007590603083372116, + "learning_rate": 0.00017759301600498857, + "loss": 0.0004, + "step": 559 + }, + { + "epoch": 0.11591802939350031, + "grad_norm": 0.005805825348943472, + "learning_rate": 0.00017755144460611098, + "loss": 0.0019, + "step": 560 + }, + { + "epoch": 0.11612502587456014, + "grad_norm": 0.008781611919403076, + "learning_rate": 0.00017750987320723342, + "loss": 0.0017, + "step": 561 + }, + { + "epoch": 0.11633202235561996, + "grad_norm": 0.0010139975929632783, + "learning_rate": 0.00017746830180835586, + "loss": 0.0, + "step": 562 + }, + { + "epoch": 0.11653901883667978, + "grad_norm": 0.003940982278436422, + "learning_rate": 0.0001774267304094783, + "loss": 0.0001, + "step": 563 + }, + { + "epoch": 0.1167460153177396, + "grad_norm": 0.013145407661795616, + "learning_rate": 0.0001773851590106007, + "loss": 0.0012, + "step": 564 + }, + { + "epoch": 0.11695301179879942, + "grad_norm": 0.0011159079149365425, + "learning_rate": 0.00017734358761172314, + "loss": 0.0001, + "step": 565 + }, + { + "epoch": 0.11716000827985924, + "grad_norm": 0.01270979829132557, + "learning_rate": 0.00017730201621284558, + "loss": 0.0017, + "step": 566 + }, + { + "epoch": 0.11736700476091906, + "grad_norm": 0.004431411158293486, + "learning_rate": 0.000177260444813968, + "loss": 0.0001, + "step": 567 + }, + { + "epoch": 0.11757400124197889, + "grad_norm": 0.0041249035857617855, + "learning_rate": 0.00017721887341509042, + "loss": 0.0003, + "step": 568 + }, + { + "epoch": 0.11778099772303871, + "grad_norm": 0.0006257002823986113, + "learning_rate": 0.00017717730201621286, + "loss": 0.0, + "step": 569 + }, + { + "epoch": 0.11798799420409853, + "grad_norm": 0.004501596093177795, + "learning_rate": 0.00017713573061733527, + "loss": 0.0001, + "step": 570 + }, + { + "epoch": 0.11819499068515835, + "grad_norm": 0.00742512010037899, + "learning_rate": 0.0001770941592184577, + "loss": 0.0002, + "step": 571 + }, + { + "epoch": 0.11840198716621818, + "grad_norm": 0.010421551764011383, + "learning_rate": 0.00017705258781958015, + "loss": 0.0008, + "step": 572 + }, + { + "epoch": 0.118608983647278, + "grad_norm": 0.0010451058624312282, + "learning_rate": 0.00017701101642070256, + "loss": 0.0001, + "step": 573 + }, + { + "epoch": 0.11881598012833781, + "grad_norm": 0.0014272347325459123, + "learning_rate": 0.000176969445021825, + "loss": 0.0001, + "step": 574 + }, + { + "epoch": 0.11902297660939765, + "grad_norm": 0.007021667901426554, + "learning_rate": 0.00017692787362294743, + "loss": 0.0002, + "step": 575 + }, + { + "epoch": 0.11922997309045746, + "grad_norm": 0.0050498368218541145, + "learning_rate": 0.00017688630222406987, + "loss": 0.001, + "step": 576 + }, + { + "epoch": 0.11943696957151728, + "grad_norm": 0.0004365240456536412, + "learning_rate": 0.00017684473082519228, + "loss": 0.0, + "step": 577 + }, + { + "epoch": 0.1196439660525771, + "grad_norm": 0.007502545602619648, + "learning_rate": 0.00017680315942631471, + "loss": 0.0002, + "step": 578 + }, + { + "epoch": 0.11985096253363693, + "grad_norm": 0.00824455451220274, + "learning_rate": 0.00017676158802743715, + "loss": 0.0004, + "step": 579 + }, + { + "epoch": 0.12005795901469675, + "grad_norm": 0.003414528677240014, + "learning_rate": 0.00017672001662855956, + "loss": 0.0011, + "step": 580 + }, + { + "epoch": 0.12026495549575657, + "grad_norm": 0.004608092829585075, + "learning_rate": 0.00017667844522968197, + "loss": 0.0014, + "step": 581 + }, + { + "epoch": 0.1204719519768164, + "grad_norm": 0.006573988124728203, + "learning_rate": 0.0001766368738308044, + "loss": 0.0002, + "step": 582 + }, + { + "epoch": 0.12067894845787622, + "grad_norm": 0.006878604646772146, + "learning_rate": 0.00017659530243192685, + "loss": 0.0005, + "step": 583 + }, + { + "epoch": 0.12088594493893604, + "grad_norm": 0.0013765916228294373, + "learning_rate": 0.00017655373103304926, + "loss": 0.0001, + "step": 584 + }, + { + "epoch": 0.12109294141999585, + "grad_norm": 0.009517376311123371, + "learning_rate": 0.0001765121596341717, + "loss": 0.0005, + "step": 585 + }, + { + "epoch": 0.12129993790105568, + "grad_norm": 0.02729586698114872, + "learning_rate": 0.00017647058823529413, + "loss": 0.0013, + "step": 586 + }, + { + "epoch": 0.1215069343821155, + "grad_norm": 0.005033944733440876, + "learning_rate": 0.00017642901683641654, + "loss": 0.0011, + "step": 587 + }, + { + "epoch": 0.12171393086317532, + "grad_norm": 0.001488934038206935, + "learning_rate": 0.00017638744543753898, + "loss": 0.0001, + "step": 588 + }, + { + "epoch": 0.12192092734423515, + "grad_norm": 0.004233523737639189, + "learning_rate": 0.0001763458740386614, + "loss": 0.0013, + "step": 589 + }, + { + "epoch": 0.12212792382529497, + "grad_norm": 0.001819688593968749, + "learning_rate": 0.00017630430263978382, + "loss": 0.0002, + "step": 590 + }, + { + "epoch": 0.12233492030635479, + "grad_norm": 0.0051133958622813225, + "learning_rate": 0.00017626273124090626, + "loss": 0.0003, + "step": 591 + }, + { + "epoch": 0.12254191678741462, + "grad_norm": 0.007632863707840443, + "learning_rate": 0.0001762211598420287, + "loss": 0.0005, + "step": 592 + }, + { + "epoch": 0.12274891326847444, + "grad_norm": 0.0009289845474995673, + "learning_rate": 0.00017617958844315113, + "loss": 0.0, + "step": 593 + }, + { + "epoch": 0.12295590974953426, + "grad_norm": 0.00543027650564909, + "learning_rate": 0.00017613801704427354, + "loss": 0.0001, + "step": 594 + }, + { + "epoch": 0.12316290623059407, + "grad_norm": 0.002607417991384864, + "learning_rate": 0.00017609644564539598, + "loss": 0.0002, + "step": 595 + }, + { + "epoch": 0.1233699027116539, + "grad_norm": 0.025557972490787506, + "learning_rate": 0.00017605487424651842, + "loss": 0.0002, + "step": 596 + }, + { + "epoch": 0.12357689919271372, + "grad_norm": 0.0016189507441595197, + "learning_rate": 0.00017601330284764083, + "loss": 0.0001, + "step": 597 + }, + { + "epoch": 0.12378389567377354, + "grad_norm": 0.004612909164279699, + "learning_rate": 0.00017597173144876327, + "loss": 0.0002, + "step": 598 + }, + { + "epoch": 0.12399089215483337, + "grad_norm": 0.00019464526849333197, + "learning_rate": 0.0001759301600498857, + "loss": 0.0, + "step": 599 + }, + { + "epoch": 0.12419788863589319, + "grad_norm": 0.0013309603091329336, + "learning_rate": 0.0001758885886510081, + "loss": 0.0007, + "step": 600 + }, + { + "epoch": 0.12440488511695301, + "grad_norm": 0.002917417325079441, + "learning_rate": 0.00017584701725213055, + "loss": 0.0001, + "step": 601 + }, + { + "epoch": 0.12461188159801283, + "grad_norm": 0.004730269778519869, + "learning_rate": 0.000175805445853253, + "loss": 0.0001, + "step": 602 + }, + { + "epoch": 0.12481887807907266, + "grad_norm": 0.0036635478027164936, + "learning_rate": 0.0001757638744543754, + "loss": 0.0001, + "step": 603 + }, + { + "epoch": 0.12502587456013248, + "grad_norm": 0.002084661042317748, + "learning_rate": 0.0001757223030554978, + "loss": 0.0001, + "step": 604 + }, + { + "epoch": 0.1252328710411923, + "grad_norm": 0.006881284527480602, + "learning_rate": 0.00017568073165662024, + "loss": 0.0002, + "step": 605 + }, + { + "epoch": 0.1254398675222521, + "grad_norm": 0.0007496286416426301, + "learning_rate": 0.00017563916025774268, + "loss": 0.0, + "step": 606 + }, + { + "epoch": 0.12564686400331193, + "grad_norm": 0.0013991744490340352, + "learning_rate": 0.0001755975888588651, + "loss": 0.0, + "step": 607 + }, + { + "epoch": 0.12585386048437178, + "grad_norm": 0.00578208127990365, + "learning_rate": 0.00017555601745998753, + "loss": 0.0016, + "step": 608 + }, + { + "epoch": 0.1260608569654316, + "grad_norm": 0.0005476415390148759, + "learning_rate": 0.00017551444606110996, + "loss": 0.0, + "step": 609 + }, + { + "epoch": 0.1262678534464914, + "grad_norm": 0.003824407234787941, + "learning_rate": 0.0001754728746622324, + "loss": 0.0001, + "step": 610 + }, + { + "epoch": 0.12647484992755123, + "grad_norm": 0.0068860347382724285, + "learning_rate": 0.0001754313032633548, + "loss": 0.0003, + "step": 611 + }, + { + "epoch": 0.12668184640861105, + "grad_norm": 0.001763600972481072, + "learning_rate": 0.00017538973186447725, + "loss": 0.0002, + "step": 612 + }, + { + "epoch": 0.12688884288967087, + "grad_norm": 0.0029042328242212534, + "learning_rate": 0.00017534816046559969, + "loss": 0.0019, + "step": 613 + }, + { + "epoch": 0.12709583937073068, + "grad_norm": 0.026835285127162933, + "learning_rate": 0.0001753065890667221, + "loss": 0.0006, + "step": 614 + }, + { + "epoch": 0.12730283585179053, + "grad_norm": 0.0025784820318222046, + "learning_rate": 0.00017526501766784453, + "loss": 0.0002, + "step": 615 + }, + { + "epoch": 0.12750983233285035, + "grad_norm": 0.000811999780125916, + "learning_rate": 0.00017522344626896697, + "loss": 0.0, + "step": 616 + }, + { + "epoch": 0.12771682881391017, + "grad_norm": 0.0023158304393291473, + "learning_rate": 0.00017518187487008938, + "loss": 0.0001, + "step": 617 + }, + { + "epoch": 0.12792382529496998, + "grad_norm": 0.00527742225676775, + "learning_rate": 0.00017514030347121182, + "loss": 0.0011, + "step": 618 + }, + { + "epoch": 0.1281308217760298, + "grad_norm": 0.004715193063020706, + "learning_rate": 0.00017509873207233425, + "loss": 0.0026, + "step": 619 + }, + { + "epoch": 0.12833781825708962, + "grad_norm": 0.001638007932342589, + "learning_rate": 0.00017505716067345666, + "loss": 0.0001, + "step": 620 + }, + { + "epoch": 0.12854481473814944, + "grad_norm": 0.0012813376961275935, + "learning_rate": 0.0001750155892745791, + "loss": 0.0001, + "step": 621 + }, + { + "epoch": 0.12875181121920928, + "grad_norm": 0.006484480109065771, + "learning_rate": 0.00017497401787570154, + "loss": 0.0004, + "step": 622 + }, + { + "epoch": 0.1289588077002691, + "grad_norm": 0.00035095165367238224, + "learning_rate": 0.00017493244647682398, + "loss": 0.0, + "step": 623 + }, + { + "epoch": 0.12916580418132892, + "grad_norm": 0.004927014000713825, + "learning_rate": 0.00017489087507794639, + "loss": 0.0012, + "step": 624 + }, + { + "epoch": 0.12937280066238874, + "grad_norm": 0.00287305167876184, + "learning_rate": 0.0001748493036790688, + "loss": 0.0001, + "step": 625 + }, + { + "epoch": 0.12957979714344856, + "grad_norm": 0.003079169662669301, + "learning_rate": 0.00017480773228019123, + "loss": 0.0001, + "step": 626 + }, + { + "epoch": 0.12978679362450837, + "grad_norm": 0.0018820518162101507, + "learning_rate": 0.00017476616088131367, + "loss": 0.0001, + "step": 627 + }, + { + "epoch": 0.12999379010556822, + "grad_norm": 0.004426770843565464, + "learning_rate": 0.00017472458948243608, + "loss": 0.0001, + "step": 628 + }, + { + "epoch": 0.13020078658662804, + "grad_norm": 0.008074757643043995, + "learning_rate": 0.00017468301808355852, + "loss": 0.0002, + "step": 629 + }, + { + "epoch": 0.13040778306768785, + "grad_norm": 0.004479815252125263, + "learning_rate": 0.00017464144668468095, + "loss": 0.0002, + "step": 630 + }, + { + "epoch": 0.13061477954874767, + "grad_norm": 0.0016544251702725887, + "learning_rate": 0.00017459987528580336, + "loss": 0.0, + "step": 631 + }, + { + "epoch": 0.1308217760298075, + "grad_norm": 0.0007902836659923196, + "learning_rate": 0.0001745583038869258, + "loss": 0.0, + "step": 632 + }, + { + "epoch": 0.1310287725108673, + "grad_norm": 0.011000900529325008, + "learning_rate": 0.00017451673248804824, + "loss": 0.0005, + "step": 633 + }, + { + "epoch": 0.13123576899192713, + "grad_norm": 0.00046783004654571414, + "learning_rate": 0.00017447516108917065, + "loss": 0.0, + "step": 634 + }, + { + "epoch": 0.13144276547298697, + "grad_norm": 0.003358067711815238, + "learning_rate": 0.00017443358969029308, + "loss": 0.0018, + "step": 635 + }, + { + "epoch": 0.1316497619540468, + "grad_norm": 0.0025496368762105703, + "learning_rate": 0.00017439201829141552, + "loss": 0.0015, + "step": 636 + }, + { + "epoch": 0.1318567584351066, + "grad_norm": 0.0016015061410143971, + "learning_rate": 0.00017435044689253793, + "loss": 0.0002, + "step": 637 + }, + { + "epoch": 0.13206375491616643, + "grad_norm": 0.0038993649650365114, + "learning_rate": 0.00017430887549366037, + "loss": 0.0001, + "step": 638 + }, + { + "epoch": 0.13227075139722624, + "grad_norm": 0.0033800469245761633, + "learning_rate": 0.0001742673040947828, + "loss": 0.0001, + "step": 639 + }, + { + "epoch": 0.13247774787828606, + "grad_norm": 0.0008187236380763352, + "learning_rate": 0.00017422573269590524, + "loss": 0.0, + "step": 640 + }, + { + "epoch": 0.13268474435934588, + "grad_norm": 0.005097914487123489, + "learning_rate": 0.00017418416129702765, + "loss": 0.0004, + "step": 641 + }, + { + "epoch": 0.13289174084040573, + "grad_norm": 0.0009978336747735739, + "learning_rate": 0.0001741425898981501, + "loss": 0.0, + "step": 642 + }, + { + "epoch": 0.13309873732146554, + "grad_norm": 0.004832749720662832, + "learning_rate": 0.00017410101849927253, + "loss": 0.0012, + "step": 643 + }, + { + "epoch": 0.13330573380252536, + "grad_norm": 0.0038694944232702255, + "learning_rate": 0.00017405944710039494, + "loss": 0.0001, + "step": 644 + }, + { + "epoch": 0.13351273028358518, + "grad_norm": 0.001419690903276205, + "learning_rate": 0.00017401787570151737, + "loss": 0.0, + "step": 645 + }, + { + "epoch": 0.133719726764645, + "grad_norm": 0.006202602293342352, + "learning_rate": 0.0001739763043026398, + "loss": 0.0002, + "step": 646 + }, + { + "epoch": 0.13392672324570482, + "grad_norm": 0.0008485604776069522, + "learning_rate": 0.00017393473290376222, + "loss": 0.0, + "step": 647 + }, + { + "epoch": 0.13413371972676463, + "grad_norm": 0.0050230189226567745, + "learning_rate": 0.00017389316150488463, + "loss": 0.0022, + "step": 648 + }, + { + "epoch": 0.13434071620782448, + "grad_norm": 0.002081549260765314, + "learning_rate": 0.00017385159010600707, + "loss": 0.0001, + "step": 649 + }, + { + "epoch": 0.1345477126888843, + "grad_norm": 0.001964141381904483, + "learning_rate": 0.0001738100187071295, + "loss": 0.0002, + "step": 650 + }, + { + "epoch": 0.13475470916994411, + "grad_norm": 0.0006888345233164728, + "learning_rate": 0.00017376844730825192, + "loss": 0.0, + "step": 651 + }, + { + "epoch": 0.13496170565100393, + "grad_norm": 0.002313450677320361, + "learning_rate": 0.00017372687590937435, + "loss": 0.0, + "step": 652 + }, + { + "epoch": 0.13516870213206375, + "grad_norm": 0.007078672293573618, + "learning_rate": 0.0001736853045104968, + "loss": 0.0006, + "step": 653 + }, + { + "epoch": 0.13537569861312357, + "grad_norm": 0.005166813265532255, + "learning_rate": 0.0001736437331116192, + "loss": 0.0011, + "step": 654 + }, + { + "epoch": 0.1355826950941834, + "grad_norm": 0.007185124326497316, + "learning_rate": 0.00017360216171274164, + "loss": 0.0001, + "step": 655 + }, + { + "epoch": 0.13578969157524323, + "grad_norm": 0.005528238136321306, + "learning_rate": 0.00017356059031386407, + "loss": 0.0001, + "step": 656 + }, + { + "epoch": 0.13599668805630305, + "grad_norm": 0.0077844299376010895, + "learning_rate": 0.0001735190189149865, + "loss": 0.0011, + "step": 657 + }, + { + "epoch": 0.13620368453736287, + "grad_norm": 0.00246329209767282, + "learning_rate": 0.00017347744751610892, + "loss": 0.0008, + "step": 658 + }, + { + "epoch": 0.13641068101842269, + "grad_norm": 0.005287639796733856, + "learning_rate": 0.00017343587611723136, + "loss": 0.0002, + "step": 659 + }, + { + "epoch": 0.1366176774994825, + "grad_norm": 0.006681959610432386, + "learning_rate": 0.0001733943047183538, + "loss": 0.0013, + "step": 660 + }, + { + "epoch": 0.13682467398054232, + "grad_norm": 0.0063599334098398685, + "learning_rate": 0.0001733527333194762, + "loss": 0.0003, + "step": 661 + }, + { + "epoch": 0.13703167046160214, + "grad_norm": 0.007015643175691366, + "learning_rate": 0.00017331116192059864, + "loss": 0.0003, + "step": 662 + }, + { + "epoch": 0.13723866694266199, + "grad_norm": 0.0003168722032569349, + "learning_rate": 0.00017326959052172108, + "loss": 0.0, + "step": 663 + }, + { + "epoch": 0.1374456634237218, + "grad_norm": 0.006562775932252407, + "learning_rate": 0.0001732280191228435, + "loss": 0.0008, + "step": 664 + }, + { + "epoch": 0.13765265990478162, + "grad_norm": 0.003267984837293625, + "learning_rate": 0.00017318644772396593, + "loss": 0.0001, + "step": 665 + }, + { + "epoch": 0.13785965638584144, + "grad_norm": 0.007215241901576519, + "learning_rate": 0.00017314487632508836, + "loss": 0.0021, + "step": 666 + }, + { + "epoch": 0.13806665286690126, + "grad_norm": 0.001962031237781048, + "learning_rate": 0.00017310330492621077, + "loss": 0.0, + "step": 667 + }, + { + "epoch": 0.13827364934796108, + "grad_norm": 0.007086516357958317, + "learning_rate": 0.0001730617335273332, + "loss": 0.0021, + "step": 668 + }, + { + "epoch": 0.1384806458290209, + "grad_norm": 0.0063016172498464584, + "learning_rate": 0.00017302016212845562, + "loss": 0.0001, + "step": 669 + }, + { + "epoch": 0.13868764231008074, + "grad_norm": 0.007975582964718342, + "learning_rate": 0.00017297859072957806, + "loss": 0.0002, + "step": 670 + }, + { + "epoch": 0.13889463879114056, + "grad_norm": 0.0030251971911638975, + "learning_rate": 0.00017293701933070047, + "loss": 0.0002, + "step": 671 + }, + { + "epoch": 0.13910163527220037, + "grad_norm": 0.00741973053663969, + "learning_rate": 0.0001728954479318229, + "loss": 0.0003, + "step": 672 + }, + { + "epoch": 0.1393086317532602, + "grad_norm": 0.002640543272718787, + "learning_rate": 0.00017285387653294534, + "loss": 0.0015, + "step": 673 + }, + { + "epoch": 0.13951562823432, + "grad_norm": 0.0004313603858463466, + "learning_rate": 0.00017281230513406778, + "loss": 0.0, + "step": 674 + }, + { + "epoch": 0.13972262471537983, + "grad_norm": 0.0020118863321840763, + "learning_rate": 0.0001727707337351902, + "loss": 0.0005, + "step": 675 + }, + { + "epoch": 0.13992962119643967, + "grad_norm": 0.003337120870128274, + "learning_rate": 0.00017272916233631262, + "loss": 0.0007, + "step": 676 + }, + { + "epoch": 0.1401366176774995, + "grad_norm": 0.014386707916855812, + "learning_rate": 0.00017268759093743506, + "loss": 0.0004, + "step": 677 + }, + { + "epoch": 0.1403436141585593, + "grad_norm": 0.006729326210916042, + "learning_rate": 0.00017264601953855747, + "loss": 0.001, + "step": 678 + }, + { + "epoch": 0.14055061063961913, + "grad_norm": 0.001671936479397118, + "learning_rate": 0.0001726044481396799, + "loss": 0.0002, + "step": 679 + }, + { + "epoch": 0.14075760712067895, + "grad_norm": 0.007516622077673674, + "learning_rate": 0.00017256287674080235, + "loss": 0.001, + "step": 680 + }, + { + "epoch": 0.14096460360173876, + "grad_norm": 0.0027280249632894993, + "learning_rate": 0.00017252130534192476, + "loss": 0.0006, + "step": 681 + }, + { + "epoch": 0.14117160008279858, + "grad_norm": 0.010556796565651894, + "learning_rate": 0.0001724797339430472, + "loss": 0.0003, + "step": 682 + }, + { + "epoch": 0.14137859656385843, + "grad_norm": 0.0027946115005761385, + "learning_rate": 0.00017243816254416963, + "loss": 0.0001, + "step": 683 + }, + { + "epoch": 0.14158559304491825, + "grad_norm": 0.00467882351949811, + "learning_rate": 0.00017239659114529204, + "loss": 0.0001, + "step": 684 + }, + { + "epoch": 0.14179258952597806, + "grad_norm": 0.004167881328612566, + "learning_rate": 0.00017235501974641448, + "loss": 0.0012, + "step": 685 + }, + { + "epoch": 0.14199958600703788, + "grad_norm": 0.0034762704744935036, + "learning_rate": 0.00017231344834753691, + "loss": 0.0007, + "step": 686 + }, + { + "epoch": 0.1422065824880977, + "grad_norm": 0.0005650786333717406, + "learning_rate": 0.00017227187694865935, + "loss": 0.0, + "step": 687 + }, + { + "epoch": 0.14241357896915752, + "grad_norm": 0.0043237158097326756, + "learning_rate": 0.00017223030554978176, + "loss": 0.0001, + "step": 688 + }, + { + "epoch": 0.14262057545021734, + "grad_norm": 0.0071853832341730595, + "learning_rate": 0.0001721887341509042, + "loss": 0.0004, + "step": 689 + }, + { + "epoch": 0.14282757193127718, + "grad_norm": 0.01868472993373871, + "learning_rate": 0.00017214716275202664, + "loss": 0.0009, + "step": 690 + }, + { + "epoch": 0.143034568412337, + "grad_norm": 0.001339295064099133, + "learning_rate": 0.00017210559135314902, + "loss": 0.0009, + "step": 691 + }, + { + "epoch": 0.14324156489339682, + "grad_norm": 0.00664726085960865, + "learning_rate": 0.00017206401995427146, + "loss": 0.0002, + "step": 692 + }, + { + "epoch": 0.14344856137445663, + "grad_norm": 0.006592089310288429, + "learning_rate": 0.0001720224485553939, + "loss": 0.0001, + "step": 693 + }, + { + "epoch": 0.14365555785551645, + "grad_norm": 0.0005503061693161726, + "learning_rate": 0.00017198087715651633, + "loss": 0.0, + "step": 694 + }, + { + "epoch": 0.14386255433657627, + "grad_norm": 0.003913522697985172, + "learning_rate": 0.00017193930575763874, + "loss": 0.0002, + "step": 695 + }, + { + "epoch": 0.1440695508176361, + "grad_norm": 0.004871245473623276, + "learning_rate": 0.00017189773435876118, + "loss": 0.0002, + "step": 696 + }, + { + "epoch": 0.14427654729869593, + "grad_norm": 0.007188999559730291, + "learning_rate": 0.0001718561629598836, + "loss": 0.0002, + "step": 697 + }, + { + "epoch": 0.14448354377975575, + "grad_norm": 0.003864140482619405, + "learning_rate": 0.00017181459156100602, + "loss": 0.0013, + "step": 698 + }, + { + "epoch": 0.14469054026081557, + "grad_norm": 0.005774588789790869, + "learning_rate": 0.00017177302016212846, + "loss": 0.0004, + "step": 699 + }, + { + "epoch": 0.1448975367418754, + "grad_norm": 0.002636535558849573, + "learning_rate": 0.0001717314487632509, + "loss": 0.0007, + "step": 700 + }, + { + "epoch": 0.1451045332229352, + "grad_norm": 0.03907289355993271, + "learning_rate": 0.0001716898773643733, + "loss": 0.0019, + "step": 701 + }, + { + "epoch": 0.14531152970399502, + "grad_norm": 0.005653630942106247, + "learning_rate": 0.00017164830596549574, + "loss": 0.0003, + "step": 702 + }, + { + "epoch": 0.14551852618505484, + "grad_norm": 0.003644258715212345, + "learning_rate": 0.00017160673456661818, + "loss": 0.0001, + "step": 703 + }, + { + "epoch": 0.1457255226661147, + "grad_norm": 0.0028953952714800835, + "learning_rate": 0.00017156516316774062, + "loss": 0.0001, + "step": 704 + }, + { + "epoch": 0.1459325191471745, + "grad_norm": 0.005685892421752214, + "learning_rate": 0.00017152359176886303, + "loss": 0.0002, + "step": 705 + }, + { + "epoch": 0.14613951562823432, + "grad_norm": 0.00946901086717844, + "learning_rate": 0.00017148202036998547, + "loss": 0.0008, + "step": 706 + }, + { + "epoch": 0.14634651210929414, + "grad_norm": 0.004027761984616518, + "learning_rate": 0.0001714404489711079, + "loss": 0.0001, + "step": 707 + }, + { + "epoch": 0.14655350859035396, + "grad_norm": 0.0014218458672985435, + "learning_rate": 0.0001713988775722303, + "loss": 0.0, + "step": 708 + }, + { + "epoch": 0.14676050507141378, + "grad_norm": 0.0058472915552556515, + "learning_rate": 0.00017135730617335275, + "loss": 0.0002, + "step": 709 + }, + { + "epoch": 0.1469675015524736, + "grad_norm": 0.004684192128479481, + "learning_rate": 0.0001713157347744752, + "loss": 0.0002, + "step": 710 + }, + { + "epoch": 0.14717449803353344, + "grad_norm": 0.002729298546910286, + "learning_rate": 0.0001712741633755976, + "loss": 0.0001, + "step": 711 + }, + { + "epoch": 0.14738149451459326, + "grad_norm": 0.003782545682042837, + "learning_rate": 0.00017123259197672, + "loss": 0.0022, + "step": 712 + }, + { + "epoch": 0.14758849099565308, + "grad_norm": 0.004307260736823082, + "learning_rate": 0.00017119102057784244, + "loss": 0.0009, + "step": 713 + }, + { + "epoch": 0.1477954874767129, + "grad_norm": 0.01339892577379942, + "learning_rate": 0.00017114944917896488, + "loss": 0.001, + "step": 714 + }, + { + "epoch": 0.1480024839577727, + "grad_norm": 0.0017793363658711314, + "learning_rate": 0.0001711078777800873, + "loss": 0.0, + "step": 715 + }, + { + "epoch": 0.14820948043883253, + "grad_norm": 0.0005680687027052045, + "learning_rate": 0.00017106630638120973, + "loss": 0.0, + "step": 716 + }, + { + "epoch": 0.14841647691989235, + "grad_norm": 0.0010823605116456747, + "learning_rate": 0.00017102473498233216, + "loss": 0.0, + "step": 717 + }, + { + "epoch": 0.1486234734009522, + "grad_norm": 0.006135303992778063, + "learning_rate": 0.00017098316358345457, + "loss": 0.0002, + "step": 718 + }, + { + "epoch": 0.148830469882012, + "grad_norm": 0.003215776290744543, + "learning_rate": 0.000170941592184577, + "loss": 0.0008, + "step": 719 + }, + { + "epoch": 0.14903746636307183, + "grad_norm": 0.00970076397061348, + "learning_rate": 0.00017090002078569945, + "loss": 0.0008, + "step": 720 + }, + { + "epoch": 0.14924446284413165, + "grad_norm": 0.0037311650812625885, + "learning_rate": 0.00017085844938682189, + "loss": 0.0007, + "step": 721 + }, + { + "epoch": 0.14945145932519147, + "grad_norm": 0.0035531132016330957, + "learning_rate": 0.0001708168779879443, + "loss": 0.0007, + "step": 722 + }, + { + "epoch": 0.14965845580625128, + "grad_norm": 0.0013675455702468753, + "learning_rate": 0.00017077530658906673, + "loss": 0.0002, + "step": 723 + }, + { + "epoch": 0.14986545228731113, + "grad_norm": 0.003667420009151101, + "learning_rate": 0.00017073373519018917, + "loss": 0.0001, + "step": 724 + }, + { + "epoch": 0.15007244876837095, + "grad_norm": 0.0006531656836159527, + "learning_rate": 0.00017069216379131158, + "loss": 0.0, + "step": 725 + }, + { + "epoch": 0.15027944524943077, + "grad_norm": 0.0029405278619378805, + "learning_rate": 0.00017065059239243402, + "loss": 0.0001, + "step": 726 + }, + { + "epoch": 0.15048644173049058, + "grad_norm": 0.0020144616719335318, + "learning_rate": 0.00017060902099355645, + "loss": 0.0001, + "step": 727 + }, + { + "epoch": 0.1506934382115504, + "grad_norm": 0.003123146714642644, + "learning_rate": 0.00017056744959467886, + "loss": 0.0007, + "step": 728 + }, + { + "epoch": 0.15090043469261022, + "grad_norm": 0.005841000005602837, + "learning_rate": 0.0001705258781958013, + "loss": 0.0001, + "step": 729 + }, + { + "epoch": 0.15110743117367004, + "grad_norm": 0.001898916088975966, + "learning_rate": 0.00017048430679692374, + "loss": 0.0, + "step": 730 + }, + { + "epoch": 0.15131442765472988, + "grad_norm": 0.0005505726439878345, + "learning_rate": 0.00017044273539804615, + "loss": 0.0, + "step": 731 + }, + { + "epoch": 0.1515214241357897, + "grad_norm": 0.022630905732512474, + "learning_rate": 0.00017040116399916859, + "loss": 0.0003, + "step": 732 + }, + { + "epoch": 0.15172842061684952, + "grad_norm": 0.0018513459945097566, + "learning_rate": 0.00017035959260029102, + "loss": 0.0, + "step": 733 + }, + { + "epoch": 0.15193541709790934, + "grad_norm": 0.006640856619924307, + "learning_rate": 0.00017031802120141343, + "loss": 0.0015, + "step": 734 + }, + { + "epoch": 0.15214241357896915, + "grad_norm": 0.010431594215333462, + "learning_rate": 0.00017027644980253584, + "loss": 0.0007, + "step": 735 + }, + { + "epoch": 0.15234941006002897, + "grad_norm": 0.0009595350711606443, + "learning_rate": 0.00017023487840365828, + "loss": 0.0001, + "step": 736 + }, + { + "epoch": 0.1525564065410888, + "grad_norm": 0.0019930503331124783, + "learning_rate": 0.00017019330700478072, + "loss": 0.0001, + "step": 737 + }, + { + "epoch": 0.15276340302214864, + "grad_norm": 0.0020235483534634113, + "learning_rate": 0.00017015173560590313, + "loss": 0.0012, + "step": 738 + }, + { + "epoch": 0.15297039950320845, + "grad_norm": 0.000323964050039649, + "learning_rate": 0.00017011016420702556, + "loss": 0.0, + "step": 739 + }, + { + "epoch": 0.15317739598426827, + "grad_norm": 0.004805979318916798, + "learning_rate": 0.000170068592808148, + "loss": 0.0016, + "step": 740 + }, + { + "epoch": 0.1533843924653281, + "grad_norm": 0.0007103011594153941, + "learning_rate": 0.00017002702140927044, + "loss": 0.0, + "step": 741 + }, + { + "epoch": 0.1535913889463879, + "grad_norm": 0.00901501253247261, + "learning_rate": 0.00016998545001039285, + "loss": 0.0001, + "step": 742 + }, + { + "epoch": 0.15379838542744773, + "grad_norm": 0.01626206934452057, + "learning_rate": 0.00016994387861151528, + "loss": 0.0001, + "step": 743 + }, + { + "epoch": 0.15400538190850754, + "grad_norm": 0.006600509863346815, + "learning_rate": 0.00016990230721263772, + "loss": 0.0022, + "step": 744 + }, + { + "epoch": 0.1542123783895674, + "grad_norm": 0.0031586415134370327, + "learning_rate": 0.00016986073581376013, + "loss": 0.0003, + "step": 745 + }, + { + "epoch": 0.1544193748706272, + "grad_norm": 0.00408458337187767, + "learning_rate": 0.00016981916441488257, + "loss": 0.0014, + "step": 746 + }, + { + "epoch": 0.15462637135168703, + "grad_norm": 0.006417447701096535, + "learning_rate": 0.000169777593016005, + "loss": 0.0013, + "step": 747 + }, + { + "epoch": 0.15483336783274684, + "grad_norm": 0.002676580101251602, + "learning_rate": 0.00016973602161712742, + "loss": 0.0009, + "step": 748 + }, + { + "epoch": 0.15504036431380666, + "grad_norm": 0.003124868730083108, + "learning_rate": 0.00016969445021824985, + "loss": 0.0001, + "step": 749 + }, + { + "epoch": 0.15524736079486648, + "grad_norm": 0.005617608781903982, + "learning_rate": 0.0001696528788193723, + "loss": 0.0009, + "step": 750 + }, + { + "epoch": 0.1554543572759263, + "grad_norm": 0.0029069185256958008, + "learning_rate": 0.00016961130742049473, + "loss": 0.0008, + "step": 751 + }, + { + "epoch": 0.15566135375698614, + "grad_norm": 0.008944474160671234, + "learning_rate": 0.00016956973602161714, + "loss": 0.0017, + "step": 752 + }, + { + "epoch": 0.15586835023804596, + "grad_norm": 0.004935794975608587, + "learning_rate": 0.00016952816462273957, + "loss": 0.0012, + "step": 753 + }, + { + "epoch": 0.15607534671910578, + "grad_norm": 0.0005579759599640965, + "learning_rate": 0.000169486593223862, + "loss": 0.0, + "step": 754 + }, + { + "epoch": 0.1562823432001656, + "grad_norm": 0.00902874581515789, + "learning_rate": 0.00016944502182498442, + "loss": 0.0004, + "step": 755 + }, + { + "epoch": 0.15648933968122541, + "grad_norm": 0.00498725613579154, + "learning_rate": 0.00016940345042610683, + "loss": 0.0001, + "step": 756 + }, + { + "epoch": 0.15669633616228523, + "grad_norm": 0.0004982489626854658, + "learning_rate": 0.00016936187902722927, + "loss": 0.0, + "step": 757 + }, + { + "epoch": 0.15690333264334505, + "grad_norm": 0.0011680921306833625, + "learning_rate": 0.0001693203076283517, + "loss": 0.0, + "step": 758 + }, + { + "epoch": 0.1571103291244049, + "grad_norm": 0.0013553223107010126, + "learning_rate": 0.00016927873622947412, + "loss": 0.0, + "step": 759 + }, + { + "epoch": 0.15731732560546471, + "grad_norm": 0.00549361202865839, + "learning_rate": 0.00016923716483059655, + "loss": 0.0016, + "step": 760 + }, + { + "epoch": 0.15752432208652453, + "grad_norm": 0.004852932877838612, + "learning_rate": 0.000169195593431719, + "loss": 0.0002, + "step": 761 + }, + { + "epoch": 0.15773131856758435, + "grad_norm": 0.0046032629907131195, + "learning_rate": 0.0001691540220328414, + "loss": 0.002, + "step": 762 + }, + { + "epoch": 0.15793831504864417, + "grad_norm": 0.009385612793266773, + "learning_rate": 0.00016911245063396384, + "loss": 0.0003, + "step": 763 + }, + { + "epoch": 0.15814531152970399, + "grad_norm": 0.0024257150944322348, + "learning_rate": 0.00016907087923508627, + "loss": 0.0009, + "step": 764 + }, + { + "epoch": 0.1583523080107638, + "grad_norm": 0.002726235194131732, + "learning_rate": 0.00016902930783620868, + "loss": 0.0012, + "step": 765 + }, + { + "epoch": 0.15855930449182365, + "grad_norm": 0.006497920490801334, + "learning_rate": 0.00016898773643733112, + "loss": 0.0005, + "step": 766 + }, + { + "epoch": 0.15876630097288347, + "grad_norm": 0.012873928062617779, + "learning_rate": 0.00016894616503845356, + "loss": 0.0021, + "step": 767 + }, + { + "epoch": 0.15897329745394329, + "grad_norm": 0.009931253269314766, + "learning_rate": 0.000168904593639576, + "loss": 0.0004, + "step": 768 + }, + { + "epoch": 0.1591802939350031, + "grad_norm": 0.0012783849379047751, + "learning_rate": 0.0001688630222406984, + "loss": 0.0, + "step": 769 + }, + { + "epoch": 0.15938729041606292, + "grad_norm": 0.0025215751957148314, + "learning_rate": 0.00016882145084182084, + "loss": 0.0001, + "step": 770 + }, + { + "epoch": 0.15959428689712274, + "grad_norm": 0.00574857834726572, + "learning_rate": 0.00016877987944294328, + "loss": 0.0002, + "step": 771 + }, + { + "epoch": 0.15980128337818258, + "grad_norm": 0.0008691879920661449, + "learning_rate": 0.0001687383080440657, + "loss": 0.0, + "step": 772 + }, + { + "epoch": 0.1600082798592424, + "grad_norm": 0.008225478231906891, + "learning_rate": 0.00016869673664518813, + "loss": 0.0002, + "step": 773 + }, + { + "epoch": 0.16021527634030222, + "grad_norm": 0.003890304360538721, + "learning_rate": 0.00016865516524631056, + "loss": 0.0014, + "step": 774 + }, + { + "epoch": 0.16042227282136204, + "grad_norm": 0.0011641031596809626, + "learning_rate": 0.00016861359384743297, + "loss": 0.0001, + "step": 775 + }, + { + "epoch": 0.16062926930242186, + "grad_norm": 0.008769871667027473, + "learning_rate": 0.0001685720224485554, + "loss": 0.0011, + "step": 776 + }, + { + "epoch": 0.16083626578348167, + "grad_norm": 0.005050954408943653, + "learning_rate": 0.00016853045104967785, + "loss": 0.0002, + "step": 777 + }, + { + "epoch": 0.1610432622645415, + "grad_norm": 0.002180990530177951, + "learning_rate": 0.00016848887965080026, + "loss": 0.0008, + "step": 778 + }, + { + "epoch": 0.16125025874560134, + "grad_norm": 0.0015876460820436478, + "learning_rate": 0.00016844730825192267, + "loss": 0.0, + "step": 779 + }, + { + "epoch": 0.16145725522666116, + "grad_norm": 0.004357179626822472, + "learning_rate": 0.0001684057368530451, + "loss": 0.0003, + "step": 780 + }, + { + "epoch": 0.16166425170772097, + "grad_norm": 0.0034056720323860645, + "learning_rate": 0.00016836416545416754, + "loss": 0.0012, + "step": 781 + }, + { + "epoch": 0.1618712481887808, + "grad_norm": 0.005545208230614662, + "learning_rate": 0.00016832259405528995, + "loss": 0.0001, + "step": 782 + }, + { + "epoch": 0.1620782446698406, + "grad_norm": 0.0002129770437022671, + "learning_rate": 0.0001682810226564124, + "loss": 0.0, + "step": 783 + }, + { + "epoch": 0.16228524115090043, + "grad_norm": 0.0036753590684384108, + "learning_rate": 0.00016823945125753482, + "loss": 0.0001, + "step": 784 + }, + { + "epoch": 0.16249223763196025, + "grad_norm": 0.0018491502851247787, + "learning_rate": 0.00016819787985865723, + "loss": 0.0008, + "step": 785 + }, + { + "epoch": 0.1626992341130201, + "grad_norm": 0.0006519712042063475, + "learning_rate": 0.00016815630845977967, + "loss": 0.0, + "step": 786 + }, + { + "epoch": 0.1629062305940799, + "grad_norm": 0.011139947921037674, + "learning_rate": 0.0001681147370609021, + "loss": 0.0011, + "step": 787 + }, + { + "epoch": 0.16311322707513973, + "grad_norm": 0.0020866713020950556, + "learning_rate": 0.00016807316566202455, + "loss": 0.0002, + "step": 788 + }, + { + "epoch": 0.16332022355619955, + "grad_norm": 0.0034007905051112175, + "learning_rate": 0.00016803159426314696, + "loss": 0.0008, + "step": 789 + }, + { + "epoch": 0.16352722003725936, + "grad_norm": 0.0017938032979145646, + "learning_rate": 0.0001679900228642694, + "loss": 0.0009, + "step": 790 + }, + { + "epoch": 0.16373421651831918, + "grad_norm": 0.005385685246437788, + "learning_rate": 0.00016794845146539183, + "loss": 0.001, + "step": 791 + }, + { + "epoch": 0.163941212999379, + "grad_norm": 0.010079730302095413, + "learning_rate": 0.00016790688006651424, + "loss": 0.0004, + "step": 792 + }, + { + "epoch": 0.16414820948043884, + "grad_norm": 0.005826961249113083, + "learning_rate": 0.00016786530866763668, + "loss": 0.0001, + "step": 793 + }, + { + "epoch": 0.16435520596149866, + "grad_norm": 0.002885566558688879, + "learning_rate": 0.00016782373726875911, + "loss": 0.0001, + "step": 794 + }, + { + "epoch": 0.16456220244255848, + "grad_norm": 0.004031067714095116, + "learning_rate": 0.00016778216586988152, + "loss": 0.0002, + "step": 795 + }, + { + "epoch": 0.1647691989236183, + "grad_norm": 0.0019721402786672115, + "learning_rate": 0.00016774059447100396, + "loss": 0.0001, + "step": 796 + }, + { + "epoch": 0.16497619540467812, + "grad_norm": 0.002213244093582034, + "learning_rate": 0.0001676990230721264, + "loss": 0.0006, + "step": 797 + }, + { + "epoch": 0.16518319188573793, + "grad_norm": 0.001942839939147234, + "learning_rate": 0.00016765745167324884, + "loss": 0.0, + "step": 798 + }, + { + "epoch": 0.16539018836679775, + "grad_norm": 0.003173516597598791, + "learning_rate": 0.00016761588027437125, + "loss": 0.0001, + "step": 799 + }, + { + "epoch": 0.1655971848478576, + "grad_norm": 0.004877821542322636, + "learning_rate": 0.00016757430887549366, + "loss": 0.0001, + "step": 800 + }, + { + "epoch": 0.16580418132891742, + "grad_norm": 0.006676991004496813, + "learning_rate": 0.0001675327374766161, + "loss": 0.0002, + "step": 801 + }, + { + "epoch": 0.16601117780997723, + "grad_norm": 0.0022598986979573965, + "learning_rate": 0.0001674911660777385, + "loss": 0.0003, + "step": 802 + }, + { + "epoch": 0.16621817429103705, + "grad_norm": 0.00012318776862230152, + "learning_rate": 0.00016744959467886094, + "loss": 0.0, + "step": 803 + }, + { + "epoch": 0.16642517077209687, + "grad_norm": 0.006045771297067404, + "learning_rate": 0.00016740802327998338, + "loss": 0.0003, + "step": 804 + }, + { + "epoch": 0.1666321672531567, + "grad_norm": 0.004370058421045542, + "learning_rate": 0.0001673664518811058, + "loss": 0.0001, + "step": 805 + }, + { + "epoch": 0.1668391637342165, + "grad_norm": 0.005490643437951803, + "learning_rate": 0.00016732488048222822, + "loss": 0.0022, + "step": 806 + }, + { + "epoch": 0.16704616021527635, + "grad_norm": 0.007493430282920599, + "learning_rate": 0.00016728330908335066, + "loss": 0.001, + "step": 807 + }, + { + "epoch": 0.16725315669633617, + "grad_norm": 0.0006159085314720869, + "learning_rate": 0.0001672417376844731, + "loss": 0.0, + "step": 808 + }, + { + "epoch": 0.167460153177396, + "grad_norm": 0.002211883431300521, + "learning_rate": 0.0001672001662855955, + "loss": 0.0, + "step": 809 + }, + { + "epoch": 0.1676671496584558, + "grad_norm": 0.0028680090326815844, + "learning_rate": 0.00016715859488671794, + "loss": 0.0011, + "step": 810 + }, + { + "epoch": 0.16787414613951562, + "grad_norm": 0.004992680158466101, + "learning_rate": 0.00016711702348784038, + "loss": 0.0002, + "step": 811 + }, + { + "epoch": 0.16808114262057544, + "grad_norm": 0.0024819490499794483, + "learning_rate": 0.0001670754520889628, + "loss": 0.0001, + "step": 812 + }, + { + "epoch": 0.16828813910163526, + "grad_norm": 0.001662694732658565, + "learning_rate": 0.00016703388069008523, + "loss": 0.0001, + "step": 813 + }, + { + "epoch": 0.1684951355826951, + "grad_norm": 0.0027136337012052536, + "learning_rate": 0.00016699230929120767, + "loss": 0.0015, + "step": 814 + }, + { + "epoch": 0.16870213206375492, + "grad_norm": 0.0055983890779316425, + "learning_rate": 0.00016695073789233008, + "loss": 0.0006, + "step": 815 + }, + { + "epoch": 0.16890912854481474, + "grad_norm": 0.0005543065490201116, + "learning_rate": 0.0001669091664934525, + "loss": 0.0, + "step": 816 + }, + { + "epoch": 0.16911612502587456, + "grad_norm": 0.006743449252098799, + "learning_rate": 0.00016686759509457495, + "loss": 0.0006, + "step": 817 + }, + { + "epoch": 0.16932312150693438, + "grad_norm": 0.005361751653254032, + "learning_rate": 0.0001668260236956974, + "loss": 0.0002, + "step": 818 + }, + { + "epoch": 0.1695301179879942, + "grad_norm": 0.015542850829660892, + "learning_rate": 0.0001667844522968198, + "loss": 0.0001, + "step": 819 + }, + { + "epoch": 0.169737114469054, + "grad_norm": 0.006788911763578653, + "learning_rate": 0.00016674288089794223, + "loss": 0.0004, + "step": 820 + }, + { + "epoch": 0.16994411095011386, + "grad_norm": 0.006434622220695019, + "learning_rate": 0.00016670130949906467, + "loss": 0.0004, + "step": 821 + }, + { + "epoch": 0.17015110743117368, + "grad_norm": 0.0024506154004484415, + "learning_rate": 0.00016665973810018708, + "loss": 0.0007, + "step": 822 + }, + { + "epoch": 0.1703581039122335, + "grad_norm": 0.000382046215236187, + "learning_rate": 0.0001666181667013095, + "loss": 0.0, + "step": 823 + }, + { + "epoch": 0.1705651003932933, + "grad_norm": 0.00432636309415102, + "learning_rate": 0.00016657659530243193, + "loss": 0.0009, + "step": 824 + }, + { + "epoch": 0.17077209687435313, + "grad_norm": 0.005686972755938768, + "learning_rate": 0.00016653502390355436, + "loss": 0.0002, + "step": 825 + }, + { + "epoch": 0.17097909335541295, + "grad_norm": 0.005743528716266155, + "learning_rate": 0.00016649345250467677, + "loss": 0.0003, + "step": 826 + }, + { + "epoch": 0.1711860898364728, + "grad_norm": 0.002116352552548051, + "learning_rate": 0.0001664518811057992, + "loss": 0.0003, + "step": 827 + }, + { + "epoch": 0.1713930863175326, + "grad_norm": 0.002352718496695161, + "learning_rate": 0.00016641030970692165, + "loss": 0.0, + "step": 828 + }, + { + "epoch": 0.17160008279859243, + "grad_norm": 0.0044693113304674625, + "learning_rate": 0.00016636873830804406, + "loss": 0.0004, + "step": 829 + }, + { + "epoch": 0.17180707927965225, + "grad_norm": 0.0005167116178199649, + "learning_rate": 0.0001663271669091665, + "loss": 0.0, + "step": 830 + }, + { + "epoch": 0.17201407576071207, + "grad_norm": 0.005162122659385204, + "learning_rate": 0.00016628559551028893, + "loss": 0.0002, + "step": 831 + }, + { + "epoch": 0.17222107224177188, + "grad_norm": 0.00015954635455273092, + "learning_rate": 0.00016624402411141134, + "loss": 0.0, + "step": 832 + }, + { + "epoch": 0.1724280687228317, + "grad_norm": 0.0030487151816487312, + "learning_rate": 0.00016620245271253378, + "loss": 0.0001, + "step": 833 + }, + { + "epoch": 0.17263506520389155, + "grad_norm": 0.002151534892618656, + "learning_rate": 0.00016616088131365622, + "loss": 0.0002, + "step": 834 + }, + { + "epoch": 0.17284206168495136, + "grad_norm": 0.0044494629837572575, + "learning_rate": 0.00016611930991477865, + "loss": 0.0005, + "step": 835 + }, + { + "epoch": 0.17304905816601118, + "grad_norm": 0.00033838755916804075, + "learning_rate": 0.00016607773851590106, + "loss": 0.0, + "step": 836 + }, + { + "epoch": 0.173256054647071, + "grad_norm": 0.0005302856443449855, + "learning_rate": 0.0001660361671170235, + "loss": 0.0, + "step": 837 + }, + { + "epoch": 0.17346305112813082, + "grad_norm": 0.0013208640739321709, + "learning_rate": 0.00016599459571814594, + "loss": 0.0, + "step": 838 + }, + { + "epoch": 0.17367004760919064, + "grad_norm": 0.001052051316946745, + "learning_rate": 0.00016595302431926835, + "loss": 0.0, + "step": 839 + }, + { + "epoch": 0.17387704409025045, + "grad_norm": 0.018445929512381554, + "learning_rate": 0.00016591145292039079, + "loss": 0.0004, + "step": 840 + }, + { + "epoch": 0.1740840405713103, + "grad_norm": 0.0025256802327930927, + "learning_rate": 0.00016586988152151322, + "loss": 0.0001, + "step": 841 + }, + { + "epoch": 0.17429103705237012, + "grad_norm": 0.0014724883949384093, + "learning_rate": 0.00016582831012263563, + "loss": 0.0, + "step": 842 + }, + { + "epoch": 0.17449803353342994, + "grad_norm": 0.003576815826818347, + "learning_rate": 0.00016578673872375807, + "loss": 0.0001, + "step": 843 + }, + { + "epoch": 0.17470503001448975, + "grad_norm": 0.0006163385114632547, + "learning_rate": 0.00016574516732488048, + "loss": 0.0, + "step": 844 + }, + { + "epoch": 0.17491202649554957, + "grad_norm": 0.0011656074784696102, + "learning_rate": 0.00016570359592600292, + "loss": 0.0001, + "step": 845 + }, + { + "epoch": 0.1751190229766094, + "grad_norm": 0.0018338944064453244, + "learning_rate": 0.00016566202452712533, + "loss": 0.0001, + "step": 846 + }, + { + "epoch": 0.1753260194576692, + "grad_norm": 0.005035779904574156, + "learning_rate": 0.00016562045312824776, + "loss": 0.0014, + "step": 847 + }, + { + "epoch": 0.17553301593872905, + "grad_norm": 0.006770180072635412, + "learning_rate": 0.0001655788817293702, + "loss": 0.0012, + "step": 848 + }, + { + "epoch": 0.17574001241978887, + "grad_norm": 0.0003650276339612901, + "learning_rate": 0.0001655373103304926, + "loss": 0.0, + "step": 849 + }, + { + "epoch": 0.1759470089008487, + "grad_norm": 0.00023851868172641844, + "learning_rate": 0.00016549573893161505, + "loss": 0.0, + "step": 850 + }, + { + "epoch": 0.1761540053819085, + "grad_norm": 0.014695384539663792, + "learning_rate": 0.00016545416753273748, + "loss": 0.0011, + "step": 851 + }, + { + "epoch": 0.17636100186296833, + "grad_norm": 0.00036404369166120887, + "learning_rate": 0.00016541259613385992, + "loss": 0.0, + "step": 852 + }, + { + "epoch": 0.17656799834402814, + "grad_norm": 0.002682497026398778, + "learning_rate": 0.00016537102473498233, + "loss": 0.0012, + "step": 853 + }, + { + "epoch": 0.17677499482508796, + "grad_norm": 0.007028127089142799, + "learning_rate": 0.00016532945333610477, + "loss": 0.0002, + "step": 854 + }, + { + "epoch": 0.1769819913061478, + "grad_norm": 0.0012324461713433266, + "learning_rate": 0.0001652878819372272, + "loss": 0.0, + "step": 855 + }, + { + "epoch": 0.17718898778720762, + "grad_norm": 0.00292210397310555, + "learning_rate": 0.00016524631053834962, + "loss": 0.0, + "step": 856 + }, + { + "epoch": 0.17739598426826744, + "grad_norm": 0.0014698312152177095, + "learning_rate": 0.00016520473913947205, + "loss": 0.0002, + "step": 857 + }, + { + "epoch": 0.17760298074932726, + "grad_norm": 0.0022247894667088985, + "learning_rate": 0.0001651631677405945, + "loss": 0.0, + "step": 858 + }, + { + "epoch": 0.17780997723038708, + "grad_norm": 0.0006738615338690579, + "learning_rate": 0.0001651215963417169, + "loss": 0.0, + "step": 859 + }, + { + "epoch": 0.1780169737114469, + "grad_norm": 0.004056425765156746, + "learning_rate": 0.00016508002494283934, + "loss": 0.001, + "step": 860 + }, + { + "epoch": 0.17822397019250671, + "grad_norm": 0.006607827264815569, + "learning_rate": 0.00016503845354396177, + "loss": 0.0003, + "step": 861 + }, + { + "epoch": 0.17843096667356656, + "grad_norm": 0.007498994003981352, + "learning_rate": 0.00016499688214508418, + "loss": 0.0009, + "step": 862 + }, + { + "epoch": 0.17863796315462638, + "grad_norm": 0.002715140348300338, + "learning_rate": 0.00016495531074620662, + "loss": 0.0001, + "step": 863 + }, + { + "epoch": 0.1788449596356862, + "grad_norm": 0.006559406872838736, + "learning_rate": 0.00016491373934732906, + "loss": 0.0012, + "step": 864 + }, + { + "epoch": 0.179051956116746, + "grad_norm": 0.005900564603507519, + "learning_rate": 0.0001648721679484515, + "loss": 0.0001, + "step": 865 + }, + { + "epoch": 0.17925895259780583, + "grad_norm": 0.0014680642634630203, + "learning_rate": 0.00016483059654957388, + "loss": 0.0004, + "step": 866 + }, + { + "epoch": 0.17946594907886565, + "grad_norm": 0.0017862527165561914, + "learning_rate": 0.00016478902515069632, + "loss": 0.0001, + "step": 867 + }, + { + "epoch": 0.17967294555992547, + "grad_norm": 0.0010660128900781274, + "learning_rate": 0.00016474745375181875, + "loss": 0.0001, + "step": 868 + }, + { + "epoch": 0.1798799420409853, + "grad_norm": 0.010508016683161259, + "learning_rate": 0.0001647058823529412, + "loss": 0.0004, + "step": 869 + }, + { + "epoch": 0.18008693852204513, + "grad_norm": 0.0026938568335026503, + "learning_rate": 0.0001646643109540636, + "loss": 0.0001, + "step": 870 + }, + { + "epoch": 0.18029393500310495, + "grad_norm": 0.0015470877988263965, + "learning_rate": 0.00016462273955518604, + "loss": 0.0, + "step": 871 + }, + { + "epoch": 0.18050093148416477, + "grad_norm": 0.003435211256146431, + "learning_rate": 0.00016458116815630847, + "loss": 0.0001, + "step": 872 + }, + { + "epoch": 0.18070792796522459, + "grad_norm": 0.010342281311750412, + "learning_rate": 0.00016453959675743088, + "loss": 0.0007, + "step": 873 + }, + { + "epoch": 0.1809149244462844, + "grad_norm": 0.0007751841330900788, + "learning_rate": 0.00016449802535855332, + "loss": 0.0, + "step": 874 + }, + { + "epoch": 0.18112192092734425, + "grad_norm": 0.0003991715202573687, + "learning_rate": 0.00016445645395967576, + "loss": 0.0, + "step": 875 + }, + { + "epoch": 0.18132891740840407, + "grad_norm": 0.004742010496556759, + "learning_rate": 0.00016441488256079817, + "loss": 0.0001, + "step": 876 + }, + { + "epoch": 0.18153591388946388, + "grad_norm": 0.0010139705846086144, + "learning_rate": 0.0001643733111619206, + "loss": 0.0, + "step": 877 + }, + { + "epoch": 0.1817429103705237, + "grad_norm": 0.00697368336841464, + "learning_rate": 0.00016433173976304304, + "loss": 0.0003, + "step": 878 + }, + { + "epoch": 0.18194990685158352, + "grad_norm": 0.0056029148399829865, + "learning_rate": 0.00016429016836416545, + "loss": 0.0003, + "step": 879 + }, + { + "epoch": 0.18215690333264334, + "grad_norm": 0.0031287583988159895, + "learning_rate": 0.0001642485969652879, + "loss": 0.0001, + "step": 880 + }, + { + "epoch": 0.18236389981370316, + "grad_norm": 0.0005836491473019123, + "learning_rate": 0.00016420702556641033, + "loss": 0.0, + "step": 881 + }, + { + "epoch": 0.182570896294763, + "grad_norm": 0.006221551448106766, + "learning_rate": 0.00016416545416753276, + "loss": 0.0014, + "step": 882 + }, + { + "epoch": 0.18277789277582282, + "grad_norm": 0.00045936627429910004, + "learning_rate": 0.00016412388276865517, + "loss": 0.0, + "step": 883 + }, + { + "epoch": 0.18298488925688264, + "grad_norm": 0.0006924067274667323, + "learning_rate": 0.0001640823113697776, + "loss": 0.0, + "step": 884 + }, + { + "epoch": 0.18319188573794246, + "grad_norm": 0.010869395919144154, + "learning_rate": 0.00016404073997090005, + "loss": 0.0013, + "step": 885 + }, + { + "epoch": 0.18339888221900227, + "grad_norm": 0.00480787456035614, + "learning_rate": 0.00016399916857202246, + "loss": 0.0009, + "step": 886 + }, + { + "epoch": 0.1836058787000621, + "grad_norm": 0.0004685772000811994, + "learning_rate": 0.0001639575971731449, + "loss": 0.0, + "step": 887 + }, + { + "epoch": 0.1838128751811219, + "grad_norm": 0.00303410436026752, + "learning_rate": 0.0001639160257742673, + "loss": 0.0007, + "step": 888 + }, + { + "epoch": 0.18401987166218176, + "grad_norm": 0.001141547691076994, + "learning_rate": 0.00016387445437538974, + "loss": 0.0001, + "step": 889 + }, + { + "epoch": 0.18422686814324157, + "grad_norm": 0.007433968596160412, + "learning_rate": 0.00016383288297651215, + "loss": 0.0001, + "step": 890 + }, + { + "epoch": 0.1844338646243014, + "grad_norm": 0.003386344527825713, + "learning_rate": 0.0001637913115776346, + "loss": 0.0003, + "step": 891 + }, + { + "epoch": 0.1846408611053612, + "grad_norm": 0.0011372484732419252, + "learning_rate": 0.00016374974017875702, + "loss": 0.0, + "step": 892 + }, + { + "epoch": 0.18484785758642103, + "grad_norm": 0.0039020997937768698, + "learning_rate": 0.00016370816877987943, + "loss": 0.0001, + "step": 893 + }, + { + "epoch": 0.18505485406748085, + "grad_norm": 0.003088288474828005, + "learning_rate": 0.00016366659738100187, + "loss": 0.0009, + "step": 894 + }, + { + "epoch": 0.18526185054854066, + "grad_norm": 0.001126794726587832, + "learning_rate": 0.0001636250259821243, + "loss": 0.0001, + "step": 895 + }, + { + "epoch": 0.1854688470296005, + "grad_norm": 0.007449139375239611, + "learning_rate": 0.00016358345458324672, + "loss": 0.0004, + "step": 896 + }, + { + "epoch": 0.18567584351066033, + "grad_norm": 0.005704225040972233, + "learning_rate": 0.00016354188318436916, + "loss": 0.0016, + "step": 897 + }, + { + "epoch": 0.18588283999172014, + "grad_norm": 0.004983640741556883, + "learning_rate": 0.0001635003117854916, + "loss": 0.0016, + "step": 898 + }, + { + "epoch": 0.18608983647277996, + "grad_norm": 0.00034120268537662923, + "learning_rate": 0.00016345874038661403, + "loss": 0.0, + "step": 899 + }, + { + "epoch": 0.18629683295383978, + "grad_norm": 0.007043101824820042, + "learning_rate": 0.00016341716898773644, + "loss": 0.001, + "step": 900 + }, + { + "epoch": 0.1865038294348996, + "grad_norm": 0.004050271585583687, + "learning_rate": 0.00016337559758885888, + "loss": 0.0001, + "step": 901 + }, + { + "epoch": 0.18671082591595942, + "grad_norm": 0.001882696757093072, + "learning_rate": 0.00016333402618998131, + "loss": 0.0005, + "step": 902 + }, + { + "epoch": 0.18691782239701926, + "grad_norm": 0.002479350659996271, + "learning_rate": 0.00016329245479110372, + "loss": 0.0003, + "step": 903 + }, + { + "epoch": 0.18712481887807908, + "grad_norm": 0.00246567465364933, + "learning_rate": 0.00016325088339222616, + "loss": 0.0001, + "step": 904 + }, + { + "epoch": 0.1873318153591389, + "grad_norm": 0.0021426973398774862, + "learning_rate": 0.0001632093119933486, + "loss": 0.0006, + "step": 905 + }, + { + "epoch": 0.18753881184019872, + "grad_norm": 0.004363594576716423, + "learning_rate": 0.000163167740594471, + "loss": 0.0002, + "step": 906 + }, + { + "epoch": 0.18774580832125853, + "grad_norm": 0.004984852857887745, + "learning_rate": 0.00016312616919559345, + "loss": 0.0004, + "step": 907 + }, + { + "epoch": 0.18795280480231835, + "grad_norm": 0.004489907994866371, + "learning_rate": 0.00016308459779671588, + "loss": 0.0001, + "step": 908 + }, + { + "epoch": 0.18815980128337817, + "grad_norm": 0.0013233786448836327, + "learning_rate": 0.0001630430263978383, + "loss": 0.0, + "step": 909 + }, + { + "epoch": 0.18836679776443802, + "grad_norm": 0.00894436426460743, + "learning_rate": 0.0001630014549989607, + "loss": 0.0016, + "step": 910 + }, + { + "epoch": 0.18857379424549783, + "grad_norm": 0.001729694427922368, + "learning_rate": 0.00016295988360008314, + "loss": 0.001, + "step": 911 + }, + { + "epoch": 0.18878079072655765, + "grad_norm": 0.005005873739719391, + "learning_rate": 0.00016291831220120558, + "loss": 0.0005, + "step": 912 + }, + { + "epoch": 0.18898778720761747, + "grad_norm": 0.0007573101902380586, + "learning_rate": 0.00016287674080232799, + "loss": 0.0, + "step": 913 + }, + { + "epoch": 0.1891947836886773, + "grad_norm": 0.005315006244927645, + "learning_rate": 0.00016283516940345042, + "loss": 0.0001, + "step": 914 + }, + { + "epoch": 0.1894017801697371, + "grad_norm": 0.001140634878538549, + "learning_rate": 0.00016279359800457286, + "loss": 0.0001, + "step": 915 + }, + { + "epoch": 0.18960877665079692, + "grad_norm": 0.003881396260112524, + "learning_rate": 0.0001627520266056953, + "loss": 0.0001, + "step": 916 + }, + { + "epoch": 0.18981577313185677, + "grad_norm": 0.001353550935164094, + "learning_rate": 0.0001627104552068177, + "loss": 0.0, + "step": 917 + }, + { + "epoch": 0.1900227696129166, + "grad_norm": 0.001415180740877986, + "learning_rate": 0.00016266888380794014, + "loss": 0.0, + "step": 918 + }, + { + "epoch": 0.1902297660939764, + "grad_norm": 0.006398684810847044, + "learning_rate": 0.00016262731240906258, + "loss": 0.0006, + "step": 919 + }, + { + "epoch": 0.19043676257503622, + "grad_norm": 0.005204816348850727, + "learning_rate": 0.000162585741010185, + "loss": 0.0016, + "step": 920 + }, + { + "epoch": 0.19064375905609604, + "grad_norm": 0.0015194268198683858, + "learning_rate": 0.00016254416961130743, + "loss": 0.0007, + "step": 921 + }, + { + "epoch": 0.19085075553715586, + "grad_norm": 0.002916971454396844, + "learning_rate": 0.00016250259821242987, + "loss": 0.0015, + "step": 922 + }, + { + "epoch": 0.1910577520182157, + "grad_norm": 0.00017840563668869436, + "learning_rate": 0.00016246102681355228, + "loss": 0.0, + "step": 923 + }, + { + "epoch": 0.19126474849927552, + "grad_norm": 0.0017515165964141488, + "learning_rate": 0.0001624194554146747, + "loss": 0.0005, + "step": 924 + }, + { + "epoch": 0.19147174498033534, + "grad_norm": 0.0011207156348973513, + "learning_rate": 0.00016237788401579715, + "loss": 0.0, + "step": 925 + }, + { + "epoch": 0.19167874146139516, + "grad_norm": 0.00647772429510951, + "learning_rate": 0.00016233631261691956, + "loss": 0.0004, + "step": 926 + }, + { + "epoch": 0.19188573794245498, + "grad_norm": 0.0009239514474757016, + "learning_rate": 0.000162294741218042, + "loss": 0.0, + "step": 927 + }, + { + "epoch": 0.1920927344235148, + "grad_norm": 0.0009718858054839075, + "learning_rate": 0.00016225316981916443, + "loss": 0.0, + "step": 928 + }, + { + "epoch": 0.1922997309045746, + "grad_norm": 0.0009835285600274801, + "learning_rate": 0.00016221159842028687, + "loss": 0.0001, + "step": 929 + }, + { + "epoch": 0.19250672738563446, + "grad_norm": 0.003986849449574947, + "learning_rate": 0.00016217002702140928, + "loss": 0.0005, + "step": 930 + }, + { + "epoch": 0.19271372386669428, + "grad_norm": 0.0055690668523311615, + "learning_rate": 0.00016212845562253172, + "loss": 0.0002, + "step": 931 + }, + { + "epoch": 0.1929207203477541, + "grad_norm": 0.006283191032707691, + "learning_rate": 0.00016208688422365413, + "loss": 0.0003, + "step": 932 + }, + { + "epoch": 0.1931277168288139, + "grad_norm": 0.00035167241003364325, + "learning_rate": 0.00016204531282477656, + "loss": 0.0, + "step": 933 + }, + { + "epoch": 0.19333471330987373, + "grad_norm": 0.001550202607177198, + "learning_rate": 0.00016200374142589897, + "loss": 0.0001, + "step": 934 + }, + { + "epoch": 0.19354170979093355, + "grad_norm": 0.0009650330757722259, + "learning_rate": 0.0001619621700270214, + "loss": 0.0, + "step": 935 + }, + { + "epoch": 0.19374870627199337, + "grad_norm": 0.006459403783082962, + "learning_rate": 0.00016192059862814385, + "loss": 0.0009, + "step": 936 + }, + { + "epoch": 0.1939557027530532, + "grad_norm": 0.0006884423783048987, + "learning_rate": 0.00016187902722926626, + "loss": 0.0, + "step": 937 + }, + { + "epoch": 0.19416269923411303, + "grad_norm": 0.0034009867813438177, + "learning_rate": 0.0001618374558303887, + "loss": 0.0001, + "step": 938 + }, + { + "epoch": 0.19436969571517285, + "grad_norm": 0.001749175600707531, + "learning_rate": 0.00016179588443151113, + "loss": 0.0012, + "step": 939 + }, + { + "epoch": 0.19457669219623266, + "grad_norm": 0.005409194156527519, + "learning_rate": 0.00016175431303263354, + "loss": 0.0002, + "step": 940 + }, + { + "epoch": 0.19478368867729248, + "grad_norm": 0.0033904362935572863, + "learning_rate": 0.00016171274163375598, + "loss": 0.0001, + "step": 941 + }, + { + "epoch": 0.1949906851583523, + "grad_norm": 0.005800081882625818, + "learning_rate": 0.00016167117023487842, + "loss": 0.0014, + "step": 942 + }, + { + "epoch": 0.19519768163941212, + "grad_norm": 0.001085714902728796, + "learning_rate": 0.00016162959883600083, + "loss": 0.0001, + "step": 943 + }, + { + "epoch": 0.19540467812047196, + "grad_norm": 0.0017082407139241695, + "learning_rate": 0.00016158802743712326, + "loss": 0.0001, + "step": 944 + }, + { + "epoch": 0.19561167460153178, + "grad_norm": 0.0016056247986853123, + "learning_rate": 0.0001615464560382457, + "loss": 0.0009, + "step": 945 + }, + { + "epoch": 0.1958186710825916, + "grad_norm": 0.0005931173800490797, + "learning_rate": 0.00016150488463936814, + "loss": 0.0, + "step": 946 + }, + { + "epoch": 0.19602566756365142, + "grad_norm": 0.0002708766842260957, + "learning_rate": 0.00016146331324049055, + "loss": 0.0, + "step": 947 + }, + { + "epoch": 0.19623266404471124, + "grad_norm": 0.003350366372615099, + "learning_rate": 0.00016142174184161299, + "loss": 0.0004, + "step": 948 + }, + { + "epoch": 0.19643966052577105, + "grad_norm": 0.00030215582228265703, + "learning_rate": 0.00016138017044273542, + "loss": 0.0, + "step": 949 + }, + { + "epoch": 0.19664665700683087, + "grad_norm": 0.0013855715515092015, + "learning_rate": 0.00016133859904385783, + "loss": 0.0009, + "step": 950 + }, + { + "epoch": 0.19685365348789072, + "grad_norm": 0.0005864354898221791, + "learning_rate": 0.00016129702764498027, + "loss": 0.0, + "step": 951 + }, + { + "epoch": 0.19706064996895054, + "grad_norm": 0.0006372429197654128, + "learning_rate": 0.0001612554562461027, + "loss": 0.0, + "step": 952 + }, + { + "epoch": 0.19726764645001035, + "grad_norm": 0.0005041586118750274, + "learning_rate": 0.00016121388484722512, + "loss": 0.0, + "step": 953 + }, + { + "epoch": 0.19747464293107017, + "grad_norm": 0.0023472902830690145, + "learning_rate": 0.00016117231344834753, + "loss": 0.001, + "step": 954 + }, + { + "epoch": 0.19768163941213, + "grad_norm": 0.00015194782463368028, + "learning_rate": 0.00016113074204946996, + "loss": 0.0, + "step": 955 + }, + { + "epoch": 0.1978886358931898, + "grad_norm": 0.001190232578665018, + "learning_rate": 0.0001610891706505924, + "loss": 0.0002, + "step": 956 + }, + { + "epoch": 0.19809563237424963, + "grad_norm": 0.0018357646185904741, + "learning_rate": 0.0001610475992517148, + "loss": 0.0002, + "step": 957 + }, + { + "epoch": 0.19830262885530947, + "grad_norm": 0.007886867970228195, + "learning_rate": 0.00016100602785283725, + "loss": 0.0015, + "step": 958 + }, + { + "epoch": 0.1985096253363693, + "grad_norm": 0.00014407855633180588, + "learning_rate": 0.00016096445645395968, + "loss": 0.0, + "step": 959 + }, + { + "epoch": 0.1987166218174291, + "grad_norm": 0.0008407345740124583, + "learning_rate": 0.0001609228850550821, + "loss": 0.0001, + "step": 960 + }, + { + "epoch": 0.19892361829848892, + "grad_norm": 0.0005690194084309042, + "learning_rate": 0.00016088131365620453, + "loss": 0.0, + "step": 961 + }, + { + "epoch": 0.19913061477954874, + "grad_norm": 0.0001164446584880352, + "learning_rate": 0.00016083974225732697, + "loss": 0.0, + "step": 962 + }, + { + "epoch": 0.19933761126060856, + "grad_norm": 0.004800689872354269, + "learning_rate": 0.0001607981708584494, + "loss": 0.0009, + "step": 963 + }, + { + "epoch": 0.19954460774166838, + "grad_norm": 0.0038551113102585077, + "learning_rate": 0.00016075659945957182, + "loss": 0.0002, + "step": 964 + }, + { + "epoch": 0.19975160422272822, + "grad_norm": 0.00023845378018449992, + "learning_rate": 0.00016071502806069425, + "loss": 0.0, + "step": 965 + }, + { + "epoch": 0.19995860070378804, + "grad_norm": 0.0006543719209730625, + "learning_rate": 0.0001606734566618167, + "loss": 0.0, + "step": 966 + }, + { + "epoch": 0.20016559718484786, + "grad_norm": 0.0024344087578356266, + "learning_rate": 0.0001606318852629391, + "loss": 0.0011, + "step": 967 + }, + { + "epoch": 0.20037259366590768, + "grad_norm": 0.0006420607678592205, + "learning_rate": 0.00016059031386406154, + "loss": 0.0, + "step": 968 + }, + { + "epoch": 0.2005795901469675, + "grad_norm": 0.0016330952057614923, + "learning_rate": 0.00016054874246518397, + "loss": 0.0, + "step": 969 + }, + { + "epoch": 0.2007865866280273, + "grad_norm": 0.0013299377169460058, + "learning_rate": 0.00016050717106630638, + "loss": 0.0001, + "step": 970 + }, + { + "epoch": 0.20099358310908713, + "grad_norm": 0.0023206102196127176, + "learning_rate": 0.00016046559966742882, + "loss": 0.0006, + "step": 971 + }, + { + "epoch": 0.20120057959014698, + "grad_norm": 0.0030964380130171776, + "learning_rate": 0.00016042402826855126, + "loss": 0.0004, + "step": 972 + }, + { + "epoch": 0.2014075760712068, + "grad_norm": 0.0015272133750841022, + "learning_rate": 0.00016038245686967367, + "loss": 0.0, + "step": 973 + }, + { + "epoch": 0.2016145725522666, + "grad_norm": 0.0036174836568534374, + "learning_rate": 0.0001603408854707961, + "loss": 0.0002, + "step": 974 + }, + { + "epoch": 0.20182156903332643, + "grad_norm": 0.0014752513961866498, + "learning_rate": 0.00016029931407191854, + "loss": 0.0, + "step": 975 + }, + { + "epoch": 0.20202856551438625, + "grad_norm": 0.0008594008395448327, + "learning_rate": 0.00016025774267304095, + "loss": 0.0001, + "step": 976 + }, + { + "epoch": 0.20223556199544607, + "grad_norm": 0.007280942518264055, + "learning_rate": 0.00016021617127416336, + "loss": 0.001, + "step": 977 + }, + { + "epoch": 0.2024425584765059, + "grad_norm": 0.0013399991439655423, + "learning_rate": 0.0001601745998752858, + "loss": 0.0, + "step": 978 + }, + { + "epoch": 0.20264955495756573, + "grad_norm": 0.0015200217021629214, + "learning_rate": 0.00016013302847640824, + "loss": 0.0003, + "step": 979 + }, + { + "epoch": 0.20285655143862555, + "grad_norm": 0.008712020702660084, + "learning_rate": 0.00016009145707753067, + "loss": 0.0002, + "step": 980 + }, + { + "epoch": 0.20306354791968537, + "grad_norm": 0.004218498710542917, + "learning_rate": 0.00016004988567865308, + "loss": 0.0006, + "step": 981 + }, + { + "epoch": 0.20327054440074518, + "grad_norm": 0.0019195530330762267, + "learning_rate": 0.00016000831427977552, + "loss": 0.0008, + "step": 982 + }, + { + "epoch": 0.203477540881805, + "grad_norm": 0.0017306975787505507, + "learning_rate": 0.00015996674288089796, + "loss": 0.001, + "step": 983 + }, + { + "epoch": 0.20368453736286482, + "grad_norm": 0.006045056506991386, + "learning_rate": 0.00015992517148202037, + "loss": 0.0002, + "step": 984 + }, + { + "epoch": 0.20389153384392467, + "grad_norm": 0.002741064177826047, + "learning_rate": 0.0001598836000831428, + "loss": 0.0001, + "step": 985 + }, + { + "epoch": 0.20409853032498448, + "grad_norm": 0.0026846020482480526, + "learning_rate": 0.00015984202868426524, + "loss": 0.0003, + "step": 986 + }, + { + "epoch": 0.2043055268060443, + "grad_norm": 0.009860471822321415, + "learning_rate": 0.00015980045728538765, + "loss": 0.0002, + "step": 987 + }, + { + "epoch": 0.20451252328710412, + "grad_norm": 0.0001563982223160565, + "learning_rate": 0.0001597588858865101, + "loss": 0.0, + "step": 988 + }, + { + "epoch": 0.20471951976816394, + "grad_norm": 0.0035680129658430815, + "learning_rate": 0.00015971731448763253, + "loss": 0.0009, + "step": 989 + }, + { + "epoch": 0.20492651624922376, + "grad_norm": 0.00037079930189065635, + "learning_rate": 0.00015967574308875494, + "loss": 0.0, + "step": 990 + }, + { + "epoch": 0.20513351273028357, + "grad_norm": 0.006476435344666243, + "learning_rate": 0.00015963417168987737, + "loss": 0.0004, + "step": 991 + }, + { + "epoch": 0.20534050921134342, + "grad_norm": 0.0008909539901651442, + "learning_rate": 0.0001595926002909998, + "loss": 0.0001, + "step": 992 + }, + { + "epoch": 0.20554750569240324, + "grad_norm": 0.005418546497821808, + "learning_rate": 0.00015955102889212225, + "loss": 0.0009, + "step": 993 + }, + { + "epoch": 0.20575450217346306, + "grad_norm": 0.004925790708512068, + "learning_rate": 0.00015950945749324466, + "loss": 0.0009, + "step": 994 + }, + { + "epoch": 0.20596149865452287, + "grad_norm": 0.003130319295451045, + "learning_rate": 0.0001594678860943671, + "loss": 0.001, + "step": 995 + }, + { + "epoch": 0.2061684951355827, + "grad_norm": 0.0062978435307741165, + "learning_rate": 0.00015942631469548953, + "loss": 0.0002, + "step": 996 + }, + { + "epoch": 0.2063754916166425, + "grad_norm": 0.006842117290943861, + "learning_rate": 0.00015938474329661194, + "loss": 0.0003, + "step": 997 + }, + { + "epoch": 0.20658248809770233, + "grad_norm": 0.0006231727893464267, + "learning_rate": 0.00015934317189773435, + "loss": 0.0, + "step": 998 + }, + { + "epoch": 0.20678948457876217, + "grad_norm": 0.0008572920341975987, + "learning_rate": 0.0001593016004988568, + "loss": 0.0, + "step": 999 + }, + { + "epoch": 0.206996481059822, + "grad_norm": 0.0018585945945233107, + "learning_rate": 0.00015926002909997922, + "loss": 0.0001, + "step": 1000 + }, + { + "epoch": 0.2072034775408818, + "grad_norm": 0.002304868074133992, + "learning_rate": 0.00015921845770110163, + "loss": 0.0009, + "step": 1001 + }, + { + "epoch": 0.20741047402194163, + "grad_norm": 0.004541350062936544, + "learning_rate": 0.00015917688630222407, + "loss": 0.0, + "step": 1002 + }, + { + "epoch": 0.20761747050300144, + "grad_norm": 0.003366716904565692, + "learning_rate": 0.0001591353149033465, + "loss": 0.0011, + "step": 1003 + }, + { + "epoch": 0.20782446698406126, + "grad_norm": 0.002917301142588258, + "learning_rate": 0.00015909374350446892, + "loss": 0.0017, + "step": 1004 + }, + { + "epoch": 0.20803146346512108, + "grad_norm": 0.0004168320447206497, + "learning_rate": 0.00015905217210559136, + "loss": 0.0, + "step": 1005 + }, + { + "epoch": 0.20823845994618093, + "grad_norm": 0.001863375655375421, + "learning_rate": 0.0001590106007067138, + "loss": 0.001, + "step": 1006 + }, + { + "epoch": 0.20844545642724074, + "grad_norm": 0.001271730288863182, + "learning_rate": 0.0001589690293078362, + "loss": 0.0001, + "step": 1007 + }, + { + "epoch": 0.20865245290830056, + "grad_norm": 0.002366506028920412, + "learning_rate": 0.00015892745790895864, + "loss": 0.0009, + "step": 1008 + }, + { + "epoch": 0.20885944938936038, + "grad_norm": 0.0031757184769958258, + "learning_rate": 0.00015888588651008108, + "loss": 0.0001, + "step": 1009 + }, + { + "epoch": 0.2090664458704202, + "grad_norm": 0.0019097458571195602, + "learning_rate": 0.00015884431511120351, + "loss": 0.0, + "step": 1010 + }, + { + "epoch": 0.20927344235148002, + "grad_norm": 0.0008379930513910949, + "learning_rate": 0.00015880274371232592, + "loss": 0.0, + "step": 1011 + }, + { + "epoch": 0.20948043883253983, + "grad_norm": 0.005206478293985128, + "learning_rate": 0.00015876117231344836, + "loss": 0.002, + "step": 1012 + }, + { + "epoch": 0.20968743531359968, + "grad_norm": 0.0008720169425942004, + "learning_rate": 0.0001587196009145708, + "loss": 0.0, + "step": 1013 + }, + { + "epoch": 0.2098944317946595, + "grad_norm": 0.0041591702029109, + "learning_rate": 0.0001586780295156932, + "loss": 0.001, + "step": 1014 + }, + { + "epoch": 0.21010142827571932, + "grad_norm": 0.0004002148343715817, + "learning_rate": 0.00015863645811681565, + "loss": 0.0, + "step": 1015 + }, + { + "epoch": 0.21030842475677913, + "grad_norm": 0.00017360522178933024, + "learning_rate": 0.00015859488671793808, + "loss": 0.0, + "step": 1016 + }, + { + "epoch": 0.21051542123783895, + "grad_norm": 0.004276643507182598, + "learning_rate": 0.0001585533153190605, + "loss": 0.0, + "step": 1017 + }, + { + "epoch": 0.21072241771889877, + "grad_norm": 0.0027010890189558268, + "learning_rate": 0.00015851174392018293, + "loss": 0.0014, + "step": 1018 + }, + { + "epoch": 0.2109294141999586, + "grad_norm": 0.0048659988678991795, + "learning_rate": 0.00015847017252130537, + "loss": 0.0005, + "step": 1019 + }, + { + "epoch": 0.21113641068101843, + "grad_norm": 0.0003106594958808273, + "learning_rate": 0.00015842860112242778, + "loss": 0.0, + "step": 1020 + }, + { + "epoch": 0.21134340716207825, + "grad_norm": 0.0032943710684776306, + "learning_rate": 0.00015838702972355019, + "loss": 0.0012, + "step": 1021 + }, + { + "epoch": 0.21155040364313807, + "grad_norm": 0.0022477346938103437, + "learning_rate": 0.00015834545832467262, + "loss": 0.0002, + "step": 1022 + }, + { + "epoch": 0.2117574001241979, + "grad_norm": 0.0007089116843417287, + "learning_rate": 0.00015830388692579506, + "loss": 0.0, + "step": 1023 + }, + { + "epoch": 0.2119643966052577, + "grad_norm": 0.003983316943049431, + "learning_rate": 0.00015826231552691747, + "loss": 0.0002, + "step": 1024 + }, + { + "epoch": 0.21217139308631752, + "grad_norm": 0.0038651269860565662, + "learning_rate": 0.0001582207441280399, + "loss": 0.0017, + "step": 1025 + }, + { + "epoch": 0.21237838956737737, + "grad_norm": 0.006369102746248245, + "learning_rate": 0.00015817917272916234, + "loss": 0.0002, + "step": 1026 + }, + { + "epoch": 0.21258538604843719, + "grad_norm": 0.002382291480898857, + "learning_rate": 0.00015813760133028478, + "loss": 0.0004, + "step": 1027 + }, + { + "epoch": 0.212792382529497, + "grad_norm": 0.0016512110596522689, + "learning_rate": 0.0001580960299314072, + "loss": 0.0002, + "step": 1028 + }, + { + "epoch": 0.21299937901055682, + "grad_norm": 0.008482804521918297, + "learning_rate": 0.00015805445853252963, + "loss": 0.0002, + "step": 1029 + }, + { + "epoch": 0.21320637549161664, + "grad_norm": 0.005470529198646545, + "learning_rate": 0.00015801288713365207, + "loss": 0.0007, + "step": 1030 + }, + { + "epoch": 0.21341337197267646, + "grad_norm": 0.004424599930644035, + "learning_rate": 0.00015797131573477448, + "loss": 0.0007, + "step": 1031 + }, + { + "epoch": 0.21362036845373628, + "grad_norm": 0.0011165774194523692, + "learning_rate": 0.0001579297443358969, + "loss": 0.0, + "step": 1032 + }, + { + "epoch": 0.21382736493479612, + "grad_norm": 0.0008872200851328671, + "learning_rate": 0.00015788817293701935, + "loss": 0.0, + "step": 1033 + }, + { + "epoch": 0.21403436141585594, + "grad_norm": 0.010052971541881561, + "learning_rate": 0.00015784660153814176, + "loss": 0.0011, + "step": 1034 + }, + { + "epoch": 0.21424135789691576, + "grad_norm": 0.002929918933659792, + "learning_rate": 0.0001578050301392642, + "loss": 0.0015, + "step": 1035 + }, + { + "epoch": 0.21444835437797558, + "grad_norm": 0.008970585651695728, + "learning_rate": 0.00015776345874038663, + "loss": 0.0003, + "step": 1036 + }, + { + "epoch": 0.2146553508590354, + "grad_norm": 0.06671982258558273, + "learning_rate": 0.00015772188734150904, + "loss": 0.0008, + "step": 1037 + }, + { + "epoch": 0.2148623473400952, + "grad_norm": 0.0003781789855565876, + "learning_rate": 0.00015768031594263148, + "loss": 0.0, + "step": 1038 + }, + { + "epoch": 0.21506934382115503, + "grad_norm": 0.0004683547012973577, + "learning_rate": 0.00015763874454375392, + "loss": 0.0, + "step": 1039 + }, + { + "epoch": 0.21527634030221487, + "grad_norm": 0.005777016282081604, + "learning_rate": 0.00015759717314487635, + "loss": 0.0015, + "step": 1040 + }, + { + "epoch": 0.2154833367832747, + "grad_norm": 0.0068507413379848, + "learning_rate": 0.00015755560174599876, + "loss": 0.0002, + "step": 1041 + }, + { + "epoch": 0.2156903332643345, + "grad_norm": 0.0021482266020029783, + "learning_rate": 0.00015751403034712117, + "loss": 0.0004, + "step": 1042 + }, + { + "epoch": 0.21589732974539433, + "grad_norm": 0.003305132733657956, + "learning_rate": 0.0001574724589482436, + "loss": 0.0009, + "step": 1043 + }, + { + "epoch": 0.21610432622645415, + "grad_norm": 0.005324844736605883, + "learning_rate": 0.00015743088754936602, + "loss": 0.0012, + "step": 1044 + }, + { + "epoch": 0.21631132270751396, + "grad_norm": 0.004400115925818682, + "learning_rate": 0.00015738931615048846, + "loss": 0.0002, + "step": 1045 + }, + { + "epoch": 0.21651831918857378, + "grad_norm": 0.0030595625285059214, + "learning_rate": 0.0001573477447516109, + "loss": 0.0005, + "step": 1046 + }, + { + "epoch": 0.21672531566963363, + "grad_norm": 0.004036907572299242, + "learning_rate": 0.00015730617335273333, + "loss": 0.0, + "step": 1047 + }, + { + "epoch": 0.21693231215069345, + "grad_norm": 0.005875944159924984, + "learning_rate": 0.00015726460195385574, + "loss": 0.0004, + "step": 1048 + }, + { + "epoch": 0.21713930863175326, + "grad_norm": 0.003494358854368329, + "learning_rate": 0.00015722303055497818, + "loss": 0.0015, + "step": 1049 + }, + { + "epoch": 0.21734630511281308, + "grad_norm": 0.00041328632505610585, + "learning_rate": 0.00015718145915610062, + "loss": 0.0, + "step": 1050 + }, + { + "epoch": 0.2175533015938729, + "grad_norm": 0.0010599165689200163, + "learning_rate": 0.00015713988775722303, + "loss": 0.0, + "step": 1051 + }, + { + "epoch": 0.21776029807493272, + "grad_norm": 0.00022103896480984986, + "learning_rate": 0.00015709831635834546, + "loss": 0.0, + "step": 1052 + }, + { + "epoch": 0.21796729455599254, + "grad_norm": 0.00018703911337070167, + "learning_rate": 0.0001570567449594679, + "loss": 0.0, + "step": 1053 + }, + { + "epoch": 0.21817429103705238, + "grad_norm": 0.0001905701938085258, + "learning_rate": 0.0001570151735605903, + "loss": 0.0, + "step": 1054 + }, + { + "epoch": 0.2183812875181122, + "grad_norm": 0.01590561680495739, + "learning_rate": 0.00015697360216171275, + "loss": 0.0004, + "step": 1055 + }, + { + "epoch": 0.21858828399917202, + "grad_norm": 0.007658824324607849, + "learning_rate": 0.00015693203076283519, + "loss": 0.0001, + "step": 1056 + }, + { + "epoch": 0.21879528048023184, + "grad_norm": 0.0036896623205393553, + "learning_rate": 0.00015689045936395762, + "loss": 0.0011, + "step": 1057 + }, + { + "epoch": 0.21900227696129165, + "grad_norm": 0.006060061044991016, + "learning_rate": 0.00015684888796508003, + "loss": 0.0001, + "step": 1058 + }, + { + "epoch": 0.21920927344235147, + "grad_norm": 0.010098116472363472, + "learning_rate": 0.00015680731656620247, + "loss": 0.0012, + "step": 1059 + }, + { + "epoch": 0.2194162699234113, + "grad_norm": 0.0016395826824009418, + "learning_rate": 0.0001567657451673249, + "loss": 0.0007, + "step": 1060 + }, + { + "epoch": 0.21962326640447113, + "grad_norm": 0.004565931856632233, + "learning_rate": 0.00015672417376844732, + "loss": 0.0011, + "step": 1061 + }, + { + "epoch": 0.21983026288553095, + "grad_norm": 0.004525905009359121, + "learning_rate": 0.00015668260236956975, + "loss": 0.0002, + "step": 1062 + }, + { + "epoch": 0.22003725936659077, + "grad_norm": 0.0008571099024266005, + "learning_rate": 0.0001566410309706922, + "loss": 0.0, + "step": 1063 + } + ], + "logging_steps": 1, + "max_steps": 4831, + "num_input_tokens_seen": 0, + "num_train_epochs": 1, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 5.0739590310395904e+17, + "train_batch_size": 12, + "trial_name": null, + "trial_params": null +}