{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.01714383679067375, "eval_steps": 500, "global_step": 100, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0, "grad_norm": 1.4420362710952759, "learning_rate": 4e-05, "loss": 3.3293, "step": 1 }, { "epoch": 0.0, "grad_norm": NaN, "learning_rate": 4e-05, "loss": 3.8207, "step": 2 }, { "epoch": 0.0, "grad_norm": 1.3726439476013184, "learning_rate": 8e-05, "loss": 3.4594, "step": 3 }, { "epoch": 0.0, "grad_norm": 1.2832139730453491, "learning_rate": 0.00012, "loss": 3.297, "step": 4 }, { "epoch": 0.0, "grad_norm": 1.2156065702438354, "learning_rate": 0.00016, "loss": 3.0036, "step": 5 }, { "epoch": 0.0, "grad_norm": 1.196576476097107, "learning_rate": 0.0002, "loss": 2.9488, "step": 6 }, { "epoch": 0.0, "grad_norm": NaN, "learning_rate": 0.0002, "loss": 2.7249, "step": 7 }, { "epoch": 0.0, "grad_norm": 1.1029444932937622, "learning_rate": 0.00019996568291008924, "loss": 2.8809, "step": 8 }, { "epoch": 0.0, "grad_norm": 1.0251176357269287, "learning_rate": 0.00019993136582017846, "loss": 2.728, "step": 9 }, { "epoch": 0.0, "grad_norm": 1.7124559879302979, "learning_rate": 0.0001998970487302677, "loss": 2.6262, "step": 10 }, { "epoch": 0.0, "grad_norm": 1.3183832168579102, "learning_rate": 0.00019986273164035692, "loss": 2.765, "step": 11 }, { "epoch": 0.0, "grad_norm": NaN, "learning_rate": 0.00019986273164035692, "loss": 2.682, "step": 12 }, { "epoch": 0.0, "grad_norm": 1.2592570781707764, "learning_rate": 0.00019982841455044615, "loss": 2.7837, "step": 13 }, { "epoch": 0.0, "grad_norm": 0.9486250877380371, "learning_rate": 0.00019979409746053535, "loss": 2.8407, "step": 14 }, { "epoch": 0.0, "grad_norm": 0.8232390284538269, "learning_rate": 0.00019975978037062458, "loss": 2.6573, "step": 15 }, { "epoch": 0.0, "grad_norm": 1.1916477680206299, "learning_rate": 0.0001997254632807138, "loss": 2.8644, "step": 16 }, { "epoch": 0.0, "grad_norm": 0.6067786812782288, "learning_rate": 0.00019969114619080303, "loss": 2.582, "step": 17 }, { "epoch": 0.0, "grad_norm": 0.9011898040771484, "learning_rate": 0.00019965682910089226, "loss": 2.7302, "step": 18 }, { "epoch": 0.0, "grad_norm": 0.8688616156578064, "learning_rate": 0.00019962251201098149, "loss": 2.3353, "step": 19 }, { "epoch": 0.0, "grad_norm": 0.7956060767173767, "learning_rate": 0.0001995881949210707, "loss": 2.5263, "step": 20 }, { "epoch": 0.0, "grad_norm": NaN, "learning_rate": 0.0001995881949210707, "loss": 2.4818, "step": 21 }, { "epoch": 0.0, "grad_norm": 0.8461011052131653, "learning_rate": 0.00019955387783115994, "loss": 2.542, "step": 22 }, { "epoch": 0.0, "grad_norm": 0.6973548531532288, "learning_rate": 0.00019951956074124917, "loss": 2.4787, "step": 23 }, { "epoch": 0.0, "grad_norm": 0.6046115756034851, "learning_rate": 0.0001994852436513384, "loss": 2.4436, "step": 24 }, { "epoch": 0.0, "grad_norm": 0.4711577594280243, "learning_rate": 0.0001994509265614276, "loss": 2.3658, "step": 25 }, { "epoch": 0.0, "grad_norm": NaN, "learning_rate": 0.0001994509265614276, "loss": 2.4518, "step": 26 }, { "epoch": 0.0, "grad_norm": 0.5012961030006409, "learning_rate": 0.00019941660947151682, "loss": 2.3782, "step": 27 }, { "epoch": 0.0, "grad_norm": 0.6144160628318787, "learning_rate": 0.00019938229238160605, "loss": 2.4487, "step": 28 }, { "epoch": 0.0, "grad_norm": NaN, "learning_rate": 0.00019938229238160605, "loss": 2.3935, "step": 29 }, { "epoch": 0.01, "grad_norm": 0.6648868322372437, "learning_rate": 0.00019934797529169528, "loss": 2.5339, "step": 30 }, { "epoch": 0.01, "grad_norm": 0.5968282222747803, "learning_rate": 0.00019931365820178448, "loss": 2.4238, "step": 31 }, { "epoch": 0.01, "grad_norm": NaN, "learning_rate": 0.00019931365820178448, "loss": 2.4432, "step": 32 }, { "epoch": 0.01, "grad_norm": NaN, "learning_rate": 0.00019931365820178448, "loss": 2.3153, "step": 33 }, { "epoch": 0.01, "grad_norm": 0.9555236101150513, "learning_rate": 0.0001992793411118737, "loss": 2.2495, "step": 34 }, { "epoch": 0.01, "grad_norm": 0.8204368352890015, "learning_rate": 0.00019924502402196293, "loss": 2.401, "step": 35 }, { "epoch": 0.01, "grad_norm": NaN, "learning_rate": 0.00019924502402196293, "loss": 2.31, "step": 36 }, { "epoch": 0.01, "grad_norm": 0.8206217885017395, "learning_rate": 0.00019921070693205216, "loss": 2.3511, "step": 37 }, { "epoch": 0.01, "grad_norm": NaN, "learning_rate": 0.00019921070693205216, "loss": 2.2848, "step": 38 }, { "epoch": 0.01, "grad_norm": 0.5887894630432129, "learning_rate": 0.0001991763898421414, "loss": 2.191, "step": 39 }, { "epoch": 0.01, "grad_norm": 0.5945093035697937, "learning_rate": 0.00019914207275223062, "loss": 2.345, "step": 40 }, { "epoch": 0.01, "grad_norm": NaN, "learning_rate": 0.00019914207275223062, "loss": 2.4744, "step": 41 }, { "epoch": 0.01, "grad_norm": 0.7511911988258362, "learning_rate": 0.00019910775566231984, "loss": 2.3624, "step": 42 }, { "epoch": 0.01, "grad_norm": NaN, "learning_rate": 0.00019910775566231984, "loss": 2.3179, "step": 43 }, { "epoch": 0.01, "grad_norm": 0.8779334425926208, "learning_rate": 0.00019907343857240907, "loss": 2.4634, "step": 44 }, { "epoch": 0.01, "grad_norm": 0.9247270226478577, "learning_rate": 0.0001990391214824983, "loss": 2.2699, "step": 45 }, { "epoch": 0.01, "grad_norm": 0.785952627658844, "learning_rate": 0.0001990048043925875, "loss": 2.2791, "step": 46 }, { "epoch": 0.01, "grad_norm": 1.0079773664474487, "learning_rate": 0.00019897048730267673, "loss": 2.4368, "step": 47 }, { "epoch": 0.01, "grad_norm": 1.1102200746536255, "learning_rate": 0.00019893617021276595, "loss": 2.4482, "step": 48 }, { "epoch": 0.01, "grad_norm": 1.1956760883331299, "learning_rate": 0.00019890185312285518, "loss": 2.4036, "step": 49 }, { "epoch": 0.01, "grad_norm": 0.7190561294555664, "learning_rate": 0.0001988675360329444, "loss": 2.2854, "step": 50 }, { "epoch": 0.01, "grad_norm": NaN, "learning_rate": 0.0001988675360329444, "loss": 2.404, "step": 51 }, { "epoch": 0.01, "grad_norm": 1.0126006603240967, "learning_rate": 0.00019883321894303364, "loss": 2.1978, "step": 52 }, { "epoch": 0.01, "grad_norm": NaN, "learning_rate": 0.00019883321894303364, "loss": 2.3899, "step": 53 }, { "epoch": 0.01, "grad_norm": 1.213659405708313, "learning_rate": 0.00019879890185312286, "loss": 2.0958, "step": 54 }, { "epoch": 0.01, "grad_norm": 1.3466947078704834, "learning_rate": 0.0001987645847632121, "loss": 2.4106, "step": 55 }, { "epoch": 0.01, "grad_norm": 1.2530690431594849, "learning_rate": 0.00019873026767330132, "loss": 2.5628, "step": 56 }, { "epoch": 0.01, "grad_norm": 1.6688309907913208, "learning_rate": 0.00019869595058339055, "loss": 2.233, "step": 57 }, { "epoch": 0.01, "grad_norm": 2.2483224868774414, "learning_rate": 0.00019866163349347977, "loss": 2.2923, "step": 58 }, { "epoch": 0.01, "grad_norm": 2.017726182937622, "learning_rate": 0.000198627316403569, "loss": 2.354, "step": 59 }, { "epoch": 0.01, "grad_norm": NaN, "learning_rate": 0.000198627316403569, "loss": 2.5311, "step": 60 }, { "epoch": 0.01, "grad_norm": 1.810524582862854, "learning_rate": 0.0001985929993136582, "loss": 2.24, "step": 61 }, { "epoch": 0.01, "grad_norm": 2.439291000366211, "learning_rate": 0.00019855868222374743, "loss": 2.8512, "step": 62 }, { "epoch": 0.01, "grad_norm": NaN, "learning_rate": 0.00019855868222374743, "loss": 2.5943, "step": 63 }, { "epoch": 0.01, "grad_norm": 2.2657182216644287, "learning_rate": 0.00019852436513383666, "loss": 2.8789, "step": 64 }, { "epoch": 0.01, "grad_norm": 2.1179282665252686, "learning_rate": 0.00019849004804392589, "loss": 2.6996, "step": 65 }, { "epoch": 0.01, "grad_norm": 2.0068299770355225, "learning_rate": 0.0001984557309540151, "loss": 2.3677, "step": 66 }, { "epoch": 0.01, "grad_norm": NaN, "learning_rate": 0.0001984557309540151, "loss": 2.577, "step": 67 }, { "epoch": 0.01, "grad_norm": NaN, "learning_rate": 0.0001984557309540151, "loss": 2.7072, "step": 68 }, { "epoch": 0.01, "grad_norm": 4.678086280822754, "learning_rate": 0.00019842141386410434, "loss": 2.658, "step": 69 }, { "epoch": 0.01, "grad_norm": 4.056044101715088, "learning_rate": 0.00019838709677419357, "loss": 2.5174, "step": 70 }, { "epoch": 0.01, "grad_norm": 4.784575462341309, "learning_rate": 0.0001983527796842828, "loss": 2.6965, "step": 71 }, { "epoch": 0.01, "grad_norm": 7.531276226043701, "learning_rate": 0.00019831846259437202, "loss": 2.9831, "step": 72 }, { "epoch": 0.01, "grad_norm": 4.478590965270996, "learning_rate": 0.00019828414550446125, "loss": 3.0535, "step": 73 }, { "epoch": 0.01, "grad_norm": NaN, "learning_rate": 0.00019828414550446125, "loss": 2.7639, "step": 74 }, { "epoch": 0.01, "grad_norm": 7.104541778564453, "learning_rate": 0.00019824982841455048, "loss": 3.0957, "step": 75 }, { "epoch": 0.01, "grad_norm": 5.70428991317749, "learning_rate": 0.00019821551132463968, "loss": 3.0107, "step": 76 }, { "epoch": 0.01, "grad_norm": 3.2724623680114746, "learning_rate": 0.0001981811942347289, "loss": 2.9372, "step": 77 }, { "epoch": 0.01, "grad_norm": NaN, "learning_rate": 0.0001981811942347289, "loss": 3.4289, "step": 78 }, { "epoch": 0.01, "grad_norm": NaN, "learning_rate": 0.0001981811942347289, "loss": 3.4832, "step": 79 }, { "epoch": 0.01, "grad_norm": 20.135986328125, "learning_rate": 0.00019814687714481813, "loss": 3.2707, "step": 80 }, { "epoch": 0.01, "grad_norm": 22.686079025268555, "learning_rate": 0.00019811256005490736, "loss": 3.8144, "step": 81 }, { "epoch": 0.01, "grad_norm": NaN, "learning_rate": 0.00019811256005490736, "loss": 4.5631, "step": 82 }, { "epoch": 0.01, "grad_norm": 28.4752197265625, "learning_rate": 0.00019807824296499656, "loss": 4.8363, "step": 83 }, { "epoch": 0.01, "grad_norm": 23.301496505737305, "learning_rate": 0.0001980439258750858, "loss": 5.3773, "step": 84 }, { "epoch": 0.01, "grad_norm": 19.411945343017578, "learning_rate": 0.00019800960878517502, "loss": 5.1105, "step": 85 }, { "epoch": 0.01, "grad_norm": NaN, "learning_rate": 0.00019800960878517502, "loss": 5.8669, "step": 86 }, { "epoch": 0.01, "grad_norm": 35.982608795166016, "learning_rate": 0.00019797529169526424, "loss": 5.4417, "step": 87 }, { "epoch": 0.02, "grad_norm": NaN, "learning_rate": 0.00019797529169526424, "loss": 5.4101, "step": 88 }, { "epoch": 0.02, "grad_norm": 69.57833862304688, "learning_rate": 0.00019794097460535347, "loss": 5.2831, "step": 89 }, { "epoch": 0.02, "grad_norm": 76.1761703491211, "learning_rate": 0.0001979066575154427, "loss": 6.0605, "step": 90 }, { "epoch": 0.02, "grad_norm": NaN, "learning_rate": 0.0001979066575154427, "loss": 6.9964, "step": 91 }, { "epoch": 0.02, "grad_norm": 104.58732604980469, "learning_rate": 0.00019787234042553193, "loss": 7.346, "step": 92 }, { "epoch": 0.02, "grad_norm": 45.56397247314453, "learning_rate": 0.00019783802333562115, "loss": 8.6069, "step": 93 }, { "epoch": 0.02, "grad_norm": NaN, "learning_rate": 0.00019783802333562115, "loss": 9.6728, "step": 94 }, { "epoch": 0.02, "grad_norm": NaN, "learning_rate": 0.00019783802333562115, "loss": 8.1841, "step": 95 }, { "epoch": 0.02, "grad_norm": NaN, "learning_rate": 0.00019783802333562115, "loss": 9.4986, "step": 96 }, { "epoch": 0.02, "grad_norm": 25.033235549926758, "learning_rate": 0.00019780370624571035, "loss": 9.8723, "step": 97 }, { "epoch": 0.02, "grad_norm": 74.37015533447266, "learning_rate": 0.00019776938915579958, "loss": 9.7078, "step": 98 }, { "epoch": 0.02, "grad_norm": 72.31209564208984, "learning_rate": 0.0001977350720658888, "loss": 12.0863, "step": 99 }, { "epoch": 0.02, "grad_norm": 47.87614059448242, "learning_rate": 0.00019770075497597804, "loss": 14.7985, "step": 100 } ], "logging_steps": 1, "max_steps": 5833, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "total_flos": 1.2504616383074304e+16, "train_batch_size": 1, "trial_name": null, "trial_params": null }