{ "best_metric": 1.2894105911254883, "best_model_checkpoint": "miner_id_24/checkpoint-80", "epoch": 0.03106247793858101, "eval_steps": 10, "global_step": 110, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0002823861630780092, "grad_norm": 0.6790897250175476, "learning_rate": 2e-05, "loss": 1.6543, "step": 1 }, { "epoch": 0.0002823861630780092, "eval_loss": 2.1089015007019043, "eval_runtime": 417.1011, "eval_samples_per_second": 1.789, "eval_steps_per_second": 1.789, "step": 1 }, { "epoch": 0.0005647723261560184, "grad_norm": 1.1645948886871338, "learning_rate": 4e-05, "loss": 2.2171, "step": 2 }, { "epoch": 0.0008471584892340275, "grad_norm": 0.4623192846775055, "learning_rate": 6e-05, "loss": 0.9606, "step": 3 }, { "epoch": 0.0011295446523120368, "grad_norm": 0.9102379679679871, "learning_rate": 8e-05, "loss": 2.5406, "step": 4 }, { "epoch": 0.0014119308153900459, "grad_norm": 1.1047648191452026, "learning_rate": 0.0001, "loss": 2.2713, "step": 5 }, { "epoch": 0.001694316978468055, "grad_norm": 0.4400641918182373, "learning_rate": 0.00012, "loss": 2.4362, "step": 6 }, { "epoch": 0.001976703141546064, "grad_norm": 0.7926576137542725, "learning_rate": 0.00014, "loss": 1.75, "step": 7 }, { "epoch": 0.0022590893046240735, "grad_norm": 0.6660397052764893, "learning_rate": 0.00016, "loss": 2.1309, "step": 8 }, { "epoch": 0.0025414754677020824, "grad_norm": 0.6158621311187744, "learning_rate": 0.00018, "loss": 2.4602, "step": 9 }, { "epoch": 0.0028238616307800918, "grad_norm": 0.9061938524246216, "learning_rate": 0.0002, "loss": 1.9124, "step": 10 }, { "epoch": 0.0028238616307800918, "eval_loss": 1.7352350950241089, "eval_runtime": 416.8713, "eval_samples_per_second": 1.79, "eval_steps_per_second": 1.79, "step": 10 }, { "epoch": 0.003106247793858101, "grad_norm": 0.9380114674568176, "learning_rate": 0.0001999979446958366, "loss": 1.5976, "step": 11 }, { "epoch": 0.00338863395693611, "grad_norm": 1.0424156188964844, "learning_rate": 0.00019999177886783194, "loss": 1.9498, "step": 12 }, { "epoch": 0.0036710201200141194, "grad_norm": 0.7817176580429077, "learning_rate": 0.00019998150276943902, "loss": 1.407, "step": 13 }, { "epoch": 0.003953406283092128, "grad_norm": 1.3307476043701172, "learning_rate": 0.000199967116823068, "loss": 1.9723, "step": 14 }, { "epoch": 0.004235792446170138, "grad_norm": 0.9394961595535278, "learning_rate": 0.0001999486216200688, "loss": 1.1677, "step": 15 }, { "epoch": 0.004518178609248147, "grad_norm": 3.413137435913086, "learning_rate": 0.00019992601792070679, "loss": 1.0187, "step": 16 }, { "epoch": 0.004800564772326156, "grad_norm": 3.870213747024536, "learning_rate": 0.00019989930665413147, "loss": 1.4127, "step": 17 }, { "epoch": 0.005082950935404165, "grad_norm": 0.6383346915245056, "learning_rate": 0.00019986848891833845, "loss": 1.338, "step": 18 }, { "epoch": 0.005365337098482174, "grad_norm": 1.3503261804580688, "learning_rate": 0.0001998335659801241, "loss": 0.4498, "step": 19 }, { "epoch": 0.0056477232615601836, "grad_norm": 1.6426069736480713, "learning_rate": 0.00019979453927503364, "loss": 1.7376, "step": 20 }, { "epoch": 0.0056477232615601836, "eval_loss": 1.337227702140808, "eval_runtime": 418.4568, "eval_samples_per_second": 1.783, "eval_steps_per_second": 1.783, "step": 20 }, { "epoch": 0.005930109424638193, "grad_norm": 1.5207312107086182, "learning_rate": 0.00019975141040730207, "loss": 1.0652, "step": 21 }, { "epoch": 0.006212495587716202, "grad_norm": 0.9025612473487854, "learning_rate": 0.0001997041811497882, "loss": 1.9069, "step": 22 }, { "epoch": 0.006494881750794211, "grad_norm": 0.662822961807251, "learning_rate": 0.00019965285344390184, "loss": 2.1184, "step": 23 }, { "epoch": 0.00677726791387222, "grad_norm": 1.2998898029327393, "learning_rate": 0.00019959742939952392, "loss": 1.3041, "step": 24 }, { "epoch": 0.0070596540769502295, "grad_norm": 0.87566077709198, "learning_rate": 0.00019953791129491983, "loss": 1.8995, "step": 25 }, { "epoch": 0.007342040240028239, "grad_norm": 1.7948976755142212, "learning_rate": 0.00019947430157664576, "loss": 1.0098, "step": 26 }, { "epoch": 0.007624426403106248, "grad_norm": 0.8442519307136536, "learning_rate": 0.00019940660285944803, "loss": 1.9566, "step": 27 }, { "epoch": 0.007906812566184257, "grad_norm": 1.40757155418396, "learning_rate": 0.00019933481792615583, "loss": 1.6962, "step": 28 }, { "epoch": 0.008189198729262267, "grad_norm": 1.3247369527816772, "learning_rate": 0.0001992589497275665, "loss": 1.6803, "step": 29 }, { "epoch": 0.008471584892340275, "grad_norm": 2.1888298988342285, "learning_rate": 0.0001991790013823246, "loss": 1.1217, "step": 30 }, { "epoch": 0.008471584892340275, "eval_loss": 1.3264524936676025, "eval_runtime": 422.0198, "eval_samples_per_second": 1.768, "eval_steps_per_second": 1.768, "step": 30 }, { "epoch": 0.008753971055418284, "grad_norm": 0.716585636138916, "learning_rate": 0.00019909497617679348, "loss": 2.0163, "step": 31 }, { "epoch": 0.009036357218496294, "grad_norm": 1.1441899538040161, "learning_rate": 0.0001990068775649202, "loss": 1.9939, "step": 32 }, { "epoch": 0.009318743381574303, "grad_norm": 0.9873970746994019, "learning_rate": 0.00019891470916809362, "loss": 2.0328, "step": 33 }, { "epoch": 0.009601129544652313, "grad_norm": 1.0919432640075684, "learning_rate": 0.00019881847477499557, "loss": 1.4307, "step": 34 }, { "epoch": 0.009883515707730321, "grad_norm": 1.9606109857559204, "learning_rate": 0.00019871817834144504, "loss": 0.8573, "step": 35 }, { "epoch": 0.01016590187080833, "grad_norm": 0.7053022384643555, "learning_rate": 0.0001986138239902355, "loss": 1.7692, "step": 36 }, { "epoch": 0.01044828803388634, "grad_norm": 1.2274115085601807, "learning_rate": 0.0001985054160109657, "loss": 1.3894, "step": 37 }, { "epoch": 0.010730674196964348, "grad_norm": 0.6820394992828369, "learning_rate": 0.00019839295885986296, "loss": 1.4337, "step": 38 }, { "epoch": 0.011013060360042359, "grad_norm": 1.0677088499069214, "learning_rate": 0.0001982764571596004, "loss": 1.6154, "step": 39 }, { "epoch": 0.011295446523120367, "grad_norm": 1.6393396854400635, "learning_rate": 0.00019815591569910654, "loss": 2.0251, "step": 40 }, { "epoch": 0.011295446523120367, "eval_loss": 1.3246878385543823, "eval_runtime": 422.1745, "eval_samples_per_second": 1.767, "eval_steps_per_second": 1.767, "step": 40 }, { "epoch": 0.011577832686198376, "grad_norm": 1.3773140907287598, "learning_rate": 0.00019803133943336874, "loss": 1.7091, "step": 41 }, { "epoch": 0.011860218849276386, "grad_norm": 0.8763225078582764, "learning_rate": 0.0001979027334832293, "loss": 1.7084, "step": 42 }, { "epoch": 0.012142605012354394, "grad_norm": 0.8546161651611328, "learning_rate": 0.00019777010313517518, "loss": 1.7023, "step": 43 }, { "epoch": 0.012424991175432405, "grad_norm": 2.4073739051818848, "learning_rate": 0.00019763345384112043, "loss": 1.1566, "step": 44 }, { "epoch": 0.012707377338510413, "grad_norm": 1.2686834335327148, "learning_rate": 0.00019749279121818235, "loss": 0.3033, "step": 45 }, { "epoch": 0.012989763501588421, "grad_norm": 0.7953050136566162, "learning_rate": 0.00019734812104845047, "loss": 1.5463, "step": 46 }, { "epoch": 0.013272149664666432, "grad_norm": 1.074453592300415, "learning_rate": 0.00019719944927874881, "loss": 1.5074, "step": 47 }, { "epoch": 0.01355453582774444, "grad_norm": 1.1829779148101807, "learning_rate": 0.0001970467820203915, "loss": 1.2887, "step": 48 }, { "epoch": 0.01383692199082245, "grad_norm": 0.631193995475769, "learning_rate": 0.00019689012554893154, "loss": 1.614, "step": 49 }, { "epoch": 0.014119308153900459, "grad_norm": 0.6255627274513245, "learning_rate": 0.00019672948630390294, "loss": 0.7312, "step": 50 }, { "epoch": 0.014119308153900459, "eval_loss": 1.303961157798767, "eval_runtime": 420.1974, "eval_samples_per_second": 1.775, "eval_steps_per_second": 1.775, "step": 50 }, { "epoch": 0.014401694316978467, "grad_norm": 1.264630913734436, "learning_rate": 0.00019656487088855592, "loss": 1.1731, "step": 51 }, { "epoch": 0.014684080480056478, "grad_norm": 1.3137420415878296, "learning_rate": 0.00019639628606958533, "loss": 0.795, "step": 52 }, { "epoch": 0.014966466643134486, "grad_norm": 1.4407997131347656, "learning_rate": 0.0001962237387768529, "loss": 1.3326, "step": 53 }, { "epoch": 0.015248852806212496, "grad_norm": 1.2762479782104492, "learning_rate": 0.00019604723610310194, "loss": 0.7972, "step": 54 }, { "epoch": 0.015531238969290505, "grad_norm": 1.0771076679229736, "learning_rate": 0.00019586678530366606, "loss": 1.3158, "step": 55 }, { "epoch": 0.015813625132368513, "grad_norm": 0.8992990851402283, "learning_rate": 0.00019568239379617088, "loss": 1.4787, "step": 56 }, { "epoch": 0.016096011295446522, "grad_norm": 1.2085480690002441, "learning_rate": 0.00019549406916022905, "loss": 1.3324, "step": 57 }, { "epoch": 0.016378397458524534, "grad_norm": 0.7203757166862488, "learning_rate": 0.00019530181913712872, "loss": 1.548, "step": 58 }, { "epoch": 0.016660783621602542, "grad_norm": 1.4223425388336182, "learning_rate": 0.00019510565162951537, "loss": 1.8365, "step": 59 }, { "epoch": 0.01694316978468055, "grad_norm": 0.46909332275390625, "learning_rate": 0.00019490557470106686, "loss": 1.6222, "step": 60 }, { "epoch": 0.01694316978468055, "eval_loss": 1.3196920156478882, "eval_runtime": 419.9731, "eval_samples_per_second": 1.776, "eval_steps_per_second": 1.776, "step": 60 }, { "epoch": 0.01722555594775856, "grad_norm": 0.7459226846694946, "learning_rate": 0.00019470159657616215, "loss": 1.4639, "step": 61 }, { "epoch": 0.017507942110836568, "grad_norm": 1.4355456829071045, "learning_rate": 0.00019449372563954293, "loss": 1.4994, "step": 62 }, { "epoch": 0.01779032827391458, "grad_norm": 0.7244583964347839, "learning_rate": 0.0001942819704359693, "loss": 0.3904, "step": 63 }, { "epoch": 0.018072714436992588, "grad_norm": 0.9388204216957092, "learning_rate": 0.00019406633966986828, "loss": 1.8876, "step": 64 }, { "epoch": 0.018355100600070597, "grad_norm": 2.272127866744995, "learning_rate": 0.00019384684220497605, "loss": 1.5339, "step": 65 }, { "epoch": 0.018637486763148605, "grad_norm": 0.8985967636108398, "learning_rate": 0.00019362348706397373, "loss": 1.5562, "step": 66 }, { "epoch": 0.018919872926226614, "grad_norm": 1.5650131702423096, "learning_rate": 0.00019339628342811632, "loss": 0.6815, "step": 67 }, { "epoch": 0.019202259089304625, "grad_norm": 1.2985599040985107, "learning_rate": 0.0001931652406368554, "loss": 0.8239, "step": 68 }, { "epoch": 0.019484645252382634, "grad_norm": 1.2683568000793457, "learning_rate": 0.0001929303681874552, "loss": 1.6103, "step": 69 }, { "epoch": 0.019767031415460642, "grad_norm": 1.2072068452835083, "learning_rate": 0.0001926916757346022, "loss": 2.0554, "step": 70 }, { "epoch": 0.019767031415460642, "eval_loss": 1.293800711631775, "eval_runtime": 418.574, "eval_samples_per_second": 1.782, "eval_steps_per_second": 1.782, "step": 70 }, { "epoch": 0.02004941757853865, "grad_norm": 1.6358360052108765, "learning_rate": 0.00019244917309000817, "loss": 1.3294, "step": 71 }, { "epoch": 0.02033180374161666, "grad_norm": 1.2779337167739868, "learning_rate": 0.00019220287022200707, "loss": 1.4654, "step": 72 }, { "epoch": 0.02061418990469467, "grad_norm": 1.0031236410140991, "learning_rate": 0.0001919527772551451, "loss": 1.4763, "step": 73 }, { "epoch": 0.02089657606777268, "grad_norm": 1.0776865482330322, "learning_rate": 0.00019169890446976454, "loss": 1.9505, "step": 74 }, { "epoch": 0.02117896223085069, "grad_norm": 1.0557630062103271, "learning_rate": 0.00019144126230158127, "loss": 1.425, "step": 75 }, { "epoch": 0.021461348393928697, "grad_norm": 1.4582068920135498, "learning_rate": 0.0001911798613412557, "loss": 2.1393, "step": 76 }, { "epoch": 0.021743734557006705, "grad_norm": 0.8082997798919678, "learning_rate": 0.0001909147123339575, "loss": 1.94, "step": 77 }, { "epoch": 0.022026120720084717, "grad_norm": 0.6186602711677551, "learning_rate": 0.0001906458261789238, "loss": 2.4405, "step": 78 }, { "epoch": 0.022308506883162726, "grad_norm": 0.9546780586242676, "learning_rate": 0.00019037321392901136, "loss": 1.4176, "step": 79 }, { "epoch": 0.022590893046240734, "grad_norm": 0.5143202543258667, "learning_rate": 0.0001900968867902419, "loss": 2.0072, "step": 80 }, { "epoch": 0.022590893046240734, "eval_loss": 1.2894105911254883, "eval_runtime": 419.1859, "eval_samples_per_second": 1.78, "eval_steps_per_second": 1.78, "step": 80 }, { "epoch": 0.022873279209318743, "grad_norm": 0.835146963596344, "learning_rate": 0.0001898168561213419, "loss": 1.6153, "step": 81 }, { "epoch": 0.02315566537239675, "grad_norm": 0.6441218852996826, "learning_rate": 0.0001895331334332753, "loss": 1.3453, "step": 82 }, { "epoch": 0.023438051535474763, "grad_norm": 1.0550341606140137, "learning_rate": 0.0001892457303887706, "loss": 0.7759, "step": 83 }, { "epoch": 0.02372043769855277, "grad_norm": 1.1421196460723877, "learning_rate": 0.0001889546588018412, "loss": 1.4067, "step": 84 }, { "epoch": 0.02400282386163078, "grad_norm": 0.683531641960144, "learning_rate": 0.00018865993063730004, "loss": 2.1907, "step": 85 }, { "epoch": 0.02428521002470879, "grad_norm": 1.0629582405090332, "learning_rate": 0.00018836155801026753, "loss": 1.656, "step": 86 }, { "epoch": 0.024567596187786797, "grad_norm": 1.9884428977966309, "learning_rate": 0.0001880595531856738, "loss": 1.3818, "step": 87 }, { "epoch": 0.02484998235086481, "grad_norm": 0.8152889013290405, "learning_rate": 0.00018775392857775432, "loss": 2.4539, "step": 88 }, { "epoch": 0.025132368513942818, "grad_norm": 1.8213200569152832, "learning_rate": 0.00018744469674953956, "loss": 0.8031, "step": 89 }, { "epoch": 0.025414754677020826, "grad_norm": 1.3625519275665283, "learning_rate": 0.00018713187041233896, "loss": 1.5481, "step": 90 }, { "epoch": 0.025414754677020826, "eval_loss": 1.2990045547485352, "eval_runtime": 420.4752, "eval_samples_per_second": 1.774, "eval_steps_per_second": 1.774, "step": 90 }, { "epoch": 0.025697140840098835, "grad_norm": 0.922839343547821, "learning_rate": 0.00018681546242521786, "loss": 2.2968, "step": 91 }, { "epoch": 0.025979527003176843, "grad_norm": 0.9457669258117676, "learning_rate": 0.00018649548579446936, "loss": 1.5159, "step": 92 }, { "epoch": 0.026261913166254855, "grad_norm": 1.3212480545043945, "learning_rate": 0.0001861719536730795, "loss": 1.8084, "step": 93 }, { "epoch": 0.026544299329332863, "grad_norm": 1.1744962930679321, "learning_rate": 0.00018584487936018661, "loss": 2.3983, "step": 94 }, { "epoch": 0.026826685492410872, "grad_norm": 0.9709725379943848, "learning_rate": 0.00018551427630053463, "loss": 1.8034, "step": 95 }, { "epoch": 0.02710907165548888, "grad_norm": 0.878976047039032, "learning_rate": 0.00018518015808392045, "loss": 1.2276, "step": 96 }, { "epoch": 0.02739145781856689, "grad_norm": 1.220984935760498, "learning_rate": 0.00018484253844463526, "loss": 1.0919, "step": 97 }, { "epoch": 0.0276738439816449, "grad_norm": 0.41714727878570557, "learning_rate": 0.00018450143126090015, "loss": 1.0902, "step": 98 }, { "epoch": 0.02795623014472291, "grad_norm": 0.9148246049880981, "learning_rate": 0.00018415685055429533, "loss": 2.1429, "step": 99 }, { "epoch": 0.028238616307800918, "grad_norm": 0.8029168844223022, "learning_rate": 0.00018380881048918405, "loss": 1.5813, "step": 100 }, { "epoch": 0.028238616307800918, "eval_loss": 1.3071253299713135, "eval_runtime": 421.1658, "eval_samples_per_second": 1.771, "eval_steps_per_second": 1.771, "step": 100 }, { "epoch": 0.028521002470878926, "grad_norm": 0.8574671745300293, "learning_rate": 0.00018345732537213027, "loss": 1.9177, "step": 101 }, { "epoch": 0.028803388633956935, "grad_norm": 0.9129525423049927, "learning_rate": 0.00018310240965131041, "loss": 2.4622, "step": 102 }, { "epoch": 0.029085774797034947, "grad_norm": 0.8835414052009583, "learning_rate": 0.00018274407791591966, "loss": 1.051, "step": 103 }, { "epoch": 0.029368160960112955, "grad_norm": 0.508120059967041, "learning_rate": 0.00018238234489557215, "loss": 0.4317, "step": 104 }, { "epoch": 0.029650547123190964, "grad_norm": 1.259400725364685, "learning_rate": 0.0001820172254596956, "loss": 1.9737, "step": 105 }, { "epoch": 0.029932933286268972, "grad_norm": 1.45259428024292, "learning_rate": 0.00018164873461691986, "loss": 0.6959, "step": 106 }, { "epoch": 0.03021531944934698, "grad_norm": 0.8846643567085266, "learning_rate": 0.00018127688751446027, "loss": 1.3487, "step": 107 }, { "epoch": 0.030497705612424993, "grad_norm": 0.7302697896957397, "learning_rate": 0.00018090169943749476, "loss": 1.7445, "step": 108 }, { "epoch": 0.030780091775503, "grad_norm": 1.3288211822509766, "learning_rate": 0.0001805231858085356, "loss": 1.2461, "step": 109 }, { "epoch": 0.03106247793858101, "grad_norm": 0.4046940207481384, "learning_rate": 0.00018014136218679567, "loss": 1.9418, "step": 110 }, { "epoch": 0.03106247793858101, "eval_loss": 1.3085497617721558, "eval_runtime": 418.6203, "eval_samples_per_second": 1.782, "eval_steps_per_second": 1.782, "step": 110 } ], "logging_steps": 1, "max_steps": 500, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 10, "stateful_callbacks": { "EarlyStoppingCallback": { "args": { "early_stopping_patience": 3, "early_stopping_threshold": 0.0 }, "attributes": { "early_stopping_patience_counter": 3 } }, "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 8.206114973810688e+16, "train_batch_size": 1, "trial_name": null, "trial_params": null }