|
{ |
|
"best_metric": 1.49005126953125, |
|
"best_model_checkpoint": "miner_id_24/checkpoint-100", |
|
"epoch": 0.03751641343087601, |
|
"eval_steps": 50, |
|
"global_step": 100, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.00037516413430876007, |
|
"grad_norm": 1.1359233856201172, |
|
"learning_rate": 1.6666666666666668e-07, |
|
"loss": 1.7936, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.00037516413430876007, |
|
"eval_loss": 2.12178373336792, |
|
"eval_runtime": 365.7986, |
|
"eval_samples_per_second": 12.275, |
|
"eval_steps_per_second": 1.536, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0007503282686175201, |
|
"grad_norm": 2.703493118286133, |
|
"learning_rate": 3.3333333333333335e-07, |
|
"loss": 1.5961, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.0011254924029262803, |
|
"grad_norm": 2.4857826232910156, |
|
"learning_rate": 5.000000000000001e-07, |
|
"loss": 1.3727, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.0015006565372350403, |
|
"grad_norm": 2.0098776817321777, |
|
"learning_rate": 6.666666666666667e-07, |
|
"loss": 1.1361, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.0018758206715438004, |
|
"grad_norm": 2.490281581878662, |
|
"learning_rate": 8.333333333333333e-07, |
|
"loss": 1.4598, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.0022509848058525606, |
|
"grad_norm": 2.235729694366455, |
|
"learning_rate": 1.0000000000000002e-06, |
|
"loss": 1.2877, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.0026261489401613208, |
|
"grad_norm": 2.2408883571624756, |
|
"learning_rate": 1.1666666666666668e-06, |
|
"loss": 1.2887, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.0030013130744700805, |
|
"grad_norm": 2.0082004070281982, |
|
"learning_rate": 1.3333333333333334e-06, |
|
"loss": 1.2898, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.0033764772087788407, |
|
"grad_norm": 2.294994592666626, |
|
"learning_rate": 1.5e-06, |
|
"loss": 1.2422, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.003751641343087601, |
|
"grad_norm": 2.1499993801116943, |
|
"learning_rate": 1.6666666666666667e-06, |
|
"loss": 1.5298, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.004126805477396361, |
|
"grad_norm": 2.7161669731140137, |
|
"learning_rate": 1.8333333333333333e-06, |
|
"loss": 1.5488, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.004501969611705121, |
|
"grad_norm": 3.114898920059204, |
|
"learning_rate": 2.0000000000000003e-06, |
|
"loss": 1.9435, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.004877133746013881, |
|
"grad_norm": 3.019930839538574, |
|
"learning_rate": 2.166666666666667e-06, |
|
"loss": 1.9317, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.0052522978803226416, |
|
"grad_norm": 2.8244283199310303, |
|
"learning_rate": 2.3333333333333336e-06, |
|
"loss": 1.966, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.005627462014631401, |
|
"grad_norm": 2.8965017795562744, |
|
"learning_rate": 2.5e-06, |
|
"loss": 1.9727, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.006002626148940161, |
|
"grad_norm": 2.852724552154541, |
|
"learning_rate": 2.666666666666667e-06, |
|
"loss": 1.8128, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.006377790283248921, |
|
"grad_norm": 2.867581605911255, |
|
"learning_rate": 2.8333333333333335e-06, |
|
"loss": 2.011, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.006752954417557681, |
|
"grad_norm": 2.7044379711151123, |
|
"learning_rate": 3e-06, |
|
"loss": 1.5948, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.0071281185518664416, |
|
"grad_norm": 2.3761465549468994, |
|
"learning_rate": 3.1666666666666667e-06, |
|
"loss": 1.7501, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.007503282686175202, |
|
"grad_norm": 2.6399307250976562, |
|
"learning_rate": 3.3333333333333333e-06, |
|
"loss": 1.495, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.007878446820483961, |
|
"grad_norm": 3.001854419708252, |
|
"learning_rate": 3.5e-06, |
|
"loss": 1.8462, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.008253610954792722, |
|
"grad_norm": 3.6531243324279785, |
|
"learning_rate": 3.6666666666666666e-06, |
|
"loss": 2.0036, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.008628775089101481, |
|
"grad_norm": 3.283379554748535, |
|
"learning_rate": 3.833333333333334e-06, |
|
"loss": 2.4422, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.009003939223410242, |
|
"grad_norm": 1.9670974016189575, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 1.3793, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.009379103357719002, |
|
"grad_norm": 2.6124558448791504, |
|
"learning_rate": 4.166666666666667e-06, |
|
"loss": 1.7191, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.009754267492027763, |
|
"grad_norm": 3.230952024459839, |
|
"learning_rate": 4.333333333333334e-06, |
|
"loss": 2.2802, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.010129431626336522, |
|
"grad_norm": 2.728804588317871, |
|
"learning_rate": 4.5e-06, |
|
"loss": 1.479, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.010504595760645283, |
|
"grad_norm": 3.584338903427124, |
|
"learning_rate": 4.666666666666667e-06, |
|
"loss": 1.893, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.010879759894954042, |
|
"grad_norm": 2.7680435180664062, |
|
"learning_rate": 4.833333333333333e-06, |
|
"loss": 1.7687, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.011254924029262802, |
|
"grad_norm": 2.725539445877075, |
|
"learning_rate": 5e-06, |
|
"loss": 1.6371, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.011630088163571563, |
|
"grad_norm": 2.4085865020751953, |
|
"learning_rate": 4.997482666353287e-06, |
|
"loss": 1.6609, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.012005252297880322, |
|
"grad_norm": 2.4707136154174805, |
|
"learning_rate": 4.989935734988098e-06, |
|
"loss": 1.6657, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.012380416432189083, |
|
"grad_norm": 2.5936341285705566, |
|
"learning_rate": 4.977374404419838e-06, |
|
"loss": 2.0111, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.012755580566497842, |
|
"grad_norm": 2.875962018966675, |
|
"learning_rate": 4.959823971496575e-06, |
|
"loss": 1.6358, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.013130744700806603, |
|
"grad_norm": 3.3064959049224854, |
|
"learning_rate": 4.937319780454559e-06, |
|
"loss": 1.6326, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.013505908835115363, |
|
"grad_norm": 3.9652161598205566, |
|
"learning_rate": 4.909907151739634e-06, |
|
"loss": 2.1585, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.013881072969424124, |
|
"grad_norm": 2.84902286529541, |
|
"learning_rate": 4.8776412907378845e-06, |
|
"loss": 1.9896, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.014256237103732883, |
|
"grad_norm": 2.7080459594726562, |
|
"learning_rate": 4.8405871765993435e-06, |
|
"loss": 1.9475, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.014631401238041642, |
|
"grad_norm": 2.5884978771209717, |
|
"learning_rate": 4.7988194313786275e-06, |
|
"loss": 1.9678, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.015006565372350403, |
|
"grad_norm": 2.8800384998321533, |
|
"learning_rate": 4.752422169756048e-06, |
|
"loss": 1.9445, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.015381729506659163, |
|
"grad_norm": 2.643221378326416, |
|
"learning_rate": 4.701488829641845e-06, |
|
"loss": 2.6071, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.015756893640967922, |
|
"grad_norm": 2.8539741039276123, |
|
"learning_rate": 4.646121984004666e-06, |
|
"loss": 2.2393, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.016132057775276685, |
|
"grad_norm": 2.8243980407714844, |
|
"learning_rate": 4.586433134303257e-06, |
|
"loss": 2.3024, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.016507221909585444, |
|
"grad_norm": 2.7712883949279785, |
|
"learning_rate": 4.522542485937369e-06, |
|
"loss": 2.2517, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.016882386043894203, |
|
"grad_norm": 2.8007943630218506, |
|
"learning_rate": 4.454578706170075e-06, |
|
"loss": 2.3495, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.017257550178202963, |
|
"grad_norm": 3.1787521839141846, |
|
"learning_rate": 4.382678665009028e-06, |
|
"loss": 2.353, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.017632714312511726, |
|
"grad_norm": 2.6534879207611084, |
|
"learning_rate": 4.3069871595684795e-06, |
|
"loss": 2.4552, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.018007878446820485, |
|
"grad_norm": 3.0257785320281982, |
|
"learning_rate": 4.227656622467162e-06, |
|
"loss": 2.1897, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.018383042581129244, |
|
"grad_norm": 4.02520227432251, |
|
"learning_rate": 4.144846814849282e-06, |
|
"loss": 2.4987, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.018758206715438003, |
|
"grad_norm": 5.180730819702148, |
|
"learning_rate": 4.058724504646834e-06, |
|
"loss": 3.2648, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.018758206715438003, |
|
"eval_loss": 1.617767095565796, |
|
"eval_runtime": 366.2259, |
|
"eval_samples_per_second": 12.26, |
|
"eval_steps_per_second": 1.535, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.019133370849746763, |
|
"grad_norm": 3.0668530464172363, |
|
"learning_rate": 3.969463130731183e-06, |
|
"loss": 1.6115, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.019508534984055526, |
|
"grad_norm": 2.0221545696258545, |
|
"learning_rate": 3.8772424536302565e-06, |
|
"loss": 1.4548, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.019883699118364285, |
|
"grad_norm": 1.181392788887024, |
|
"learning_rate": 3.782248193514766e-06, |
|
"loss": 0.7772, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.020258863252673044, |
|
"grad_norm": 1.638417363166809, |
|
"learning_rate": 3.684671656182497e-06, |
|
"loss": 0.9231, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.020634027386981803, |
|
"grad_norm": 1.695191740989685, |
|
"learning_rate": 3.5847093477938955e-06, |
|
"loss": 1.0114, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.021009191521290566, |
|
"grad_norm": 1.8865028619766235, |
|
"learning_rate": 3.4825625791348093e-06, |
|
"loss": 0.9822, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.021384355655599326, |
|
"grad_norm": 2.3858845233917236, |
|
"learning_rate": 3.3784370602033572e-06, |
|
"loss": 1.7195, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.021759519789908085, |
|
"grad_norm": 1.7190085649490356, |
|
"learning_rate": 3.272542485937369e-06, |
|
"loss": 1.3282, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.022134683924216844, |
|
"grad_norm": 2.187772750854492, |
|
"learning_rate": 3.165092113916688e-06, |
|
"loss": 1.09, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.022509848058525603, |
|
"grad_norm": 1.8837783336639404, |
|
"learning_rate": 3.056302334890786e-06, |
|
"loss": 1.0456, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.022885012192834366, |
|
"grad_norm": 1.810979962348938, |
|
"learning_rate": 2.946392236996592e-06, |
|
"loss": 1.2371, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.023260176327143126, |
|
"grad_norm": 1.6549510955810547, |
|
"learning_rate": 2.835583164544139e-06, |
|
"loss": 1.2869, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.023635340461451885, |
|
"grad_norm": 1.8585015535354614, |
|
"learning_rate": 2.724098272258584e-06, |
|
"loss": 1.2206, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.024010504595760644, |
|
"grad_norm": 2.164670467376709, |
|
"learning_rate": 2.6121620758762877e-06, |
|
"loss": 1.2996, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.024385668730069407, |
|
"grad_norm": 2.0705413818359375, |
|
"learning_rate": 2.5e-06, |
|
"loss": 1.2158, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.024760832864378166, |
|
"grad_norm": 1.8618977069854736, |
|
"learning_rate": 2.3878379241237136e-06, |
|
"loss": 1.3169, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.025135996998686926, |
|
"grad_norm": 2.355955123901367, |
|
"learning_rate": 2.2759017277414165e-06, |
|
"loss": 1.1751, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.025511161132995685, |
|
"grad_norm": 1.8614939451217651, |
|
"learning_rate": 2.1644168354558623e-06, |
|
"loss": 1.3641, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.025886325267304444, |
|
"grad_norm": 2.999551773071289, |
|
"learning_rate": 2.053607763003409e-06, |
|
"loss": 1.2753, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.026261489401613207, |
|
"grad_norm": 1.9945933818817139, |
|
"learning_rate": 1.9436976651092143e-06, |
|
"loss": 1.1216, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.026636653535921966, |
|
"grad_norm": 1.8532859086990356, |
|
"learning_rate": 1.8349078860833125e-06, |
|
"loss": 1.088, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.027011817670230726, |
|
"grad_norm": 2.1395885944366455, |
|
"learning_rate": 1.7274575140626318e-06, |
|
"loss": 1.0015, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.027386981804539485, |
|
"grad_norm": 2.2408030033111572, |
|
"learning_rate": 1.6215629397966432e-06, |
|
"loss": 1.108, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.027762145938848248, |
|
"grad_norm": 2.0609114170074463, |
|
"learning_rate": 1.5174374208651913e-06, |
|
"loss": 1.574, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.028137310073157007, |
|
"grad_norm": 2.0313756465911865, |
|
"learning_rate": 1.415290652206105e-06, |
|
"loss": 1.1664, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.028512474207465766, |
|
"grad_norm": 1.7218246459960938, |
|
"learning_rate": 1.3153283438175036e-06, |
|
"loss": 1.3208, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.028887638341774526, |
|
"grad_norm": 2.472414016723633, |
|
"learning_rate": 1.217751806485235e-06, |
|
"loss": 1.3203, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.029262802476083285, |
|
"grad_norm": 1.6580557823181152, |
|
"learning_rate": 1.122757546369744e-06, |
|
"loss": 1.1601, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.029637966610392048, |
|
"grad_norm": 1.7984857559204102, |
|
"learning_rate": 1.0305368692688175e-06, |
|
"loss": 1.1257, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.030013130744700807, |
|
"grad_norm": 2.6854803562164307, |
|
"learning_rate": 9.412754953531664e-07, |
|
"loss": 1.5246, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.030388294879009566, |
|
"grad_norm": 2.4177401065826416, |
|
"learning_rate": 8.551531851507186e-07, |
|
"loss": 1.5478, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.030763459013318326, |
|
"grad_norm": 2.418226957321167, |
|
"learning_rate": 7.723433775328385e-07, |
|
"loss": 1.5165, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.03113862314762709, |
|
"grad_norm": 2.8937528133392334, |
|
"learning_rate": 6.930128404315214e-07, |
|
"loss": 1.7612, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.031513787281935844, |
|
"grad_norm": 2.7434005737304688, |
|
"learning_rate": 6.17321334990973e-07, |
|
"loss": 1.7735, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.03188895141624461, |
|
"grad_norm": 2.3153278827667236, |
|
"learning_rate": 5.454212938299256e-07, |
|
"loss": 1.4117, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.03226411555055337, |
|
"grad_norm": 2.19724440574646, |
|
"learning_rate": 4.774575140626317e-07, |
|
"loss": 1.5057, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.032639279684862126, |
|
"grad_norm": 4.022721290588379, |
|
"learning_rate": 4.1356686569674344e-07, |
|
"loss": 1.4694, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.03301444381917089, |
|
"grad_norm": 2.6987545490264893, |
|
"learning_rate": 3.538780159953348e-07, |
|
"loss": 1.8611, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.033389607953479644, |
|
"grad_norm": 2.5568947792053223, |
|
"learning_rate": 2.98511170358155e-07, |
|
"loss": 2.0525, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.03376477208778841, |
|
"grad_norm": 2.9902150630950928, |
|
"learning_rate": 2.4757783024395244e-07, |
|
"loss": 1.9918, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.03413993622209717, |
|
"grad_norm": 2.9363043308258057, |
|
"learning_rate": 2.0118056862137358e-07, |
|
"loss": 1.8205, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.034515100356405926, |
|
"grad_norm": 2.565152168273926, |
|
"learning_rate": 1.59412823400657e-07, |
|
"loss": 2.1421, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.03489026449071469, |
|
"grad_norm": 2.2385518550872803, |
|
"learning_rate": 1.223587092621162e-07, |
|
"loss": 2.1784, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.03526542862502345, |
|
"grad_norm": 2.4541680812835693, |
|
"learning_rate": 9.00928482603669e-08, |
|
"loss": 2.1318, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.03564059275933221, |
|
"grad_norm": 3.151038408279419, |
|
"learning_rate": 6.268021954544095e-08, |
|
"loss": 2.1849, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.03601575689364097, |
|
"grad_norm": 2.4328556060791016, |
|
"learning_rate": 4.017602850342584e-08, |
|
"loss": 2.3812, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.036390921027949726, |
|
"grad_norm": 3.570035934448242, |
|
"learning_rate": 2.262559558016325e-08, |
|
"loss": 2.091, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.03676608516225849, |
|
"grad_norm": 3.09126877784729, |
|
"learning_rate": 1.006426501190233e-08, |
|
"loss": 2.4002, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.03714124929656725, |
|
"grad_norm": 3.067747116088867, |
|
"learning_rate": 2.5173336467135266e-09, |
|
"loss": 1.8106, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.03751641343087601, |
|
"grad_norm": 5.066938400268555, |
|
"learning_rate": 0.0, |
|
"loss": 2.3905, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.03751641343087601, |
|
"eval_loss": 1.49005126953125, |
|
"eval_runtime": 366.7943, |
|
"eval_samples_per_second": 12.241, |
|
"eval_steps_per_second": 1.532, |
|
"step": 100 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 100, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 50, |
|
"stateful_callbacks": { |
|
"EarlyStoppingCallback": { |
|
"args": { |
|
"early_stopping_patience": 5, |
|
"early_stopping_threshold": 0.0 |
|
}, |
|
"attributes": { |
|
"early_stopping_patience_counter": 0 |
|
} |
|
}, |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.561311554001961e+17, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|