|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 9.760827407886232, |
|
"eval_steps": 25, |
|
"global_step": 3775, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.06464124111182935, |
|
"grad_norm": 0.4801495671272278, |
|
"learning_rate": 0.0001987561544441565, |
|
"loss": 0.6331, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.06464124111182935, |
|
"eval_loss": 0.614825427532196, |
|
"eval_runtime": 13.3629, |
|
"eval_samples_per_second": 14.518, |
|
"eval_steps_per_second": 1.871, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.1292824822236587, |
|
"grad_norm": 0.5615572333335876, |
|
"learning_rate": 0.0001974604819901529, |
|
"loss": 0.4891, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.1292824822236587, |
|
"eval_loss": 0.6148936748504639, |
|
"eval_runtime": 13.3296, |
|
"eval_samples_per_second": 14.554, |
|
"eval_steps_per_second": 1.876, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.19392372333548805, |
|
"grad_norm": 0.5219907164573669, |
|
"learning_rate": 0.00019616480953614927, |
|
"loss": 0.6005, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.19392372333548805, |
|
"eval_loss": 0.5868657827377319, |
|
"eval_runtime": 13.3229, |
|
"eval_samples_per_second": 14.561, |
|
"eval_steps_per_second": 1.876, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.2585649644473174, |
|
"grad_norm": 0.38672706484794617, |
|
"learning_rate": 0.00019486913708214565, |
|
"loss": 0.4458, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.2585649644473174, |
|
"eval_loss": 0.5840005874633789, |
|
"eval_runtime": 13.3283, |
|
"eval_samples_per_second": 14.556, |
|
"eval_steps_per_second": 1.876, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.32320620555914675, |
|
"grad_norm": 0.28140130639076233, |
|
"learning_rate": 0.000193573464628142, |
|
"loss": 0.5595, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.32320620555914675, |
|
"eval_loss": 0.5389835834503174, |
|
"eval_runtime": 13.3413, |
|
"eval_samples_per_second": 14.541, |
|
"eval_steps_per_second": 1.874, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.3878474466709761, |
|
"grad_norm": 0.4100258946418762, |
|
"learning_rate": 0.00019227779217413838, |
|
"loss": 0.3954, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.3878474466709761, |
|
"eval_loss": 0.5461440086364746, |
|
"eval_runtime": 13.3424, |
|
"eval_samples_per_second": 14.54, |
|
"eval_steps_per_second": 1.874, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.45248868778280543, |
|
"grad_norm": 0.443906307220459, |
|
"learning_rate": 0.00019098211972013476, |
|
"loss": 0.5271, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.45248868778280543, |
|
"eval_loss": 0.511994481086731, |
|
"eval_runtime": 13.3406, |
|
"eval_samples_per_second": 14.542, |
|
"eval_steps_per_second": 1.874, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.5171299288946348, |
|
"grad_norm": 0.45473217964172363, |
|
"learning_rate": 0.00018968644726613114, |
|
"loss": 0.3711, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.5171299288946348, |
|
"eval_loss": 0.5170156359672546, |
|
"eval_runtime": 13.3485, |
|
"eval_samples_per_second": 14.533, |
|
"eval_steps_per_second": 1.873, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.5817711700064642, |
|
"grad_norm": 0.2725232243537903, |
|
"learning_rate": 0.0001883907748121275, |
|
"loss": 0.485, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.5817711700064642, |
|
"eval_loss": 0.48935818672180176, |
|
"eval_runtime": 13.3325, |
|
"eval_samples_per_second": 14.551, |
|
"eval_steps_per_second": 1.875, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.6464124111182935, |
|
"grad_norm": 0.37662985920906067, |
|
"learning_rate": 0.00018709510235812387, |
|
"loss": 0.3509, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.6464124111182935, |
|
"eval_loss": 0.49390536546707153, |
|
"eval_runtime": 13.3404, |
|
"eval_samples_per_second": 14.542, |
|
"eval_steps_per_second": 1.874, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.7110536522301228, |
|
"grad_norm": 0.37969282269477844, |
|
"learning_rate": 0.00018579942990412025, |
|
"loss": 0.4957, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.7110536522301228, |
|
"eval_loss": 0.4695137143135071, |
|
"eval_runtime": 13.3432, |
|
"eval_samples_per_second": 14.539, |
|
"eval_steps_per_second": 1.874, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.7756948933419522, |
|
"grad_norm": 0.47988465428352356, |
|
"learning_rate": 0.0001845037574501166, |
|
"loss": 0.3239, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.7756948933419522, |
|
"eval_loss": 0.4713852107524872, |
|
"eval_runtime": 13.3412, |
|
"eval_samples_per_second": 14.541, |
|
"eval_steps_per_second": 1.874, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.8403361344537815, |
|
"grad_norm": 0.48248979449272156, |
|
"learning_rate": 0.00018320808499611298, |
|
"loss": 0.4371, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.8403361344537815, |
|
"eval_loss": 0.4440518915653229, |
|
"eval_runtime": 13.3429, |
|
"eval_samples_per_second": 14.54, |
|
"eval_steps_per_second": 1.874, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.9049773755656109, |
|
"grad_norm": 0.7585775256156921, |
|
"learning_rate": 0.00018191241254210936, |
|
"loss": 0.2987, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.9049773755656109, |
|
"eval_loss": 0.45601731538772583, |
|
"eval_runtime": 13.3413, |
|
"eval_samples_per_second": 14.541, |
|
"eval_steps_per_second": 1.874, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.9696186166774402, |
|
"grad_norm": 0.35234636068344116, |
|
"learning_rate": 0.00018061674008810574, |
|
"loss": 0.4198, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.9696186166774402, |
|
"eval_loss": 0.4415298402309418, |
|
"eval_runtime": 13.3427, |
|
"eval_samples_per_second": 14.54, |
|
"eval_steps_per_second": 1.874, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 1.0342598577892697, |
|
"grad_norm": 0.4931798577308655, |
|
"learning_rate": 0.0001793210676341021, |
|
"loss": 0.4064, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.0342598577892697, |
|
"eval_loss": 0.42522725462913513, |
|
"eval_runtime": 13.3548, |
|
"eval_samples_per_second": 14.527, |
|
"eval_steps_per_second": 1.872, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.098901098901099, |
|
"grad_norm": 0.5019311308860779, |
|
"learning_rate": 0.00017802539518009847, |
|
"loss": 0.2946, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 1.098901098901099, |
|
"eval_loss": 0.4273912012577057, |
|
"eval_runtime": 13.3417, |
|
"eval_samples_per_second": 14.541, |
|
"eval_steps_per_second": 1.874, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 1.1635423400129283, |
|
"grad_norm": 0.5820428133010864, |
|
"learning_rate": 0.00017672972272609485, |
|
"loss": 0.3152, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.1635423400129283, |
|
"eval_loss": 0.4081675410270691, |
|
"eval_runtime": 13.3377, |
|
"eval_samples_per_second": 14.545, |
|
"eval_steps_per_second": 1.874, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.2281835811247577, |
|
"grad_norm": 0.5499022006988525, |
|
"learning_rate": 0.00017543405027209123, |
|
"loss": 0.2964, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 1.2281835811247577, |
|
"eval_loss": 0.41611921787261963, |
|
"eval_runtime": 13.3496, |
|
"eval_samples_per_second": 14.532, |
|
"eval_steps_per_second": 1.873, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 1.292824822236587, |
|
"grad_norm": 0.5704877376556396, |
|
"learning_rate": 0.00017413837781808759, |
|
"loss": 0.3104, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.292824822236587, |
|
"eval_loss": 0.4043827950954437, |
|
"eval_runtime": 13.3487, |
|
"eval_samples_per_second": 14.533, |
|
"eval_steps_per_second": 1.873, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.3574660633484164, |
|
"grad_norm": 0.410047322511673, |
|
"learning_rate": 0.000172842705364084, |
|
"loss": 0.3023, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 1.3574660633484164, |
|
"eval_loss": 0.40202653408050537, |
|
"eval_runtime": 13.351, |
|
"eval_samples_per_second": 14.531, |
|
"eval_steps_per_second": 1.873, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 1.4221073044602457, |
|
"grad_norm": 0.609170138835907, |
|
"learning_rate": 0.00017154703291008034, |
|
"loss": 0.298, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.4221073044602457, |
|
"eval_loss": 0.3982485234737396, |
|
"eval_runtime": 13.3436, |
|
"eval_samples_per_second": 14.539, |
|
"eval_steps_per_second": 1.874, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.486748545572075, |
|
"grad_norm": 0.5322830677032471, |
|
"learning_rate": 0.0001702513604560767, |
|
"loss": 0.2675, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 1.486748545572075, |
|
"eval_loss": 0.3925786018371582, |
|
"eval_runtime": 13.3528, |
|
"eval_samples_per_second": 14.529, |
|
"eval_steps_per_second": 1.872, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 1.5513897866839044, |
|
"grad_norm": 0.4065910279750824, |
|
"learning_rate": 0.00016895568800207308, |
|
"loss": 0.3029, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.5513897866839044, |
|
"eval_loss": 0.3790633976459503, |
|
"eval_runtime": 13.361, |
|
"eval_samples_per_second": 14.52, |
|
"eval_steps_per_second": 1.871, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.6160310277957337, |
|
"grad_norm": 0.5289133787155151, |
|
"learning_rate": 0.00016766001554806945, |
|
"loss": 0.2705, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 1.6160310277957337, |
|
"eval_loss": 0.3800542652606964, |
|
"eval_runtime": 13.3478, |
|
"eval_samples_per_second": 14.534, |
|
"eval_steps_per_second": 1.873, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 1.680672268907563, |
|
"grad_norm": 0.33712488412857056, |
|
"learning_rate": 0.00016636434309406583, |
|
"loss": 0.2841, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 1.680672268907563, |
|
"eval_loss": 0.3804551362991333, |
|
"eval_runtime": 13.3456, |
|
"eval_samples_per_second": 14.537, |
|
"eval_steps_per_second": 1.873, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 1.7453135100193924, |
|
"grad_norm": 0.31102249026298523, |
|
"learning_rate": 0.0001650686706400622, |
|
"loss": 0.28, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 1.7453135100193924, |
|
"eval_loss": 0.3744698762893677, |
|
"eval_runtime": 13.3601, |
|
"eval_samples_per_second": 14.521, |
|
"eval_steps_per_second": 1.871, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 1.8099547511312217, |
|
"grad_norm": 0.5691291689872742, |
|
"learning_rate": 0.00016377299818605857, |
|
"loss": 0.2652, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 1.8099547511312217, |
|
"eval_loss": 0.3774305582046509, |
|
"eval_runtime": 13.3431, |
|
"eval_samples_per_second": 14.539, |
|
"eval_steps_per_second": 1.874, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 1.874595992243051, |
|
"grad_norm": 0.4610290825366974, |
|
"learning_rate": 0.00016247732573205495, |
|
"loss": 0.2628, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 1.874595992243051, |
|
"eval_loss": 0.37386658787727356, |
|
"eval_runtime": 13.3514, |
|
"eval_samples_per_second": 14.53, |
|
"eval_steps_per_second": 1.872, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 1.9392372333548804, |
|
"grad_norm": 0.4986567497253418, |
|
"learning_rate": 0.00016118165327805132, |
|
"loss": 0.2622, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 1.9392372333548804, |
|
"eval_loss": 0.37703824043273926, |
|
"eval_runtime": 13.3525, |
|
"eval_samples_per_second": 14.529, |
|
"eval_steps_per_second": 1.872, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 2.0038784744667097, |
|
"grad_norm": 0.3808733820915222, |
|
"learning_rate": 0.00015988598082404768, |
|
"loss": 0.2308, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 2.0038784744667097, |
|
"eval_loss": 0.3765748143196106, |
|
"eval_runtime": 13.3632, |
|
"eval_samples_per_second": 14.518, |
|
"eval_steps_per_second": 1.871, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 2.0685197155785393, |
|
"grad_norm": 0.37369033694267273, |
|
"learning_rate": 0.00015859030837004406, |
|
"loss": 0.2115, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 2.0685197155785393, |
|
"eval_loss": 0.3865245580673218, |
|
"eval_runtime": 13.356, |
|
"eval_samples_per_second": 14.525, |
|
"eval_steps_per_second": 1.872, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 2.1331609566903684, |
|
"grad_norm": 0.41955631971359253, |
|
"learning_rate": 0.00015729463591604044, |
|
"loss": 0.1749, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 2.1331609566903684, |
|
"eval_loss": 0.38168737292289734, |
|
"eval_runtime": 13.3562, |
|
"eval_samples_per_second": 14.525, |
|
"eval_steps_per_second": 1.872, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 2.197802197802198, |
|
"grad_norm": 0.39571812748908997, |
|
"learning_rate": 0.0001559989634620368, |
|
"loss": 0.2131, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 2.197802197802198, |
|
"eval_loss": 0.36951744556427, |
|
"eval_runtime": 13.355, |
|
"eval_samples_per_second": 14.526, |
|
"eval_steps_per_second": 1.872, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 2.262443438914027, |
|
"grad_norm": 0.6356109380722046, |
|
"learning_rate": 0.00015470329100803317, |
|
"loss": 0.1882, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 2.262443438914027, |
|
"eval_loss": 0.37231186032295227, |
|
"eval_runtime": 13.3614, |
|
"eval_samples_per_second": 14.519, |
|
"eval_steps_per_second": 1.871, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 2.3270846800258567, |
|
"grad_norm": 0.3918738067150116, |
|
"learning_rate": 0.00015340761855402955, |
|
"loss": 0.2094, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 2.3270846800258567, |
|
"eval_loss": 0.38670334219932556, |
|
"eval_runtime": 13.3624, |
|
"eval_samples_per_second": 14.518, |
|
"eval_steps_per_second": 1.871, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 2.391725921137686, |
|
"grad_norm": 0.6116267442703247, |
|
"learning_rate": 0.00015211194610002593, |
|
"loss": 0.1987, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 2.391725921137686, |
|
"eval_loss": 0.3871639668941498, |
|
"eval_runtime": 13.3618, |
|
"eval_samples_per_second": 14.519, |
|
"eval_steps_per_second": 1.871, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 2.4563671622495153, |
|
"grad_norm": 0.49433189630508423, |
|
"learning_rate": 0.00015081627364602228, |
|
"loss": 0.2305, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 2.4563671622495153, |
|
"eval_loss": 0.366474986076355, |
|
"eval_runtime": 13.37, |
|
"eval_samples_per_second": 14.51, |
|
"eval_steps_per_second": 1.87, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 2.5210084033613445, |
|
"grad_norm": 0.43418142199516296, |
|
"learning_rate": 0.00014952060119201869, |
|
"loss": 0.1755, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 2.5210084033613445, |
|
"eval_loss": 0.3646507263183594, |
|
"eval_runtime": 13.3532, |
|
"eval_samples_per_second": 14.528, |
|
"eval_steps_per_second": 1.872, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 2.585649644473174, |
|
"grad_norm": 0.2716585397720337, |
|
"learning_rate": 0.00014822492873801504, |
|
"loss": 0.2384, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 2.585649644473174, |
|
"eval_loss": 0.35182201862335205, |
|
"eval_runtime": 13.3693, |
|
"eval_samples_per_second": 14.511, |
|
"eval_steps_per_second": 1.87, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 2.650290885585003, |
|
"grad_norm": 0.4708922207355499, |
|
"learning_rate": 0.00014692925628401142, |
|
"loss": 0.1689, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 2.650290885585003, |
|
"eval_loss": 0.3655441999435425, |
|
"eval_runtime": 13.371, |
|
"eval_samples_per_second": 14.509, |
|
"eval_steps_per_second": 1.87, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 2.7149321266968327, |
|
"grad_norm": 0.4261806607246399, |
|
"learning_rate": 0.00014563358383000777, |
|
"loss": 0.2181, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 2.7149321266968327, |
|
"eval_loss": 0.3593634366989136, |
|
"eval_runtime": 13.3548, |
|
"eval_samples_per_second": 14.527, |
|
"eval_steps_per_second": 1.872, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 2.779573367808662, |
|
"grad_norm": 0.5506019592285156, |
|
"learning_rate": 0.00014433791137600415, |
|
"loss": 0.1778, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 2.779573367808662, |
|
"eval_loss": 0.35731247067451477, |
|
"eval_runtime": 13.357, |
|
"eval_samples_per_second": 14.524, |
|
"eval_steps_per_second": 1.872, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 2.8442146089204914, |
|
"grad_norm": 0.5495656132698059, |
|
"learning_rate": 0.00014304223892200053, |
|
"loss": 0.2239, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 2.8442146089204914, |
|
"eval_loss": 0.35565704107284546, |
|
"eval_runtime": 13.3623, |
|
"eval_samples_per_second": 14.518, |
|
"eval_steps_per_second": 1.871, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 2.9088558500323205, |
|
"grad_norm": 0.7420383095741272, |
|
"learning_rate": 0.00014174656646799688, |
|
"loss": 0.1647, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 2.9088558500323205, |
|
"eval_loss": 0.36469554901123047, |
|
"eval_runtime": 13.3662, |
|
"eval_samples_per_second": 14.514, |
|
"eval_steps_per_second": 1.87, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 2.97349709114415, |
|
"grad_norm": 0.2853482961654663, |
|
"learning_rate": 0.00014045089401399326, |
|
"loss": 0.1909, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 2.97349709114415, |
|
"eval_loss": 0.3539246916770935, |
|
"eval_runtime": 13.3625, |
|
"eval_samples_per_second": 14.518, |
|
"eval_steps_per_second": 1.871, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 3.038138332255979, |
|
"grad_norm": 0.4978826940059662, |
|
"learning_rate": 0.00013915522155998964, |
|
"loss": 0.1684, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 3.038138332255979, |
|
"eval_loss": 0.3665294945240021, |
|
"eval_runtime": 13.3706, |
|
"eval_samples_per_second": 14.509, |
|
"eval_steps_per_second": 1.87, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 3.1027795733678087, |
|
"grad_norm": 0.40166956186294556, |
|
"learning_rate": 0.00013785954910598602, |
|
"loss": 0.1398, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 3.1027795733678087, |
|
"eval_loss": 0.37494078278541565, |
|
"eval_runtime": 13.3667, |
|
"eval_samples_per_second": 14.514, |
|
"eval_steps_per_second": 1.87, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 3.167420814479638, |
|
"grad_norm": 0.507607638835907, |
|
"learning_rate": 0.00013656387665198237, |
|
"loss": 0.1326, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 3.167420814479638, |
|
"eval_loss": 0.3721146583557129, |
|
"eval_runtime": 13.3726, |
|
"eval_samples_per_second": 14.507, |
|
"eval_steps_per_second": 1.869, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 3.2320620555914674, |
|
"grad_norm": 0.38957518339157104, |
|
"learning_rate": 0.00013526820419797878, |
|
"loss": 0.1328, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 3.2320620555914674, |
|
"eval_loss": 0.37096062302589417, |
|
"eval_runtime": 13.364, |
|
"eval_samples_per_second": 14.517, |
|
"eval_steps_per_second": 1.871, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 3.2967032967032965, |
|
"grad_norm": 0.7261212468147278, |
|
"learning_rate": 0.00013397253174397513, |
|
"loss": 0.1634, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 3.2967032967032965, |
|
"eval_loss": 0.36481988430023193, |
|
"eval_runtime": 13.3645, |
|
"eval_samples_per_second": 14.516, |
|
"eval_steps_per_second": 1.871, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 3.361344537815126, |
|
"grad_norm": 0.3340177834033966, |
|
"learning_rate": 0.0001326768592899715, |
|
"loss": 0.1403, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 3.361344537815126, |
|
"eval_loss": 0.36701083183288574, |
|
"eval_runtime": 13.3687, |
|
"eval_samples_per_second": 14.512, |
|
"eval_steps_per_second": 1.87, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 3.425985778926955, |
|
"grad_norm": 0.5969800353050232, |
|
"learning_rate": 0.00013138118683596786, |
|
"loss": 0.1549, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 3.425985778926955, |
|
"eval_loss": 0.3541284203529358, |
|
"eval_runtime": 13.3607, |
|
"eval_samples_per_second": 14.52, |
|
"eval_steps_per_second": 1.871, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 3.490627020038785, |
|
"grad_norm": 0.30678942799568176, |
|
"learning_rate": 0.00013008551438196424, |
|
"loss": 0.1492, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 3.490627020038785, |
|
"eval_loss": 0.35420697927474976, |
|
"eval_runtime": 13.3674, |
|
"eval_samples_per_second": 14.513, |
|
"eval_steps_per_second": 1.87, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 3.555268261150614, |
|
"grad_norm": 0.7490562796592712, |
|
"learning_rate": 0.00012878984192796062, |
|
"loss": 0.1514, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 3.555268261150614, |
|
"eval_loss": 0.35634493827819824, |
|
"eval_runtime": 13.3824, |
|
"eval_samples_per_second": 14.497, |
|
"eval_steps_per_second": 1.868, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 3.6199095022624435, |
|
"grad_norm": 0.3132079541683197, |
|
"learning_rate": 0.00012749416947395697, |
|
"loss": 0.1493, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 3.6199095022624435, |
|
"eval_loss": 0.3662796914577484, |
|
"eval_runtime": 13.3685, |
|
"eval_samples_per_second": 14.512, |
|
"eval_steps_per_second": 1.87, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 3.684550743374273, |
|
"grad_norm": 0.4222179651260376, |
|
"learning_rate": 0.00012619849701995338, |
|
"loss": 0.1403, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 3.684550743374273, |
|
"eval_loss": 0.36451268196105957, |
|
"eval_runtime": 13.3664, |
|
"eval_samples_per_second": 14.514, |
|
"eval_steps_per_second": 1.87, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 3.749191984486102, |
|
"grad_norm": 0.4552185833454132, |
|
"learning_rate": 0.00012490282456594973, |
|
"loss": 0.15, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 3.749191984486102, |
|
"eval_loss": 0.3632330298423767, |
|
"eval_runtime": 13.374, |
|
"eval_samples_per_second": 14.506, |
|
"eval_steps_per_second": 1.869, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 3.8138332255979313, |
|
"grad_norm": 0.39688166975975037, |
|
"learning_rate": 0.0001236071521119461, |
|
"loss": 0.1648, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 3.8138332255979313, |
|
"eval_loss": 0.35391995310783386, |
|
"eval_runtime": 13.3842, |
|
"eval_samples_per_second": 14.495, |
|
"eval_steps_per_second": 1.868, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 3.878474466709761, |
|
"grad_norm": 0.2690788209438324, |
|
"learning_rate": 0.00012231147965794246, |
|
"loss": 0.1486, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 3.878474466709761, |
|
"eval_loss": 0.36058586835861206, |
|
"eval_runtime": 13.3733, |
|
"eval_samples_per_second": 14.506, |
|
"eval_steps_per_second": 1.869, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 3.9431157078215904, |
|
"grad_norm": 0.3785063922405243, |
|
"learning_rate": 0.00012101580720393886, |
|
"loss": 0.1523, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 3.9431157078215904, |
|
"eval_loss": 0.3479475975036621, |
|
"eval_runtime": 13.3703, |
|
"eval_samples_per_second": 14.51, |
|
"eval_steps_per_second": 1.87, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 4.0077569489334195, |
|
"grad_norm": 0.39190673828125, |
|
"learning_rate": 0.00011972013474993522, |
|
"loss": 0.1408, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 4.0077569489334195, |
|
"eval_loss": 0.3482886254787445, |
|
"eval_runtime": 13.3744, |
|
"eval_samples_per_second": 14.505, |
|
"eval_steps_per_second": 1.869, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 4.072398190045249, |
|
"grad_norm": 0.33796021342277527, |
|
"learning_rate": 0.0001184244622959316, |
|
"loss": 0.1163, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 4.072398190045249, |
|
"eval_loss": 0.36559316515922546, |
|
"eval_runtime": 13.3799, |
|
"eval_samples_per_second": 14.499, |
|
"eval_steps_per_second": 1.868, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 4.137039431157079, |
|
"grad_norm": 0.843379020690918, |
|
"learning_rate": 0.00011712878984192797, |
|
"loss": 0.1118, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 4.137039431157079, |
|
"eval_loss": 0.36994925141334534, |
|
"eval_runtime": 13.3773, |
|
"eval_samples_per_second": 14.502, |
|
"eval_steps_per_second": 1.869, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 4.201680672268908, |
|
"grad_norm": 0.30512169003486633, |
|
"learning_rate": 0.00011583311738792433, |
|
"loss": 0.1249, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 4.201680672268908, |
|
"eval_loss": 0.36630260944366455, |
|
"eval_runtime": 13.3863, |
|
"eval_samples_per_second": 14.492, |
|
"eval_steps_per_second": 1.868, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 4.266321913380737, |
|
"grad_norm": 0.35797879099845886, |
|
"learning_rate": 0.00011453744493392071, |
|
"loss": 0.1154, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 4.266321913380737, |
|
"eval_loss": 0.36872294545173645, |
|
"eval_runtime": 13.3799, |
|
"eval_samples_per_second": 14.499, |
|
"eval_steps_per_second": 1.868, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 4.330963154492566, |
|
"grad_norm": 0.3784850239753723, |
|
"learning_rate": 0.00011324177247991708, |
|
"loss": 0.1237, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 4.330963154492566, |
|
"eval_loss": 0.3683045506477356, |
|
"eval_runtime": 13.3855, |
|
"eval_samples_per_second": 14.493, |
|
"eval_steps_per_second": 1.868, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 4.395604395604396, |
|
"grad_norm": 0.5391280651092529, |
|
"learning_rate": 0.00011194610002591346, |
|
"loss": 0.1268, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 4.395604395604396, |
|
"eval_loss": 0.366862028837204, |
|
"eval_runtime": 13.3849, |
|
"eval_samples_per_second": 14.494, |
|
"eval_steps_per_second": 1.868, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 4.460245636716225, |
|
"grad_norm": 0.3441055417060852, |
|
"learning_rate": 0.00011065042757190982, |
|
"loss": 0.1154, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 4.460245636716225, |
|
"eval_loss": 0.36934134364128113, |
|
"eval_runtime": 13.3807, |
|
"eval_samples_per_second": 14.498, |
|
"eval_steps_per_second": 1.868, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 4.524886877828054, |
|
"grad_norm": 0.38485729694366455, |
|
"learning_rate": 0.0001093547551179062, |
|
"loss": 0.1196, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 4.524886877828054, |
|
"eval_loss": 0.37078526616096497, |
|
"eval_runtime": 13.3642, |
|
"eval_samples_per_second": 14.516, |
|
"eval_steps_per_second": 1.871, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 4.589528118939883, |
|
"grad_norm": 0.27951779961586, |
|
"learning_rate": 0.00010805908266390257, |
|
"loss": 0.1271, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 4.589528118939883, |
|
"eval_loss": 0.36370983719825745, |
|
"eval_runtime": 13.3815, |
|
"eval_samples_per_second": 14.498, |
|
"eval_steps_per_second": 1.868, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 4.654169360051713, |
|
"grad_norm": 0.41593897342681885, |
|
"learning_rate": 0.00010676341020989895, |
|
"loss": 0.1227, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 4.654169360051713, |
|
"eval_loss": 0.37005409598350525, |
|
"eval_runtime": 13.3667, |
|
"eval_samples_per_second": 14.514, |
|
"eval_steps_per_second": 1.87, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 4.7188106011635425, |
|
"grad_norm": 0.44129708409309387, |
|
"learning_rate": 0.00010546773775589531, |
|
"loss": 0.1177, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 4.7188106011635425, |
|
"eval_loss": 0.36843162775039673, |
|
"eval_runtime": 13.3752, |
|
"eval_samples_per_second": 14.504, |
|
"eval_steps_per_second": 1.869, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 4.783451842275372, |
|
"grad_norm": 0.6084820032119751, |
|
"learning_rate": 0.00010417206530189169, |
|
"loss": 0.1201, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 4.783451842275372, |
|
"eval_loss": 0.37733200192451477, |
|
"eval_runtime": 13.3814, |
|
"eval_samples_per_second": 14.498, |
|
"eval_steps_per_second": 1.868, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 4.848093083387201, |
|
"grad_norm": 0.4541929364204407, |
|
"learning_rate": 0.00010287639284788806, |
|
"loss": 0.1159, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 4.848093083387201, |
|
"eval_loss": 0.3665323853492737, |
|
"eval_runtime": 13.3818, |
|
"eval_samples_per_second": 14.497, |
|
"eval_steps_per_second": 1.868, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 4.912734324499031, |
|
"grad_norm": 0.5170966982841492, |
|
"learning_rate": 0.00010158072039388442, |
|
"loss": 0.1215, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 4.912734324499031, |
|
"eval_loss": 0.37353450059890747, |
|
"eval_runtime": 13.3783, |
|
"eval_samples_per_second": 14.501, |
|
"eval_steps_per_second": 1.869, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 4.97737556561086, |
|
"grad_norm": 0.3494817912578583, |
|
"learning_rate": 0.0001002850479398808, |
|
"loss": 0.1203, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 4.97737556561086, |
|
"eval_loss": 0.3610128164291382, |
|
"eval_runtime": 13.3811, |
|
"eval_samples_per_second": 14.498, |
|
"eval_steps_per_second": 1.868, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 5.042016806722689, |
|
"grad_norm": 0.713656485080719, |
|
"learning_rate": 9.898937548587718e-05, |
|
"loss": 0.1036, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 5.042016806722689, |
|
"eval_loss": 0.3713524341583252, |
|
"eval_runtime": 13.3758, |
|
"eval_samples_per_second": 14.504, |
|
"eval_steps_per_second": 1.869, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 5.106658047834518, |
|
"grad_norm": 0.564629316329956, |
|
"learning_rate": 9.769370303187355e-05, |
|
"loss": 0.0999, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 5.106658047834518, |
|
"eval_loss": 0.39409345388412476, |
|
"eval_runtime": 13.3765, |
|
"eval_samples_per_second": 14.503, |
|
"eval_steps_per_second": 1.869, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 5.171299288946348, |
|
"grad_norm": 0.3473380208015442, |
|
"learning_rate": 9.639803057786993e-05, |
|
"loss": 0.1026, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 5.171299288946348, |
|
"eval_loss": 0.3814166486263275, |
|
"eval_runtime": 13.3768, |
|
"eval_samples_per_second": 14.503, |
|
"eval_steps_per_second": 1.869, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 5.235940530058177, |
|
"grad_norm": 0.2592593729496002, |
|
"learning_rate": 9.510235812386628e-05, |
|
"loss": 0.1072, |
|
"step": 2025 |
|
}, |
|
{ |
|
"epoch": 5.235940530058177, |
|
"eval_loss": 0.3801679313182831, |
|
"eval_runtime": 13.3829, |
|
"eval_samples_per_second": 14.496, |
|
"eval_steps_per_second": 1.868, |
|
"step": 2025 |
|
}, |
|
{ |
|
"epoch": 5.300581771170006, |
|
"grad_norm": 0.5560858249664307, |
|
"learning_rate": 9.380668566986266e-05, |
|
"loss": 0.1047, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 5.300581771170006, |
|
"eval_loss": 0.3803362548351288, |
|
"eval_runtime": 13.3828, |
|
"eval_samples_per_second": 14.496, |
|
"eval_steps_per_second": 1.868, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 5.365223012281835, |
|
"grad_norm": 0.31524232029914856, |
|
"learning_rate": 9.251101321585903e-05, |
|
"loss": 0.101, |
|
"step": 2075 |
|
}, |
|
{ |
|
"epoch": 5.365223012281835, |
|
"eval_loss": 0.3836386799812317, |
|
"eval_runtime": 13.3933, |
|
"eval_samples_per_second": 14.485, |
|
"eval_steps_per_second": 1.867, |
|
"step": 2075 |
|
}, |
|
{ |
|
"epoch": 5.429864253393665, |
|
"grad_norm": 0.4523369073867798, |
|
"learning_rate": 9.12153407618554e-05, |
|
"loss": 0.1018, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 5.429864253393665, |
|
"eval_loss": 0.38159799575805664, |
|
"eval_runtime": 13.3858, |
|
"eval_samples_per_second": 14.493, |
|
"eval_steps_per_second": 1.868, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 5.4945054945054945, |
|
"grad_norm": 0.2764829099178314, |
|
"learning_rate": 8.991966830785178e-05, |
|
"loss": 0.1034, |
|
"step": 2125 |
|
}, |
|
{ |
|
"epoch": 5.4945054945054945, |
|
"eval_loss": 0.3741620182991028, |
|
"eval_runtime": 13.3746, |
|
"eval_samples_per_second": 14.505, |
|
"eval_steps_per_second": 1.869, |
|
"step": 2125 |
|
}, |
|
{ |
|
"epoch": 5.559146735617324, |
|
"grad_norm": 0.3252582848072052, |
|
"learning_rate": 8.862399585384815e-05, |
|
"loss": 0.1047, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 5.559146735617324, |
|
"eval_loss": 0.3792659342288971, |
|
"eval_runtime": 13.3818, |
|
"eval_samples_per_second": 14.497, |
|
"eval_steps_per_second": 1.868, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 5.623787976729153, |
|
"grad_norm": 0.31311649084091187, |
|
"learning_rate": 8.732832339984453e-05, |
|
"loss": 0.1061, |
|
"step": 2175 |
|
}, |
|
{ |
|
"epoch": 5.623787976729153, |
|
"eval_loss": 0.3811538815498352, |
|
"eval_runtime": 13.3767, |
|
"eval_samples_per_second": 14.503, |
|
"eval_steps_per_second": 1.869, |
|
"step": 2175 |
|
}, |
|
{ |
|
"epoch": 5.688429217840983, |
|
"grad_norm": 0.6086262464523315, |
|
"learning_rate": 8.60326509458409e-05, |
|
"loss": 0.1041, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 5.688429217840983, |
|
"eval_loss": 0.36947333812713623, |
|
"eval_runtime": 13.3773, |
|
"eval_samples_per_second": 14.502, |
|
"eval_steps_per_second": 1.869, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 5.753070458952812, |
|
"grad_norm": 0.3273337781429291, |
|
"learning_rate": 8.473697849183728e-05, |
|
"loss": 0.1092, |
|
"step": 2225 |
|
}, |
|
{ |
|
"epoch": 5.753070458952812, |
|
"eval_loss": 0.3786868751049042, |
|
"eval_runtime": 13.371, |
|
"eval_samples_per_second": 14.509, |
|
"eval_steps_per_second": 1.87, |
|
"step": 2225 |
|
}, |
|
{ |
|
"epoch": 5.817711700064641, |
|
"grad_norm": 0.5605947375297546, |
|
"learning_rate": 8.344130603783364e-05, |
|
"loss": 0.1055, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 5.817711700064641, |
|
"eval_loss": 0.37696948647499084, |
|
"eval_runtime": 13.3873, |
|
"eval_samples_per_second": 14.491, |
|
"eval_steps_per_second": 1.867, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 5.882352941176471, |
|
"grad_norm": 0.3829572796821594, |
|
"learning_rate": 8.214563358383001e-05, |
|
"loss": 0.1047, |
|
"step": 2275 |
|
}, |
|
{ |
|
"epoch": 5.882352941176471, |
|
"eval_loss": 0.3816302716732025, |
|
"eval_runtime": 13.375, |
|
"eval_samples_per_second": 14.505, |
|
"eval_steps_per_second": 1.869, |
|
"step": 2275 |
|
}, |
|
{ |
|
"epoch": 5.9469941822883, |
|
"grad_norm": 0.36125028133392334, |
|
"learning_rate": 8.084996112982637e-05, |
|
"loss": 0.1005, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 5.9469941822883, |
|
"eval_loss": 0.37172621488571167, |
|
"eval_runtime": 13.3728, |
|
"eval_samples_per_second": 14.507, |
|
"eval_steps_per_second": 1.869, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 6.011635423400129, |
|
"grad_norm": 0.37611111998558044, |
|
"learning_rate": 7.955428867582275e-05, |
|
"loss": 0.1007, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 6.011635423400129, |
|
"eval_loss": 0.3778633773326874, |
|
"eval_runtime": 13.3736, |
|
"eval_samples_per_second": 14.506, |
|
"eval_steps_per_second": 1.869, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 6.076276664511958, |
|
"grad_norm": 0.23287895321846008, |
|
"learning_rate": 7.825861622181913e-05, |
|
"loss": 0.0853, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 6.076276664511958, |
|
"eval_loss": 0.3875974416732788, |
|
"eval_runtime": 13.3815, |
|
"eval_samples_per_second": 14.498, |
|
"eval_steps_per_second": 1.868, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 6.140917905623788, |
|
"grad_norm": 0.9243698120117188, |
|
"learning_rate": 7.69629437678155e-05, |
|
"loss": 0.0941, |
|
"step": 2375 |
|
}, |
|
{ |
|
"epoch": 6.140917905623788, |
|
"eval_loss": 0.3978659212589264, |
|
"eval_runtime": 13.3745, |
|
"eval_samples_per_second": 14.505, |
|
"eval_steps_per_second": 1.869, |
|
"step": 2375 |
|
}, |
|
{ |
|
"epoch": 6.2055591467356175, |
|
"grad_norm": 0.2637142241001129, |
|
"learning_rate": 7.566727131381188e-05, |
|
"loss": 0.087, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 6.2055591467356175, |
|
"eval_loss": 0.3923758268356323, |
|
"eval_runtime": 13.3843, |
|
"eval_samples_per_second": 14.495, |
|
"eval_steps_per_second": 1.868, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 6.270200387847447, |
|
"grad_norm": 0.48433196544647217, |
|
"learning_rate": 7.437159885980824e-05, |
|
"loss": 0.0969, |
|
"step": 2425 |
|
}, |
|
{ |
|
"epoch": 6.270200387847447, |
|
"eval_loss": 0.38719630241394043, |
|
"eval_runtime": 13.3714, |
|
"eval_samples_per_second": 14.509, |
|
"eval_steps_per_second": 1.87, |
|
"step": 2425 |
|
}, |
|
{ |
|
"epoch": 6.334841628959276, |
|
"grad_norm": 0.3420272767543793, |
|
"learning_rate": 7.307592640580462e-05, |
|
"loss": 0.091, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 6.334841628959276, |
|
"eval_loss": 0.39057913422584534, |
|
"eval_runtime": 13.3771, |
|
"eval_samples_per_second": 14.502, |
|
"eval_steps_per_second": 1.869, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 6.399482870071106, |
|
"grad_norm": 0.48248687386512756, |
|
"learning_rate": 7.178025395180099e-05, |
|
"loss": 0.0983, |
|
"step": 2475 |
|
}, |
|
{ |
|
"epoch": 6.399482870071106, |
|
"eval_loss": 0.38908788561820984, |
|
"eval_runtime": 13.3905, |
|
"eval_samples_per_second": 14.488, |
|
"eval_steps_per_second": 1.867, |
|
"step": 2475 |
|
}, |
|
{ |
|
"epoch": 6.464124111182935, |
|
"grad_norm": 0.2842746078968048, |
|
"learning_rate": 7.048458149779737e-05, |
|
"loss": 0.0908, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 6.464124111182935, |
|
"eval_loss": 0.387993723154068, |
|
"eval_runtime": 13.4025, |
|
"eval_samples_per_second": 14.475, |
|
"eval_steps_per_second": 1.865, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 6.528765352294764, |
|
"grad_norm": 0.31269609928131104, |
|
"learning_rate": 6.918890904379373e-05, |
|
"loss": 0.0942, |
|
"step": 2525 |
|
}, |
|
{ |
|
"epoch": 6.528765352294764, |
|
"eval_loss": 0.38906431198120117, |
|
"eval_runtime": 13.3831, |
|
"eval_samples_per_second": 14.496, |
|
"eval_steps_per_second": 1.868, |
|
"step": 2525 |
|
}, |
|
{ |
|
"epoch": 6.593406593406593, |
|
"grad_norm": 0.2563638687133789, |
|
"learning_rate": 6.78932365897901e-05, |
|
"loss": 0.0906, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 6.593406593406593, |
|
"eval_loss": 0.39094778895378113, |
|
"eval_runtime": 13.3827, |
|
"eval_samples_per_second": 14.496, |
|
"eval_steps_per_second": 1.868, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 6.658047834518423, |
|
"grad_norm": 0.228476420044899, |
|
"learning_rate": 6.659756413578648e-05, |
|
"loss": 0.0969, |
|
"step": 2575 |
|
}, |
|
{ |
|
"epoch": 6.658047834518423, |
|
"eval_loss": 0.3924230933189392, |
|
"eval_runtime": 13.3727, |
|
"eval_samples_per_second": 14.507, |
|
"eval_steps_per_second": 1.869, |
|
"step": 2575 |
|
}, |
|
{ |
|
"epoch": 6.722689075630252, |
|
"grad_norm": 0.19723258912563324, |
|
"learning_rate": 6.530189168178284e-05, |
|
"loss": 0.0922, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 6.722689075630252, |
|
"eval_loss": 0.3915347158908844, |
|
"eval_runtime": 13.3804, |
|
"eval_samples_per_second": 14.499, |
|
"eval_steps_per_second": 1.868, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 6.787330316742081, |
|
"grad_norm": 0.3364139199256897, |
|
"learning_rate": 6.400621922777922e-05, |
|
"loss": 0.0989, |
|
"step": 2625 |
|
}, |
|
{ |
|
"epoch": 6.787330316742081, |
|
"eval_loss": 0.3891298770904541, |
|
"eval_runtime": 13.3771, |
|
"eval_samples_per_second": 14.502, |
|
"eval_steps_per_second": 1.869, |
|
"step": 2625 |
|
}, |
|
{ |
|
"epoch": 6.85197155785391, |
|
"grad_norm": 0.31284722685813904, |
|
"learning_rate": 6.271054677377559e-05, |
|
"loss": 0.0947, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 6.85197155785391, |
|
"eval_loss": 0.3864246904850006, |
|
"eval_runtime": 13.3787, |
|
"eval_samples_per_second": 14.501, |
|
"eval_steps_per_second": 1.869, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 6.91661279896574, |
|
"grad_norm": 0.4304758906364441, |
|
"learning_rate": 6.141487431977197e-05, |
|
"loss": 0.0955, |
|
"step": 2675 |
|
}, |
|
{ |
|
"epoch": 6.91661279896574, |
|
"eval_loss": 0.38957709074020386, |
|
"eval_runtime": 13.3927, |
|
"eval_samples_per_second": 14.485, |
|
"eval_steps_per_second": 1.867, |
|
"step": 2675 |
|
}, |
|
{ |
|
"epoch": 6.98125404007757, |
|
"grad_norm": 0.21785129606723785, |
|
"learning_rate": 6.0119201865768335e-05, |
|
"loss": 0.0957, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 6.98125404007757, |
|
"eval_loss": 0.3870868384838104, |
|
"eval_runtime": 13.3762, |
|
"eval_samples_per_second": 14.503, |
|
"eval_steps_per_second": 1.869, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 7.045895281189399, |
|
"grad_norm": 0.31381458044052124, |
|
"learning_rate": 5.882352941176471e-05, |
|
"loss": 0.0844, |
|
"step": 2725 |
|
}, |
|
{ |
|
"epoch": 7.045895281189399, |
|
"eval_loss": 0.3945547640323639, |
|
"eval_runtime": 13.3799, |
|
"eval_samples_per_second": 14.499, |
|
"eval_steps_per_second": 1.868, |
|
"step": 2725 |
|
}, |
|
{ |
|
"epoch": 7.110536522301228, |
|
"grad_norm": 0.33102476596832275, |
|
"learning_rate": 5.752785695776109e-05, |
|
"loss": 0.0883, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 7.110536522301228, |
|
"eval_loss": 0.4064219295978546, |
|
"eval_runtime": 13.3763, |
|
"eval_samples_per_second": 14.503, |
|
"eval_steps_per_second": 1.869, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 7.175177763413058, |
|
"grad_norm": 0.2840661406517029, |
|
"learning_rate": 5.623218450375746e-05, |
|
"loss": 0.0812, |
|
"step": 2775 |
|
}, |
|
{ |
|
"epoch": 7.175177763413058, |
|
"eval_loss": 0.40618330240249634, |
|
"eval_runtime": 13.391, |
|
"eval_samples_per_second": 14.487, |
|
"eval_steps_per_second": 1.867, |
|
"step": 2775 |
|
}, |
|
{ |
|
"epoch": 7.239819004524887, |
|
"grad_norm": 0.3565811216831207, |
|
"learning_rate": 5.493651204975383e-05, |
|
"loss": 0.0894, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 7.239819004524887, |
|
"eval_loss": 0.4039536416530609, |
|
"eval_runtime": 13.3689, |
|
"eval_samples_per_second": 14.511, |
|
"eval_steps_per_second": 1.87, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 7.304460245636716, |
|
"grad_norm": 0.19193735718727112, |
|
"learning_rate": 5.364083959575019e-05, |
|
"loss": 0.0827, |
|
"step": 2825 |
|
}, |
|
{ |
|
"epoch": 7.304460245636716, |
|
"eval_loss": 0.40351200103759766, |
|
"eval_runtime": 13.3814, |
|
"eval_samples_per_second": 14.498, |
|
"eval_steps_per_second": 1.868, |
|
"step": 2825 |
|
}, |
|
{ |
|
"epoch": 7.369101486748546, |
|
"grad_norm": 0.28010833263397217, |
|
"learning_rate": 5.2345167141746564e-05, |
|
"loss": 0.0937, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 7.369101486748546, |
|
"eval_loss": 0.40530896186828613, |
|
"eval_runtime": 13.371, |
|
"eval_samples_per_second": 14.509, |
|
"eval_steps_per_second": 1.87, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 7.433742727860375, |
|
"grad_norm": 0.1792915165424347, |
|
"learning_rate": 5.1049494687742936e-05, |
|
"loss": 0.0839, |
|
"step": 2875 |
|
}, |
|
{ |
|
"epoch": 7.433742727860375, |
|
"eval_loss": 0.40639927983283997, |
|
"eval_runtime": 13.3697, |
|
"eval_samples_per_second": 14.51, |
|
"eval_steps_per_second": 1.87, |
|
"step": 2875 |
|
}, |
|
{ |
|
"epoch": 7.498383968972204, |
|
"grad_norm": 0.23624753952026367, |
|
"learning_rate": 4.975382223373931e-05, |
|
"loss": 0.0888, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 7.498383968972204, |
|
"eval_loss": 0.4040166437625885, |
|
"eval_runtime": 13.3871, |
|
"eval_samples_per_second": 14.492, |
|
"eval_steps_per_second": 1.867, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 7.563025210084033, |
|
"grad_norm": 0.3217020630836487, |
|
"learning_rate": 4.845814977973568e-05, |
|
"loss": 0.0839, |
|
"step": 2925 |
|
}, |
|
{ |
|
"epoch": 7.563025210084033, |
|
"eval_loss": 0.40058034658432007, |
|
"eval_runtime": 13.3755, |
|
"eval_samples_per_second": 14.504, |
|
"eval_steps_per_second": 1.869, |
|
"step": 2925 |
|
}, |
|
{ |
|
"epoch": 7.6276664511958625, |
|
"grad_norm": 0.19924822449684143, |
|
"learning_rate": 4.7162477325732054e-05, |
|
"loss": 0.0931, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 7.6276664511958625, |
|
"eval_loss": 0.3974081575870514, |
|
"eval_runtime": 13.3797, |
|
"eval_samples_per_second": 14.5, |
|
"eval_steps_per_second": 1.869, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 7.6923076923076925, |
|
"grad_norm": 0.22853316366672516, |
|
"learning_rate": 4.5866804871728434e-05, |
|
"loss": 0.0857, |
|
"step": 2975 |
|
}, |
|
{ |
|
"epoch": 7.6923076923076925, |
|
"eval_loss": 0.3978106379508972, |
|
"eval_runtime": 13.3649, |
|
"eval_samples_per_second": 14.516, |
|
"eval_steps_per_second": 1.871, |
|
"step": 2975 |
|
}, |
|
{ |
|
"epoch": 7.756948933419522, |
|
"grad_norm": 0.3232315480709076, |
|
"learning_rate": 4.4571132417724806e-05, |
|
"loss": 0.0944, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 7.756948933419522, |
|
"eval_loss": 0.39671608805656433, |
|
"eval_runtime": 13.3837, |
|
"eval_samples_per_second": 14.495, |
|
"eval_steps_per_second": 1.868, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 7.821590174531351, |
|
"grad_norm": 0.19052205979824066, |
|
"learning_rate": 4.327545996372117e-05, |
|
"loss": 0.0825, |
|
"step": 3025 |
|
}, |
|
{ |
|
"epoch": 7.821590174531351, |
|
"eval_loss": 0.39850884675979614, |
|
"eval_runtime": 13.3766, |
|
"eval_samples_per_second": 14.503, |
|
"eval_steps_per_second": 1.869, |
|
"step": 3025 |
|
}, |
|
{ |
|
"epoch": 7.886231415643181, |
|
"grad_norm": 0.24069742858409882, |
|
"learning_rate": 4.1979787509717545e-05, |
|
"loss": 0.0926, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 7.886231415643181, |
|
"eval_loss": 0.40030214190483093, |
|
"eval_runtime": 13.3678, |
|
"eval_samples_per_second": 14.512, |
|
"eval_steps_per_second": 1.87, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 7.95087265675501, |
|
"grad_norm": 0.26551541686058044, |
|
"learning_rate": 4.068411505571392e-05, |
|
"loss": 0.0858, |
|
"step": 3075 |
|
}, |
|
{ |
|
"epoch": 7.95087265675501, |
|
"eval_loss": 0.39577826857566833, |
|
"eval_runtime": 13.3848, |
|
"eval_samples_per_second": 14.494, |
|
"eval_steps_per_second": 1.868, |
|
"step": 3075 |
|
}, |
|
{ |
|
"epoch": 8.015513897866839, |
|
"grad_norm": 0.23020051419734955, |
|
"learning_rate": 3.938844260171029e-05, |
|
"loss": 0.0841, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 8.015513897866839, |
|
"eval_loss": 0.4009804129600525, |
|
"eval_runtime": 13.3828, |
|
"eval_samples_per_second": 14.496, |
|
"eval_steps_per_second": 1.868, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 8.080155138978668, |
|
"grad_norm": 0.23779013752937317, |
|
"learning_rate": 3.809277014770666e-05, |
|
"loss": 0.0795, |
|
"step": 3125 |
|
}, |
|
{ |
|
"epoch": 8.080155138978668, |
|
"eval_loss": 0.40682119131088257, |
|
"eval_runtime": 13.3724, |
|
"eval_samples_per_second": 14.507, |
|
"eval_steps_per_second": 1.87, |
|
"step": 3125 |
|
}, |
|
{ |
|
"epoch": 8.144796380090497, |
|
"grad_norm": 0.30244529247283936, |
|
"learning_rate": 3.679709769370303e-05, |
|
"loss": 0.083, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 8.144796380090497, |
|
"eval_loss": 0.40890321135520935, |
|
"eval_runtime": 13.3736, |
|
"eval_samples_per_second": 14.506, |
|
"eval_steps_per_second": 1.869, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 8.209437621202326, |
|
"grad_norm": 0.2585161328315735, |
|
"learning_rate": 3.55014252396994e-05, |
|
"loss": 0.078, |
|
"step": 3175 |
|
}, |
|
{ |
|
"epoch": 8.209437621202326, |
|
"eval_loss": 0.4109211564064026, |
|
"eval_runtime": 13.3935, |
|
"eval_samples_per_second": 14.485, |
|
"eval_steps_per_second": 1.867, |
|
"step": 3175 |
|
}, |
|
{ |
|
"epoch": 8.274078862314157, |
|
"grad_norm": 0.28934046626091003, |
|
"learning_rate": 3.420575278569578e-05, |
|
"loss": 0.0841, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 8.274078862314157, |
|
"eval_loss": 0.41459396481513977, |
|
"eval_runtime": 13.3855, |
|
"eval_samples_per_second": 14.493, |
|
"eval_steps_per_second": 1.868, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 8.338720103425986, |
|
"grad_norm": 0.23310010135173798, |
|
"learning_rate": 3.291008033169215e-05, |
|
"loss": 0.0775, |
|
"step": 3225 |
|
}, |
|
{ |
|
"epoch": 8.338720103425986, |
|
"eval_loss": 0.41512471437454224, |
|
"eval_runtime": 13.3818, |
|
"eval_samples_per_second": 14.497, |
|
"eval_steps_per_second": 1.868, |
|
"step": 3225 |
|
}, |
|
{ |
|
"epoch": 8.403361344537815, |
|
"grad_norm": 0.22439134120941162, |
|
"learning_rate": 3.1614407877688526e-05, |
|
"loss": 0.0837, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 8.403361344537815, |
|
"eval_loss": 0.41561707854270935, |
|
"eval_runtime": 13.3826, |
|
"eval_samples_per_second": 14.496, |
|
"eval_steps_per_second": 1.868, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 8.468002585649645, |
|
"grad_norm": 0.2346993237733841, |
|
"learning_rate": 3.0318735423684895e-05, |
|
"loss": 0.0799, |
|
"step": 3275 |
|
}, |
|
{ |
|
"epoch": 8.468002585649645, |
|
"eval_loss": 0.4051866829395294, |
|
"eval_runtime": 13.3812, |
|
"eval_samples_per_second": 14.498, |
|
"eval_steps_per_second": 1.868, |
|
"step": 3275 |
|
}, |
|
{ |
|
"epoch": 8.532643826761474, |
|
"grad_norm": 0.22098496556282043, |
|
"learning_rate": 2.9023062969681264e-05, |
|
"loss": 0.0849, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 8.532643826761474, |
|
"eval_loss": 0.40784117579460144, |
|
"eval_runtime": 13.3808, |
|
"eval_samples_per_second": 14.498, |
|
"eval_steps_per_second": 1.868, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 8.597285067873303, |
|
"grad_norm": 0.2729458808898926, |
|
"learning_rate": 2.7727390515677637e-05, |
|
"loss": 0.0803, |
|
"step": 3325 |
|
}, |
|
{ |
|
"epoch": 8.597285067873303, |
|
"eval_loss": 0.4111315608024597, |
|
"eval_runtime": 13.3646, |
|
"eval_samples_per_second": 14.516, |
|
"eval_steps_per_second": 1.871, |
|
"step": 3325 |
|
}, |
|
{ |
|
"epoch": 8.661926308985132, |
|
"grad_norm": 0.19786176085472107, |
|
"learning_rate": 2.643171806167401e-05, |
|
"loss": 0.0849, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 8.661926308985132, |
|
"eval_loss": 0.415315181016922, |
|
"eval_runtime": 13.3797, |
|
"eval_samples_per_second": 14.5, |
|
"eval_steps_per_second": 1.868, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 8.726567550096961, |
|
"grad_norm": 0.26568812131881714, |
|
"learning_rate": 2.5136045607670382e-05, |
|
"loss": 0.0806, |
|
"step": 3375 |
|
}, |
|
{ |
|
"epoch": 8.726567550096961, |
|
"eval_loss": 0.4111620783805847, |
|
"eval_runtime": 13.3769, |
|
"eval_samples_per_second": 14.503, |
|
"eval_steps_per_second": 1.869, |
|
"step": 3375 |
|
}, |
|
{ |
|
"epoch": 8.791208791208792, |
|
"grad_norm": 0.2054780125617981, |
|
"learning_rate": 2.3840373153666755e-05, |
|
"loss": 0.0881, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 8.791208791208792, |
|
"eval_loss": 0.4074566960334778, |
|
"eval_runtime": 13.384, |
|
"eval_samples_per_second": 14.495, |
|
"eval_steps_per_second": 1.868, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 8.855850032320621, |
|
"grad_norm": 0.29870903491973877, |
|
"learning_rate": 2.2544700699663127e-05, |
|
"loss": 0.0806, |
|
"step": 3425 |
|
}, |
|
{ |
|
"epoch": 8.855850032320621, |
|
"eval_loss": 0.40841713547706604, |
|
"eval_runtime": 13.3768, |
|
"eval_samples_per_second": 14.503, |
|
"eval_steps_per_second": 1.869, |
|
"step": 3425 |
|
}, |
|
{ |
|
"epoch": 8.92049127343245, |
|
"grad_norm": 0.21767568588256836, |
|
"learning_rate": 2.12490282456595e-05, |
|
"loss": 0.085, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 8.92049127343245, |
|
"eval_loss": 0.4094007909297943, |
|
"eval_runtime": 13.3757, |
|
"eval_samples_per_second": 14.504, |
|
"eval_steps_per_second": 1.869, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 8.98513251454428, |
|
"grad_norm": 0.20575569570064545, |
|
"learning_rate": 1.995335579165587e-05, |
|
"loss": 0.0873, |
|
"step": 3475 |
|
}, |
|
{ |
|
"epoch": 8.98513251454428, |
|
"eval_loss": 0.40901434421539307, |
|
"eval_runtime": 13.3813, |
|
"eval_samples_per_second": 14.498, |
|
"eval_steps_per_second": 1.868, |
|
"step": 3475 |
|
}, |
|
{ |
|
"epoch": 9.049773755656108, |
|
"grad_norm": 0.23535218834877014, |
|
"learning_rate": 1.8657683337652242e-05, |
|
"loss": 0.0744, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 9.049773755656108, |
|
"eval_loss": 0.41001009941101074, |
|
"eval_runtime": 13.3767, |
|
"eval_samples_per_second": 14.503, |
|
"eval_steps_per_second": 1.869, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 9.114414996767938, |
|
"grad_norm": 0.22550158202648163, |
|
"learning_rate": 1.7362010883648614e-05, |
|
"loss": 0.0815, |
|
"step": 3525 |
|
}, |
|
{ |
|
"epoch": 9.114414996767938, |
|
"eval_loss": 0.4145725667476654, |
|
"eval_runtime": 13.372, |
|
"eval_samples_per_second": 14.508, |
|
"eval_steps_per_second": 1.87, |
|
"step": 3525 |
|
}, |
|
{ |
|
"epoch": 9.179056237879767, |
|
"grad_norm": 0.21262843906879425, |
|
"learning_rate": 1.6066338429644987e-05, |
|
"loss": 0.0732, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 9.179056237879767, |
|
"eval_loss": 0.4180243909358978, |
|
"eval_runtime": 13.3877, |
|
"eval_samples_per_second": 14.491, |
|
"eval_steps_per_second": 1.867, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 9.243697478991596, |
|
"grad_norm": 0.22902533411979675, |
|
"learning_rate": 1.477066597564136e-05, |
|
"loss": 0.0821, |
|
"step": 3575 |
|
}, |
|
{ |
|
"epoch": 9.243697478991596, |
|
"eval_loss": 0.4216705858707428, |
|
"eval_runtime": 13.3663, |
|
"eval_samples_per_second": 14.514, |
|
"eval_steps_per_second": 1.87, |
|
"step": 3575 |
|
}, |
|
{ |
|
"epoch": 9.308338720103427, |
|
"grad_norm": 0.2216034233570099, |
|
"learning_rate": 1.347499352163773e-05, |
|
"loss": 0.0718, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 9.308338720103427, |
|
"eval_loss": 0.4226977825164795, |
|
"eval_runtime": 13.3768, |
|
"eval_samples_per_second": 14.503, |
|
"eval_steps_per_second": 1.869, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 9.372979961215256, |
|
"grad_norm": 0.27297767996788025, |
|
"learning_rate": 1.2179321067634103e-05, |
|
"loss": 0.0823, |
|
"step": 3625 |
|
}, |
|
{ |
|
"epoch": 9.372979961215256, |
|
"eval_loss": 0.42199140787124634, |
|
"eval_runtime": 13.3655, |
|
"eval_samples_per_second": 14.515, |
|
"eval_steps_per_second": 1.87, |
|
"step": 3625 |
|
}, |
|
{ |
|
"epoch": 9.437621202327085, |
|
"grad_norm": 0.2388414740562439, |
|
"learning_rate": 1.0883648613630474e-05, |
|
"loss": 0.073, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 9.437621202327085, |
|
"eval_loss": 0.42137593030929565, |
|
"eval_runtime": 13.3793, |
|
"eval_samples_per_second": 14.5, |
|
"eval_steps_per_second": 1.869, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 9.502262443438914, |
|
"grad_norm": 0.25000351667404175, |
|
"learning_rate": 9.587976159626847e-06, |
|
"loss": 0.0843, |
|
"step": 3675 |
|
}, |
|
{ |
|
"epoch": 9.502262443438914, |
|
"eval_loss": 0.42024168372154236, |
|
"eval_runtime": 13.3731, |
|
"eval_samples_per_second": 14.507, |
|
"eval_steps_per_second": 1.869, |
|
"step": 3675 |
|
}, |
|
{ |
|
"epoch": 9.566903684550743, |
|
"grad_norm": 0.24509260058403015, |
|
"learning_rate": 8.29230370562322e-06, |
|
"loss": 0.0753, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 9.566903684550743, |
|
"eval_loss": 0.41970470547676086, |
|
"eval_runtime": 13.378, |
|
"eval_samples_per_second": 14.501, |
|
"eval_steps_per_second": 1.869, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 9.631544925662572, |
|
"grad_norm": 0.28298047184944153, |
|
"learning_rate": 6.996631251619591e-06, |
|
"loss": 0.0835, |
|
"step": 3725 |
|
}, |
|
{ |
|
"epoch": 9.631544925662572, |
|
"eval_loss": 0.4190382957458496, |
|
"eval_runtime": 13.3719, |
|
"eval_samples_per_second": 14.508, |
|
"eval_steps_per_second": 1.87, |
|
"step": 3725 |
|
}, |
|
{ |
|
"epoch": 9.696186166774401, |
|
"grad_norm": 0.28533321619033813, |
|
"learning_rate": 5.700958797615963e-06, |
|
"loss": 0.0735, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 9.696186166774401, |
|
"eval_loss": 0.4197899401187897, |
|
"eval_runtime": 13.3702, |
|
"eval_samples_per_second": 14.51, |
|
"eval_steps_per_second": 1.87, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 9.760827407886232, |
|
"grad_norm": 0.2747386693954468, |
|
"learning_rate": 4.4052863436123355e-06, |
|
"loss": 0.0846, |
|
"step": 3775 |
|
}, |
|
{ |
|
"epoch": 9.760827407886232, |
|
"eval_loss": 0.420250803232193, |
|
"eval_runtime": 13.3816, |
|
"eval_samples_per_second": 14.498, |
|
"eval_steps_per_second": 1.868, |
|
"step": 3775 |
|
} |
|
], |
|
"logging_steps": 25, |
|
"max_steps": 3860, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 10, |
|
"save_steps": 25, |
|
"total_flos": 1.232289499803648e+16, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|