|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.20630892698727074, |
|
"eval_steps": 500, |
|
"global_step": 10000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0, |
|
"grad_norm": 12.003094673156738, |
|
"learning_rate": 0.00001, |
|
"loss": 2.4799, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0, |
|
"grad_norm": 2.0890064239501953, |
|
"learning_rate": 0.00002, |
|
"loss": 0.6329, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 2.033069133758545, |
|
"learning_rate": 0.00003, |
|
"loss": 0.5035, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 2.252149820327759, |
|
"learning_rate": 0.00004, |
|
"loss": 0.4943, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 2.118387460708618, |
|
"learning_rate": 0.00005, |
|
"loss": 0.5105, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 1.954121708869934, |
|
"learning_rate": 0.000049473684210526315, |
|
"loss": 0.5008, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 2.036804676055908, |
|
"learning_rate": 0.000048947368421052635, |
|
"loss": 0.4793, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.824407696723938, |
|
"learning_rate": 0.00004842105263157895, |
|
"loss": 0.469, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.724145531654358, |
|
"learning_rate": 0.00004789473684210526, |
|
"loss": 0.4633, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.3427892923355103, |
|
"learning_rate": 0.00004736842105263158, |
|
"loss": 0.4614, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.0959577560424805, |
|
"learning_rate": 0.0000468421052631579, |
|
"loss": 0.4605, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.5479639768600464, |
|
"learning_rate": 0.000046315789473684214, |
|
"loss": 0.46, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 1.535975456237793, |
|
"learning_rate": 0.000045789473684210527, |
|
"loss": 0.4252, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 1.5321663618087769, |
|
"learning_rate": 0.000045263157894736846, |
|
"loss": 0.4369, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 1.5459933280944824, |
|
"learning_rate": 0.00004473684210526316, |
|
"loss": 0.4276, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 1.7711249589920044, |
|
"learning_rate": 0.00004421052631578947, |
|
"loss": 0.4474, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 1.376570701599121, |
|
"learning_rate": 0.00004368421052631579, |
|
"loss": 0.4064, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 1.74457585811615, |
|
"learning_rate": 0.000043157894736842105, |
|
"loss": 0.4093, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 1.4628654718399048, |
|
"learning_rate": 0.000042631578947368425, |
|
"loss": 0.4236, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 1.0637551546096802, |
|
"learning_rate": 0.00004210526315789474, |
|
"loss": 0.4137, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 1.385728359222412, |
|
"learning_rate": 0.00004157894736842106, |
|
"loss": 0.4071, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 1.3536651134490967, |
|
"learning_rate": 0.00004105263157894737, |
|
"loss": 0.4215, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 1.5586481094360352, |
|
"learning_rate": 0.000040526315789473684, |
|
"loss": 0.4041, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 1.0527902841567993, |
|
"learning_rate": 0.00004, |
|
"loss": 0.3984, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 1.5508724451065063, |
|
"learning_rate": 0.000039473684210526316, |
|
"loss": 0.406, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 1.7268626689910889, |
|
"learning_rate": 0.00003894736842105263, |
|
"loss": 0.3978, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 1.259959101676941, |
|
"learning_rate": 0.00003842105263157895, |
|
"loss": 0.4111, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 1.5187910795211792, |
|
"learning_rate": 0.00003789473684210527, |
|
"loss": 0.3862, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 1.4480005502700806, |
|
"learning_rate": 0.00003736842105263158, |
|
"loss": 0.4021, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 1.2096664905548096, |
|
"learning_rate": 0.000036842105263157895, |
|
"loss": 0.4092, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 1.0258468389511108, |
|
"learning_rate": 0.000036315789473684214, |
|
"loss": 0.3829, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 1.336549162864685, |
|
"learning_rate": 0.00003578947368421053, |
|
"loss": 0.4082, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 1.1610171794891357, |
|
"learning_rate": 0.00003526315789473684, |
|
"loss": 0.4216, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 1.1225395202636719, |
|
"learning_rate": 0.00003473684210526316, |
|
"loss": 0.3866, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 1.2079445123672485, |
|
"learning_rate": 0.00003421052631578947, |
|
"loss": 0.3827, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.8150144815444946, |
|
"learning_rate": 0.00003368421052631579, |
|
"loss": 0.3833, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 1.1681067943572998, |
|
"learning_rate": 0.000033157894736842106, |
|
"loss": 0.3953, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 1.1494988203048706, |
|
"learning_rate": 0.000032631578947368426, |
|
"loss": 0.3785, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.8699579238891602, |
|
"learning_rate": 0.00003210526315789474, |
|
"loss": 0.3802, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 1.560229778289795, |
|
"learning_rate": 0.00003157894736842105, |
|
"loss": 0.3837, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 1.1917518377304077, |
|
"learning_rate": 0.00003105263157894737, |
|
"loss": 0.3775, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 1.050404667854309, |
|
"learning_rate": 0.000030526315789473684, |
|
"loss": 0.3781, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.8547882437705994, |
|
"learning_rate": 0.00003, |
|
"loss": 0.3629, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 1.0195717811584473, |
|
"learning_rate": 0.000029473684210526314, |
|
"loss": 0.3644, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 1.0551220178604126, |
|
"learning_rate": 0.000028947368421052634, |
|
"loss": 0.3774, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.9689894318580627, |
|
"learning_rate": 0.00002842105263157895, |
|
"loss": 0.3384, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 1.3785911798477173, |
|
"learning_rate": 0.000027894736842105263, |
|
"loss": 0.3725, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 1.1647924184799194, |
|
"learning_rate": 0.000027368421052631583, |
|
"loss": 0.3716, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 1.0317350625991821, |
|
"learning_rate": 0.000026842105263157896, |
|
"loss": 0.3454, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 1.2156306505203247, |
|
"learning_rate": 0.000026315789473684212, |
|
"loss": 0.3414, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 1.0330926179885864, |
|
"learning_rate": 0.00002578947368421053, |
|
"loss": 0.3658, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 1.1999331712722778, |
|
"learning_rate": 0.000025263157894736845, |
|
"loss": 0.3659, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.8645588159561157, |
|
"learning_rate": 0.000024736842105263158, |
|
"loss": 0.3567, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 1.3966509103775024, |
|
"learning_rate": 0.000024210526315789474, |
|
"loss": 0.3593, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.9758176207542419, |
|
"learning_rate": 0.00002368421052631579, |
|
"loss": 0.3491, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 1.0079020261764526, |
|
"learning_rate": 0.000023157894736842107, |
|
"loss": 0.3403, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 1.0078275203704834, |
|
"learning_rate": 0.000022631578947368423, |
|
"loss": 0.3396, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 1.2329459190368652, |
|
"learning_rate": 0.000022105263157894736, |
|
"loss": 0.3323, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 1.5663444995880127, |
|
"learning_rate": 0.000021578947368421053, |
|
"loss": 0.332, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 1.104218602180481, |
|
"learning_rate": 0.00002105263157894737, |
|
"loss": 0.346, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 1.1730228662490845, |
|
"learning_rate": 0.000020526315789473685, |
|
"loss": 0.3485, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 1.2409980297088623, |
|
"learning_rate": 0.00002, |
|
"loss": 0.3539, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.933708906173706, |
|
"learning_rate": 0.000019473684210526315, |
|
"loss": 0.3177, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 1.0058486461639404, |
|
"learning_rate": 0.000018947368421052634, |
|
"loss": 0.3394, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 1.2994540929794312, |
|
"learning_rate": 0.000018421052631578947, |
|
"loss": 0.3447, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 1.1113885641098022, |
|
"learning_rate": 0.000017894736842105264, |
|
"loss": 0.3572, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 1.1501245498657227, |
|
"learning_rate": 0.00001736842105263158, |
|
"loss": 0.3468, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 1.5754578113555908, |
|
"learning_rate": 0.000016842105263157896, |
|
"loss": 0.3461, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 1.214449167251587, |
|
"learning_rate": 0.000016315789473684213, |
|
"loss": 0.3463, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 1.38958740234375, |
|
"learning_rate": 0.000015789473684210526, |
|
"loss": 0.3432, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 1.1833761930465698, |
|
"learning_rate": 0.000015263157894736842, |
|
"loss": 0.336, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 1.2316564321517944, |
|
"learning_rate": 0.000014736842105263157, |
|
"loss": 0.3266, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 1.5909250974655151, |
|
"learning_rate": 0.000014210526315789475, |
|
"loss": 0.3354, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 1.1195193529129028, |
|
"learning_rate": 0.000013684210526315791, |
|
"loss": 0.3388, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 1.1586064100265503, |
|
"learning_rate": 0.000013157894736842106, |
|
"loss": 0.3413, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.935364305973053, |
|
"learning_rate": 0.000012631578947368422, |
|
"loss": 0.3359, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 1.3456859588623047, |
|
"learning_rate": 0.000012105263157894737, |
|
"loss": 0.3328, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 1.5139777660369873, |
|
"learning_rate": 0.000011578947368421053, |
|
"loss": 0.3499, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 1.0874484777450562, |
|
"learning_rate": 0.000011052631578947368, |
|
"loss": 0.3123, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.9059338569641113, |
|
"learning_rate": 0.000010526315789473684, |
|
"loss": 0.3305, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 1.1520755290985107, |
|
"learning_rate": 0.00001, |
|
"loss": 0.3235, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 1.2957168817520142, |
|
"learning_rate": 0.000009473684210526317, |
|
"loss": 0.3204, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 1.0766379833221436, |
|
"learning_rate": 0.000008947368421052632, |
|
"loss": 0.3239, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 1.1667858362197876, |
|
"learning_rate": 0.000008421052631578948, |
|
"loss": 0.3183, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 1.364989161491394, |
|
"learning_rate": 0.000007894736842105263, |
|
"loss": 0.3326, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 1.2203367948532104, |
|
"learning_rate": 0.0000073684210526315784, |
|
"loss": 0.3205, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 1.254922866821289, |
|
"learning_rate": 0.000006842105263157896, |
|
"loss": 0.3448, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 1.471166729927063, |
|
"learning_rate": 0.000006315789473684211, |
|
"loss": 0.3307, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 1.1350233554840088, |
|
"learning_rate": 0.000005789473684210527, |
|
"loss": 0.3174, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.953658401966095, |
|
"learning_rate": 0.000005263157894736842, |
|
"loss": 0.3327, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 1.2155749797821045, |
|
"learning_rate": 0.000004736842105263159, |
|
"loss": 0.3251, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 1.106632947921753, |
|
"learning_rate": 0.000004210526315789474, |
|
"loss": 0.3011, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 1.2169803380966187, |
|
"learning_rate": 0.0000036842105263157892, |
|
"loss": 0.3331, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.7549806237220764, |
|
"learning_rate": 0.0000031578947368421056, |
|
"loss": 0.319, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 1.2598366737365723, |
|
"learning_rate": 0.000002631578947368421, |
|
"loss": 0.3201, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.991303563117981, |
|
"learning_rate": 0.000002105263157894737, |
|
"loss": 0.3144, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 1.120094656944275, |
|
"learning_rate": 0.0000015789473684210528, |
|
"loss": 0.2997, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.9906071424484253, |
|
"learning_rate": 0.0000010526315789473685, |
|
"loss": 0.3244, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 1.083082675933838, |
|
"learning_rate": 5.263157894736843e-7, |
|
"loss": 0.3382, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 1.6393320560455322, |
|
"learning_rate": 0, |
|
"loss": 0.3064, |
|
"step": 10000 |
|
} |
|
], |
|
"logging_steps": 100, |
|
"max_steps": 10000, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 5000, |
|
"total_flos": 21671046021120000, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|