| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 2.5, |
| "eval_steps": 10000, |
| "global_step": 100000, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.0, |
| "grad_norm": 5.09375, |
| "learning_rate": 1.0040201005025126e-07, |
| "loss": 0.9541, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.01, |
| "grad_norm": 5.0625, |
| "learning_rate": 1.0030150753768845e-07, |
| "loss": 1.0027, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.01, |
| "grad_norm": 4.5625, |
| "learning_rate": 1.0020100502512563e-07, |
| "loss": 0.9509, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.01, |
| "grad_norm": 5.09375, |
| "learning_rate": 1.0010050251256282e-07, |
| "loss": 0.9602, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.01, |
| "grad_norm": 5.03125, |
| "learning_rate": 1e-07, |
| "loss": 0.9338, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.01, |
| "grad_norm": 6.34375, |
| "learning_rate": 9.989949748743718e-08, |
| "loss": 0.9006, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.02, |
| "grad_norm": 6.09375, |
| "learning_rate": 9.979899497487437e-08, |
| "loss": 0.9381, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.02, |
| "grad_norm": 4.40625, |
| "learning_rate": 9.969849246231156e-08, |
| "loss": 0.931, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.02, |
| "grad_norm": 6.4375, |
| "learning_rate": 9.959798994974874e-08, |
| "loss": 0.9787, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.03, |
| "grad_norm": 8.25, |
| "learning_rate": 9.949748743718593e-08, |
| "loss": 0.9649, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.03, |
| "grad_norm": 5.78125, |
| "learning_rate": 9.939698492462312e-08, |
| "loss": 0.9302, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.03, |
| "grad_norm": 5.375, |
| "learning_rate": 9.92964824120603e-08, |
| "loss": 0.891, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.03, |
| "grad_norm": 3.078125, |
| "learning_rate": 9.919597989949749e-08, |
| "loss": 0.9733, |
| "step": 1300 |
| }, |
| { |
| "epoch": 0.04, |
| "grad_norm": 12.25, |
| "learning_rate": 9.909547738693468e-08, |
| "loss": 0.8859, |
| "step": 1400 |
| }, |
| { |
| "epoch": 0.04, |
| "grad_norm": 5.4375, |
| "learning_rate": 9.899497487437185e-08, |
| "loss": 0.9646, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.04, |
| "grad_norm": 8.5, |
| "learning_rate": 9.889447236180904e-08, |
| "loss": 0.9013, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.04, |
| "grad_norm": 12.5, |
| "learning_rate": 9.879396984924623e-08, |
| "loss": 0.9703, |
| "step": 1700 |
| }, |
| { |
| "epoch": 0.04, |
| "grad_norm": 6.03125, |
| "learning_rate": 9.869346733668342e-08, |
| "loss": 1.035, |
| "step": 1800 |
| }, |
| { |
| "epoch": 0.05, |
| "grad_norm": 7.5625, |
| "learning_rate": 9.85929648241206e-08, |
| "loss": 0.9458, |
| "step": 1900 |
| }, |
| { |
| "epoch": 0.05, |
| "grad_norm": 3.78125, |
| "learning_rate": 9.849246231155779e-08, |
| "loss": 0.9454, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.05, |
| "grad_norm": 4.40625, |
| "learning_rate": 9.839195979899498e-08, |
| "loss": 0.9185, |
| "step": 2100 |
| }, |
| { |
| "epoch": 0.06, |
| "grad_norm": 5.375, |
| "learning_rate": 9.829145728643217e-08, |
| "loss": 0.9754, |
| "step": 2200 |
| }, |
| { |
| "epoch": 0.06, |
| "grad_norm": 9.625, |
| "learning_rate": 9.819095477386935e-08, |
| "loss": 0.9614, |
| "step": 2300 |
| }, |
| { |
| "epoch": 0.06, |
| "grad_norm": 6.46875, |
| "learning_rate": 9.809045226130653e-08, |
| "loss": 0.9622, |
| "step": 2400 |
| }, |
| { |
| "epoch": 0.06, |
| "grad_norm": 8.5, |
| "learning_rate": 9.798994974874371e-08, |
| "loss": 0.9266, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.07, |
| "grad_norm": 6.125, |
| "learning_rate": 9.78894472361809e-08, |
| "loss": 0.9662, |
| "step": 2600 |
| }, |
| { |
| "epoch": 0.07, |
| "grad_norm": 7.375, |
| "learning_rate": 9.778894472361809e-08, |
| "loss": 0.9502, |
| "step": 2700 |
| }, |
| { |
| "epoch": 0.07, |
| "grad_norm": 5.375, |
| "learning_rate": 9.768844221105528e-08, |
| "loss": 0.9098, |
| "step": 2800 |
| }, |
| { |
| "epoch": 0.07, |
| "grad_norm": 8.4375, |
| "learning_rate": 9.758793969849246e-08, |
| "loss": 0.9406, |
| "step": 2900 |
| }, |
| { |
| "epoch": 0.07, |
| "grad_norm": 5.40625, |
| "learning_rate": 9.748743718592965e-08, |
| "loss": 0.925, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.08, |
| "grad_norm": 5.03125, |
| "learning_rate": 9.738693467336684e-08, |
| "loss": 0.9583, |
| "step": 3100 |
| }, |
| { |
| "epoch": 0.08, |
| "grad_norm": 6.03125, |
| "learning_rate": 9.728643216080402e-08, |
| "loss": 0.9582, |
| "step": 3200 |
| }, |
| { |
| "epoch": 0.08, |
| "grad_norm": 6.1875, |
| "learning_rate": 9.718592964824121e-08, |
| "loss": 0.9629, |
| "step": 3300 |
| }, |
| { |
| "epoch": 0.09, |
| "grad_norm": 4.03125, |
| "learning_rate": 9.708542713567839e-08, |
| "loss": 0.9405, |
| "step": 3400 |
| }, |
| { |
| "epoch": 0.09, |
| "grad_norm": 6.375, |
| "learning_rate": 9.698492462311557e-08, |
| "loss": 0.9867, |
| "step": 3500 |
| }, |
| { |
| "epoch": 0.09, |
| "grad_norm": 6.90625, |
| "learning_rate": 9.688442211055276e-08, |
| "loss": 0.9221, |
| "step": 3600 |
| }, |
| { |
| "epoch": 0.09, |
| "grad_norm": 5.09375, |
| "learning_rate": 9.678391959798995e-08, |
| "loss": 0.9397, |
| "step": 3700 |
| }, |
| { |
| "epoch": 0.1, |
| "grad_norm": 5.34375, |
| "learning_rate": 9.668341708542713e-08, |
| "loss": 0.9511, |
| "step": 3800 |
| }, |
| { |
| "epoch": 0.1, |
| "grad_norm": 4.96875, |
| "learning_rate": 9.658291457286432e-08, |
| "loss": 0.9714, |
| "step": 3900 |
| }, |
| { |
| "epoch": 0.1, |
| "grad_norm": 5.15625, |
| "learning_rate": 9.648241206030151e-08, |
| "loss": 1.0139, |
| "step": 4000 |
| }, |
| { |
| "epoch": 0.1, |
| "grad_norm": 7.875, |
| "learning_rate": 9.63819095477387e-08, |
| "loss": 0.9871, |
| "step": 4100 |
| }, |
| { |
| "epoch": 0.1, |
| "grad_norm": 5.4375, |
| "learning_rate": 9.628140703517588e-08, |
| "loss": 0.9687, |
| "step": 4200 |
| }, |
| { |
| "epoch": 0.11, |
| "grad_norm": 6.78125, |
| "learning_rate": 9.618090452261306e-08, |
| "loss": 1.0132, |
| "step": 4300 |
| }, |
| { |
| "epoch": 0.11, |
| "grad_norm": 6.40625, |
| "learning_rate": 9.608040201005024e-08, |
| "loss": 0.9381, |
| "step": 4400 |
| }, |
| { |
| "epoch": 0.11, |
| "grad_norm": 10.375, |
| "learning_rate": 9.597989949748743e-08, |
| "loss": 0.9664, |
| "step": 4500 |
| }, |
| { |
| "epoch": 0.12, |
| "grad_norm": 6.53125, |
| "learning_rate": 9.587939698492462e-08, |
| "loss": 1.0202, |
| "step": 4600 |
| }, |
| { |
| "epoch": 0.12, |
| "grad_norm": 8.25, |
| "learning_rate": 9.57788944723618e-08, |
| "loss": 0.9774, |
| "step": 4700 |
| }, |
| { |
| "epoch": 0.12, |
| "grad_norm": 7.78125, |
| "learning_rate": 9.567839195979899e-08, |
| "loss": 0.9352, |
| "step": 4800 |
| }, |
| { |
| "epoch": 0.12, |
| "grad_norm": 6.8125, |
| "learning_rate": 9.557788944723618e-08, |
| "loss": 0.9894, |
| "step": 4900 |
| }, |
| { |
| "epoch": 0.12, |
| "grad_norm": 7.46875, |
| "learning_rate": 9.547738693467337e-08, |
| "loss": 0.9708, |
| "step": 5000 |
| }, |
| { |
| "epoch": 0.13, |
| "grad_norm": 6.46875, |
| "learning_rate": 9.537688442211055e-08, |
| "loss": 0.9545, |
| "step": 5100 |
| }, |
| { |
| "epoch": 0.13, |
| "grad_norm": 6.34375, |
| "learning_rate": 9.527638190954774e-08, |
| "loss": 0.9299, |
| "step": 5200 |
| }, |
| { |
| "epoch": 0.13, |
| "grad_norm": 3.890625, |
| "learning_rate": 9.517587939698492e-08, |
| "loss": 1.0221, |
| "step": 5300 |
| }, |
| { |
| "epoch": 0.14, |
| "grad_norm": 6.3125, |
| "learning_rate": 9.50753768844221e-08, |
| "loss": 1.0046, |
| "step": 5400 |
| }, |
| { |
| "epoch": 0.14, |
| "grad_norm": 5.59375, |
| "learning_rate": 9.497487437185929e-08, |
| "loss": 0.9403, |
| "step": 5500 |
| }, |
| { |
| "epoch": 0.14, |
| "grad_norm": 6.375, |
| "learning_rate": 9.487437185929648e-08, |
| "loss": 0.8997, |
| "step": 5600 |
| }, |
| { |
| "epoch": 0.14, |
| "grad_norm": 6.90625, |
| "learning_rate": 9.477386934673366e-08, |
| "loss": 1.0223, |
| "step": 5700 |
| }, |
| { |
| "epoch": 0.14, |
| "grad_norm": 11.0625, |
| "learning_rate": 9.467336683417085e-08, |
| "loss": 0.9861, |
| "step": 5800 |
| }, |
| { |
| "epoch": 0.15, |
| "grad_norm": 7.5625, |
| "learning_rate": 9.457286432160804e-08, |
| "loss": 0.8784, |
| "step": 5900 |
| }, |
| { |
| "epoch": 0.15, |
| "grad_norm": 8.75, |
| "learning_rate": 9.447236180904523e-08, |
| "loss": 0.9643, |
| "step": 6000 |
| }, |
| { |
| "epoch": 0.15, |
| "grad_norm": 7.6875, |
| "learning_rate": 9.437185929648241e-08, |
| "loss": 0.9867, |
| "step": 6100 |
| }, |
| { |
| "epoch": 0.15, |
| "grad_norm": 4.625, |
| "learning_rate": 9.42713567839196e-08, |
| "loss": 0.9421, |
| "step": 6200 |
| }, |
| { |
| "epoch": 0.16, |
| "grad_norm": 11.5, |
| "learning_rate": 9.417085427135679e-08, |
| "loss": 1.0064, |
| "step": 6300 |
| }, |
| { |
| "epoch": 0.16, |
| "grad_norm": 7.0, |
| "learning_rate": 9.407035175879398e-08, |
| "loss": 0.9879, |
| "step": 6400 |
| }, |
| { |
| "epoch": 0.16, |
| "grad_norm": 5.5625, |
| "learning_rate": 9.396984924623116e-08, |
| "loss": 0.941, |
| "step": 6500 |
| }, |
| { |
| "epoch": 0.17, |
| "grad_norm": 6.8125, |
| "learning_rate": 9.386934673366835e-08, |
| "loss": 0.9675, |
| "step": 6600 |
| }, |
| { |
| "epoch": 0.17, |
| "grad_norm": 7.59375, |
| "learning_rate": 9.376884422110554e-08, |
| "loss": 1.008, |
| "step": 6700 |
| }, |
| { |
| "epoch": 0.17, |
| "grad_norm": 4.5625, |
| "learning_rate": 9.366834170854272e-08, |
| "loss": 0.9385, |
| "step": 6800 |
| }, |
| { |
| "epoch": 0.17, |
| "grad_norm": 5.875, |
| "learning_rate": 9.356783919597991e-08, |
| "loss": 0.9606, |
| "step": 6900 |
| }, |
| { |
| "epoch": 0.17, |
| "grad_norm": 6.5625, |
| "learning_rate": 9.34673366834171e-08, |
| "loss": 0.9215, |
| "step": 7000 |
| }, |
| { |
| "epoch": 0.18, |
| "grad_norm": 8.6875, |
| "learning_rate": 9.336683417085427e-08, |
| "loss": 0.9214, |
| "step": 7100 |
| }, |
| { |
| "epoch": 0.18, |
| "grad_norm": 5.8125, |
| "learning_rate": 9.326633165829146e-08, |
| "loss": 0.9777, |
| "step": 7200 |
| }, |
| { |
| "epoch": 0.18, |
| "grad_norm": 5.59375, |
| "learning_rate": 9.316582914572865e-08, |
| "loss": 1.0172, |
| "step": 7300 |
| }, |
| { |
| "epoch": 0.18, |
| "grad_norm": 5.375, |
| "learning_rate": 9.306532663316583e-08, |
| "loss": 0.9319, |
| "step": 7400 |
| }, |
| { |
| "epoch": 0.19, |
| "grad_norm": 6.9375, |
| "learning_rate": 9.296482412060302e-08, |
| "loss": 0.9762, |
| "step": 7500 |
| }, |
| { |
| "epoch": 0.19, |
| "grad_norm": 7.4375, |
| "learning_rate": 9.286432160804021e-08, |
| "loss": 0.9869, |
| "step": 7600 |
| }, |
| { |
| "epoch": 0.19, |
| "grad_norm": 5.71875, |
| "learning_rate": 9.27638190954774e-08, |
| "loss": 0.973, |
| "step": 7700 |
| }, |
| { |
| "epoch": 0.2, |
| "grad_norm": 6.90625, |
| "learning_rate": 9.266331658291458e-08, |
| "loss": 1.0055, |
| "step": 7800 |
| }, |
| { |
| "epoch": 0.2, |
| "grad_norm": 5.21875, |
| "learning_rate": 9.256281407035177e-08, |
| "loss": 0.893, |
| "step": 7900 |
| }, |
| { |
| "epoch": 0.2, |
| "grad_norm": 5.46875, |
| "learning_rate": 9.246231155778894e-08, |
| "loss": 0.9886, |
| "step": 8000 |
| }, |
| { |
| "epoch": 0.2, |
| "grad_norm": 6.15625, |
| "learning_rate": 9.236180904522613e-08, |
| "loss": 0.9345, |
| "step": 8100 |
| }, |
| { |
| "epoch": 0.2, |
| "grad_norm": 4.28125, |
| "learning_rate": 9.226130653266332e-08, |
| "loss": 0.9994, |
| "step": 8200 |
| }, |
| { |
| "epoch": 0.21, |
| "grad_norm": 7.8125, |
| "learning_rate": 9.21608040201005e-08, |
| "loss": 0.8965, |
| "step": 8300 |
| }, |
| { |
| "epoch": 0.21, |
| "grad_norm": 6.15625, |
| "learning_rate": 9.20603015075377e-08, |
| "loss": 0.9273, |
| "step": 8400 |
| }, |
| { |
| "epoch": 0.21, |
| "grad_norm": 6.5, |
| "learning_rate": 9.195979899497488e-08, |
| "loss": 0.9655, |
| "step": 8500 |
| }, |
| { |
| "epoch": 0.21, |
| "grad_norm": 6.15625, |
| "learning_rate": 9.185929648241207e-08, |
| "loss": 0.9608, |
| "step": 8600 |
| }, |
| { |
| "epoch": 0.22, |
| "grad_norm": 6.1875, |
| "learning_rate": 9.175879396984926e-08, |
| "loss": 0.9906, |
| "step": 8700 |
| }, |
| { |
| "epoch": 0.22, |
| "grad_norm": 5.78125, |
| "learning_rate": 9.165829145728644e-08, |
| "loss": 0.9592, |
| "step": 8800 |
| }, |
| { |
| "epoch": 0.22, |
| "grad_norm": 6.21875, |
| "learning_rate": 9.155778894472363e-08, |
| "loss": 0.9491, |
| "step": 8900 |
| }, |
| { |
| "epoch": 0.23, |
| "grad_norm": 11.4375, |
| "learning_rate": 9.14572864321608e-08, |
| "loss": 0.9509, |
| "step": 9000 |
| }, |
| { |
| "epoch": 0.23, |
| "grad_norm": 6.1875, |
| "learning_rate": 9.135678391959799e-08, |
| "loss": 0.9575, |
| "step": 9100 |
| }, |
| { |
| "epoch": 0.23, |
| "grad_norm": 5.15625, |
| "learning_rate": 9.125628140703518e-08, |
| "loss": 0.9879, |
| "step": 9200 |
| }, |
| { |
| "epoch": 0.23, |
| "grad_norm": 4.5, |
| "learning_rate": 9.115577889447237e-08, |
| "loss": 0.9281, |
| "step": 9300 |
| }, |
| { |
| "epoch": 0.23, |
| "grad_norm": 7.625, |
| "learning_rate": 9.105527638190955e-08, |
| "loss": 0.9937, |
| "step": 9400 |
| }, |
| { |
| "epoch": 0.24, |
| "grad_norm": 6.21875, |
| "learning_rate": 9.095477386934674e-08, |
| "loss": 0.9812, |
| "step": 9500 |
| }, |
| { |
| "epoch": 0.24, |
| "grad_norm": 4.625, |
| "learning_rate": 9.085427135678393e-08, |
| "loss": 0.9643, |
| "step": 9600 |
| }, |
| { |
| "epoch": 0.24, |
| "grad_norm": 9.625, |
| "learning_rate": 9.075376884422111e-08, |
| "loss": 1.0037, |
| "step": 9700 |
| }, |
| { |
| "epoch": 0.24, |
| "grad_norm": 4.15625, |
| "learning_rate": 9.06532663316583e-08, |
| "loss": 1.0041, |
| "step": 9800 |
| }, |
| { |
| "epoch": 0.25, |
| "grad_norm": 4.40625, |
| "learning_rate": 9.055276381909548e-08, |
| "loss": 0.9367, |
| "step": 9900 |
| }, |
| { |
| "epoch": 0.25, |
| "grad_norm": 4.8125, |
| "learning_rate": 9.045226130653266e-08, |
| "loss": 0.9733, |
| "step": 10000 |
| }, |
| { |
| "epoch": 0.25, |
| "eval_loss": 1.025413990020752, |
| "eval_runtime": 60.1, |
| "eval_samples_per_second": 16.639, |
| "eval_steps_per_second": 4.16, |
| "step": 10000 |
| }, |
| { |
| "epoch": 0.25, |
| "grad_norm": 5.1875, |
| "learning_rate": 9.035175879396985e-08, |
| "loss": 0.9202, |
| "step": 10100 |
| }, |
| { |
| "epoch": 0.26, |
| "grad_norm": 5.78125, |
| "learning_rate": 9.025125628140704e-08, |
| "loss": 0.9826, |
| "step": 10200 |
| }, |
| { |
| "epoch": 0.26, |
| "grad_norm": 9.1875, |
| "learning_rate": 9.015075376884422e-08, |
| "loss": 0.9588, |
| "step": 10300 |
| }, |
| { |
| "epoch": 0.26, |
| "grad_norm": 7.71875, |
| "learning_rate": 9.005025125628141e-08, |
| "loss": 0.9928, |
| "step": 10400 |
| }, |
| { |
| "epoch": 0.26, |
| "grad_norm": 7.65625, |
| "learning_rate": 8.99497487437186e-08, |
| "loss": 0.9502, |
| "step": 10500 |
| }, |
| { |
| "epoch": 0.27, |
| "grad_norm": 7.59375, |
| "learning_rate": 8.984924623115579e-08, |
| "loss": 0.9946, |
| "step": 10600 |
| }, |
| { |
| "epoch": 0.27, |
| "grad_norm": 7.0, |
| "learning_rate": 8.974874371859297e-08, |
| "loss": 0.9375, |
| "step": 10700 |
| }, |
| { |
| "epoch": 0.27, |
| "grad_norm": 6.53125, |
| "learning_rate": 8.964824120603016e-08, |
| "loss": 0.9633, |
| "step": 10800 |
| }, |
| { |
| "epoch": 0.27, |
| "grad_norm": 6.625, |
| "learning_rate": 8.954773869346733e-08, |
| "loss": 0.9744, |
| "step": 10900 |
| }, |
| { |
| "epoch": 0.28, |
| "grad_norm": 4.09375, |
| "learning_rate": 8.944723618090452e-08, |
| "loss": 0.9602, |
| "step": 11000 |
| }, |
| { |
| "epoch": 0.28, |
| "grad_norm": 6.5, |
| "learning_rate": 8.934673366834171e-08, |
| "loss": 0.96, |
| "step": 11100 |
| }, |
| { |
| "epoch": 0.28, |
| "grad_norm": 7.0625, |
| "learning_rate": 8.92462311557789e-08, |
| "loss": 0.8934, |
| "step": 11200 |
| }, |
| { |
| "epoch": 0.28, |
| "grad_norm": 8.1875, |
| "learning_rate": 8.914572864321608e-08, |
| "loss": 0.9041, |
| "step": 11300 |
| }, |
| { |
| "epoch": 0.28, |
| "grad_norm": 6.03125, |
| "learning_rate": 8.904522613065327e-08, |
| "loss": 0.9754, |
| "step": 11400 |
| }, |
| { |
| "epoch": 0.29, |
| "grad_norm": 11.5, |
| "learning_rate": 8.894472361809046e-08, |
| "loss": 0.9461, |
| "step": 11500 |
| }, |
| { |
| "epoch": 0.29, |
| "grad_norm": 2.84375, |
| "learning_rate": 8.884422110552765e-08, |
| "loss": 0.9161, |
| "step": 11600 |
| }, |
| { |
| "epoch": 0.29, |
| "grad_norm": 8.25, |
| "learning_rate": 8.874371859296483e-08, |
| "loss": 0.9703, |
| "step": 11700 |
| }, |
| { |
| "epoch": 0.29, |
| "grad_norm": 5.09375, |
| "learning_rate": 8.8643216080402e-08, |
| "loss": 1.0176, |
| "step": 11800 |
| }, |
| { |
| "epoch": 0.3, |
| "grad_norm": 5.90625, |
| "learning_rate": 8.85427135678392e-08, |
| "loss": 0.9929, |
| "step": 11900 |
| }, |
| { |
| "epoch": 0.3, |
| "grad_norm": 4.90625, |
| "learning_rate": 8.844221105527638e-08, |
| "loss": 1.0158, |
| "step": 12000 |
| }, |
| { |
| "epoch": 0.3, |
| "grad_norm": 4.75, |
| "learning_rate": 8.834170854271357e-08, |
| "loss": 0.856, |
| "step": 12100 |
| }, |
| { |
| "epoch": 0.3, |
| "grad_norm": 5.75, |
| "learning_rate": 8.824120603015076e-08, |
| "loss": 0.9228, |
| "step": 12200 |
| }, |
| { |
| "epoch": 0.31, |
| "grad_norm": 6.5625, |
| "learning_rate": 8.814070351758794e-08, |
| "loss": 0.9731, |
| "step": 12300 |
| }, |
| { |
| "epoch": 0.31, |
| "grad_norm": 6.3125, |
| "learning_rate": 8.804020100502513e-08, |
| "loss": 0.9674, |
| "step": 12400 |
| }, |
| { |
| "epoch": 0.31, |
| "grad_norm": 7.15625, |
| "learning_rate": 8.793969849246232e-08, |
| "loss": 0.9545, |
| "step": 12500 |
| }, |
| { |
| "epoch": 0.32, |
| "grad_norm": 6.0, |
| "learning_rate": 8.78391959798995e-08, |
| "loss": 0.9233, |
| "step": 12600 |
| }, |
| { |
| "epoch": 0.32, |
| "grad_norm": 11.8125, |
| "learning_rate": 8.773869346733669e-08, |
| "loss": 1.0133, |
| "step": 12700 |
| }, |
| { |
| "epoch": 0.32, |
| "grad_norm": 4.5, |
| "learning_rate": 8.763819095477387e-08, |
| "loss": 0.9488, |
| "step": 12800 |
| }, |
| { |
| "epoch": 0.32, |
| "grad_norm": 6.3125, |
| "learning_rate": 8.753768844221105e-08, |
| "loss": 0.9521, |
| "step": 12900 |
| }, |
| { |
| "epoch": 0.33, |
| "grad_norm": 7.1875, |
| "learning_rate": 8.743718592964824e-08, |
| "loss": 0.9728, |
| "step": 13000 |
| }, |
| { |
| "epoch": 0.33, |
| "grad_norm": 7.875, |
| "learning_rate": 8.733668341708543e-08, |
| "loss": 0.9632, |
| "step": 13100 |
| }, |
| { |
| "epoch": 0.33, |
| "grad_norm": 6.3125, |
| "learning_rate": 8.723618090452261e-08, |
| "loss": 0.9788, |
| "step": 13200 |
| }, |
| { |
| "epoch": 0.33, |
| "grad_norm": 4.9375, |
| "learning_rate": 8.71356783919598e-08, |
| "loss": 0.9722, |
| "step": 13300 |
| }, |
| { |
| "epoch": 0.34, |
| "grad_norm": 6.21875, |
| "learning_rate": 8.703517587939699e-08, |
| "loss": 0.9887, |
| "step": 13400 |
| }, |
| { |
| "epoch": 0.34, |
| "grad_norm": 6.5, |
| "learning_rate": 8.693467336683418e-08, |
| "loss": 0.9568, |
| "step": 13500 |
| }, |
| { |
| "epoch": 0.34, |
| "grad_norm": 7.65625, |
| "learning_rate": 8.683417085427136e-08, |
| "loss": 0.9804, |
| "step": 13600 |
| }, |
| { |
| "epoch": 0.34, |
| "grad_norm": 6.125, |
| "learning_rate": 8.673366834170854e-08, |
| "loss": 0.9415, |
| "step": 13700 |
| }, |
| { |
| "epoch": 0.34, |
| "grad_norm": 6.21875, |
| "learning_rate": 8.663316582914572e-08, |
| "loss": 0.9515, |
| "step": 13800 |
| }, |
| { |
| "epoch": 0.35, |
| "grad_norm": 5.875, |
| "learning_rate": 8.653266331658291e-08, |
| "loss": 0.9264, |
| "step": 13900 |
| }, |
| { |
| "epoch": 0.35, |
| "grad_norm": 7.25, |
| "learning_rate": 8.64321608040201e-08, |
| "loss": 0.9275, |
| "step": 14000 |
| }, |
| { |
| "epoch": 0.35, |
| "grad_norm": 7.75, |
| "learning_rate": 8.633165829145729e-08, |
| "loss": 0.8601, |
| "step": 14100 |
| }, |
| { |
| "epoch": 0.35, |
| "grad_norm": 8.75, |
| "learning_rate": 8.623115577889447e-08, |
| "loss": 0.991, |
| "step": 14200 |
| }, |
| { |
| "epoch": 0.36, |
| "grad_norm": 6.78125, |
| "learning_rate": 8.613065326633166e-08, |
| "loss": 0.9526, |
| "step": 14300 |
| }, |
| { |
| "epoch": 0.36, |
| "grad_norm": 5.6875, |
| "learning_rate": 8.603015075376885e-08, |
| "loss": 1.0127, |
| "step": 14400 |
| }, |
| { |
| "epoch": 0.36, |
| "grad_norm": 8.3125, |
| "learning_rate": 8.592964824120603e-08, |
| "loss": 0.9497, |
| "step": 14500 |
| }, |
| { |
| "epoch": 0.36, |
| "grad_norm": 4.0, |
| "learning_rate": 8.582914572864322e-08, |
| "loss": 0.976, |
| "step": 14600 |
| }, |
| { |
| "epoch": 0.37, |
| "grad_norm": 6.40625, |
| "learning_rate": 8.57286432160804e-08, |
| "loss": 0.915, |
| "step": 14700 |
| }, |
| { |
| "epoch": 0.37, |
| "grad_norm": 4.875, |
| "learning_rate": 8.562814070351758e-08, |
| "loss": 0.9379, |
| "step": 14800 |
| }, |
| { |
| "epoch": 0.37, |
| "grad_norm": 6.03125, |
| "learning_rate": 8.552763819095477e-08, |
| "loss": 0.9729, |
| "step": 14900 |
| }, |
| { |
| "epoch": 0.38, |
| "grad_norm": 5.9375, |
| "learning_rate": 8.542713567839196e-08, |
| "loss": 0.9725, |
| "step": 15000 |
| }, |
| { |
| "epoch": 0.38, |
| "grad_norm": 7.3125, |
| "learning_rate": 8.532663316582914e-08, |
| "loss": 0.9677, |
| "step": 15100 |
| }, |
| { |
| "epoch": 0.38, |
| "grad_norm": 5.71875, |
| "learning_rate": 8.522613065326633e-08, |
| "loss": 0.977, |
| "step": 15200 |
| }, |
| { |
| "epoch": 0.38, |
| "grad_norm": 7.0, |
| "learning_rate": 8.512562814070352e-08, |
| "loss": 0.9989, |
| "step": 15300 |
| }, |
| { |
| "epoch": 0.39, |
| "grad_norm": 5.78125, |
| "learning_rate": 8.502512562814071e-08, |
| "loss": 0.9648, |
| "step": 15400 |
| }, |
| { |
| "epoch": 0.39, |
| "grad_norm": 13.5625, |
| "learning_rate": 8.49246231155779e-08, |
| "loss": 1.042, |
| "step": 15500 |
| }, |
| { |
| "epoch": 0.39, |
| "grad_norm": 4.96875, |
| "learning_rate": 8.482412060301508e-08, |
| "loss": 0.9511, |
| "step": 15600 |
| }, |
| { |
| "epoch": 0.39, |
| "grad_norm": 5.15625, |
| "learning_rate": 8.472361809045226e-08, |
| "loss": 0.8953, |
| "step": 15700 |
| }, |
| { |
| "epoch": 0.4, |
| "grad_norm": 8.875, |
| "learning_rate": 8.462311557788944e-08, |
| "loss": 0.968, |
| "step": 15800 |
| }, |
| { |
| "epoch": 0.4, |
| "grad_norm": 7.71875, |
| "learning_rate": 8.452261306532663e-08, |
| "loss": 0.9539, |
| "step": 15900 |
| }, |
| { |
| "epoch": 0.4, |
| "grad_norm": 7.28125, |
| "learning_rate": 8.442211055276382e-08, |
| "loss": 0.9637, |
| "step": 16000 |
| }, |
| { |
| "epoch": 0.4, |
| "grad_norm": 7.625, |
| "learning_rate": 8.4321608040201e-08, |
| "loss": 0.8945, |
| "step": 16100 |
| }, |
| { |
| "epoch": 0.41, |
| "grad_norm": 6.40625, |
| "learning_rate": 8.422110552763819e-08, |
| "loss": 0.9473, |
| "step": 16200 |
| }, |
| { |
| "epoch": 0.41, |
| "grad_norm": 5.59375, |
| "learning_rate": 8.412060301507538e-08, |
| "loss": 0.9729, |
| "step": 16300 |
| }, |
| { |
| "epoch": 0.41, |
| "grad_norm": 6.125, |
| "learning_rate": 8.402010050251257e-08, |
| "loss": 1.0077, |
| "step": 16400 |
| }, |
| { |
| "epoch": 0.41, |
| "grad_norm": 7.5, |
| "learning_rate": 8.391959798994975e-08, |
| "loss": 1.012, |
| "step": 16500 |
| }, |
| { |
| "epoch": 0.41, |
| "grad_norm": 7.09375, |
| "learning_rate": 8.381909547738693e-08, |
| "loss": 1.0129, |
| "step": 16600 |
| }, |
| { |
| "epoch": 0.42, |
| "grad_norm": 10.125, |
| "learning_rate": 8.371859296482411e-08, |
| "loss": 0.9403, |
| "step": 16700 |
| }, |
| { |
| "epoch": 0.42, |
| "grad_norm": 7.125, |
| "learning_rate": 8.36180904522613e-08, |
| "loss": 0.9256, |
| "step": 16800 |
| }, |
| { |
| "epoch": 0.42, |
| "grad_norm": 6.53125, |
| "learning_rate": 8.351758793969849e-08, |
| "loss": 1.0053, |
| "step": 16900 |
| }, |
| { |
| "epoch": 0.42, |
| "grad_norm": 9.125, |
| "learning_rate": 8.341708542713568e-08, |
| "loss": 0.9406, |
| "step": 17000 |
| }, |
| { |
| "epoch": 0.43, |
| "grad_norm": 9.5, |
| "learning_rate": 8.331658291457286e-08, |
| "loss": 0.9271, |
| "step": 17100 |
| }, |
| { |
| "epoch": 0.43, |
| "grad_norm": 6.71875, |
| "learning_rate": 8.321608040201005e-08, |
| "loss": 0.9238, |
| "step": 17200 |
| }, |
| { |
| "epoch": 0.43, |
| "grad_norm": 5.59375, |
| "learning_rate": 8.311557788944724e-08, |
| "loss": 0.9762, |
| "step": 17300 |
| }, |
| { |
| "epoch": 0.43, |
| "grad_norm": 7.15625, |
| "learning_rate": 8.301507537688442e-08, |
| "loss": 0.9427, |
| "step": 17400 |
| }, |
| { |
| "epoch": 0.44, |
| "grad_norm": 6.09375, |
| "learning_rate": 8.291457286432161e-08, |
| "loss": 0.9835, |
| "step": 17500 |
| }, |
| { |
| "epoch": 0.44, |
| "grad_norm": 5.9375, |
| "learning_rate": 8.281407035175879e-08, |
| "loss": 0.9718, |
| "step": 17600 |
| }, |
| { |
| "epoch": 0.44, |
| "grad_norm": 6.03125, |
| "learning_rate": 8.271356783919597e-08, |
| "loss": 0.9812, |
| "step": 17700 |
| }, |
| { |
| "epoch": 0.45, |
| "grad_norm": 6.40625, |
| "learning_rate": 8.261306532663316e-08, |
| "loss": 0.9781, |
| "step": 17800 |
| }, |
| { |
| "epoch": 0.45, |
| "grad_norm": 5.0, |
| "learning_rate": 8.251256281407035e-08, |
| "loss": 0.9996, |
| "step": 17900 |
| }, |
| { |
| "epoch": 0.45, |
| "grad_norm": 6.625, |
| "learning_rate": 8.241206030150753e-08, |
| "loss": 0.9603, |
| "step": 18000 |
| }, |
| { |
| "epoch": 0.45, |
| "grad_norm": 7.125, |
| "learning_rate": 8.231155778894472e-08, |
| "loss": 0.9493, |
| "step": 18100 |
| }, |
| { |
| "epoch": 0.46, |
| "grad_norm": 4.53125, |
| "learning_rate": 8.221105527638191e-08, |
| "loss": 0.9126, |
| "step": 18200 |
| }, |
| { |
| "epoch": 0.46, |
| "grad_norm": 6.0, |
| "learning_rate": 8.21105527638191e-08, |
| "loss": 0.9765, |
| "step": 18300 |
| }, |
| { |
| "epoch": 0.46, |
| "grad_norm": 10.625, |
| "learning_rate": 8.201005025125628e-08, |
| "loss": 0.9804, |
| "step": 18400 |
| }, |
| { |
| "epoch": 0.46, |
| "grad_norm": 5.875, |
| "learning_rate": 8.190954773869346e-08, |
| "loss": 1.0446, |
| "step": 18500 |
| }, |
| { |
| "epoch": 0.47, |
| "grad_norm": 7.1875, |
| "learning_rate": 8.180904522613064e-08, |
| "loss": 0.9757, |
| "step": 18600 |
| }, |
| { |
| "epoch": 0.47, |
| "grad_norm": 6.34375, |
| "learning_rate": 8.170854271356783e-08, |
| "loss": 1.0207, |
| "step": 18700 |
| }, |
| { |
| "epoch": 0.47, |
| "grad_norm": 9.75, |
| "learning_rate": 8.160804020100503e-08, |
| "loss": 1.002, |
| "step": 18800 |
| }, |
| { |
| "epoch": 0.47, |
| "grad_norm": 8.8125, |
| "learning_rate": 8.150753768844222e-08, |
| "loss": 0.9783, |
| "step": 18900 |
| }, |
| { |
| "epoch": 0.47, |
| "grad_norm": 7.125, |
| "learning_rate": 8.140703517587941e-08, |
| "loss": 0.9591, |
| "step": 19000 |
| }, |
| { |
| "epoch": 0.48, |
| "grad_norm": 5.46875, |
| "learning_rate": 8.13065326633166e-08, |
| "loss": 0.9886, |
| "step": 19100 |
| }, |
| { |
| "epoch": 0.48, |
| "grad_norm": 9.5, |
| "learning_rate": 8.120603015075378e-08, |
| "loss": 0.9721, |
| "step": 19200 |
| }, |
| { |
| "epoch": 0.48, |
| "grad_norm": 4.59375, |
| "learning_rate": 8.110552763819096e-08, |
| "loss": 1.0216, |
| "step": 19300 |
| }, |
| { |
| "epoch": 0.48, |
| "grad_norm": 6.8125, |
| "learning_rate": 8.100502512562814e-08, |
| "loss": 0.9477, |
| "step": 19400 |
| }, |
| { |
| "epoch": 0.49, |
| "grad_norm": 7.25, |
| "learning_rate": 8.090452261306533e-08, |
| "loss": 0.9425, |
| "step": 19500 |
| }, |
| { |
| "epoch": 0.49, |
| "grad_norm": 5.625, |
| "learning_rate": 8.080402010050252e-08, |
| "loss": 1.0231, |
| "step": 19600 |
| }, |
| { |
| "epoch": 0.49, |
| "grad_norm": 5.25, |
| "learning_rate": 8.07035175879397e-08, |
| "loss": 0.9594, |
| "step": 19700 |
| }, |
| { |
| "epoch": 0.49, |
| "grad_norm": 4.46875, |
| "learning_rate": 8.060301507537689e-08, |
| "loss": 0.9909, |
| "step": 19800 |
| }, |
| { |
| "epoch": 0.5, |
| "grad_norm": 14.0, |
| "learning_rate": 8.050251256281408e-08, |
| "loss": 0.9236, |
| "step": 19900 |
| }, |
| { |
| "epoch": 0.5, |
| "grad_norm": 6.09375, |
| "learning_rate": 8.040201005025127e-08, |
| "loss": 1.0047, |
| "step": 20000 |
| }, |
| { |
| "epoch": 0.5, |
| "eval_loss": 1.0322586297988892, |
| "eval_runtime": 60.0446, |
| "eval_samples_per_second": 16.654, |
| "eval_steps_per_second": 4.164, |
| "step": 20000 |
| }, |
| { |
| "epoch": 0.5, |
| "grad_norm": 7.875, |
| "learning_rate": 8.030150753768845e-08, |
| "loss": 1.0004, |
| "step": 20100 |
| }, |
| { |
| "epoch": 0.51, |
| "grad_norm": 7.21875, |
| "learning_rate": 8.020100502512564e-08, |
| "loss": 1.0313, |
| "step": 20200 |
| }, |
| { |
| "epoch": 0.51, |
| "grad_norm": 7.40625, |
| "learning_rate": 8.010050251256281e-08, |
| "loss": 0.9795, |
| "step": 20300 |
| }, |
| { |
| "epoch": 0.51, |
| "grad_norm": 9.0625, |
| "learning_rate": 8e-08, |
| "loss": 1.0003, |
| "step": 20400 |
| }, |
| { |
| "epoch": 0.51, |
| "grad_norm": 10.8125, |
| "learning_rate": 7.989949748743719e-08, |
| "loss": 0.9945, |
| "step": 20500 |
| }, |
| { |
| "epoch": 0.52, |
| "grad_norm": 7.96875, |
| "learning_rate": 7.979899497487438e-08, |
| "loss": 1.0274, |
| "step": 20600 |
| }, |
| { |
| "epoch": 0.52, |
| "grad_norm": 5.71875, |
| "learning_rate": 7.969849246231156e-08, |
| "loss": 0.9918, |
| "step": 20700 |
| }, |
| { |
| "epoch": 0.52, |
| "grad_norm": 5.0, |
| "learning_rate": 7.959798994974875e-08, |
| "loss": 0.9931, |
| "step": 20800 |
| }, |
| { |
| "epoch": 0.52, |
| "grad_norm": 6.4375, |
| "learning_rate": 7.949748743718594e-08, |
| "loss": 1.0109, |
| "step": 20900 |
| }, |
| { |
| "epoch": 0.53, |
| "grad_norm": 3.265625, |
| "learning_rate": 7.939698492462313e-08, |
| "loss": 0.9392, |
| "step": 21000 |
| }, |
| { |
| "epoch": 0.53, |
| "grad_norm": 12.875, |
| "learning_rate": 7.929648241206031e-08, |
| "loss": 0.9646, |
| "step": 21100 |
| }, |
| { |
| "epoch": 0.53, |
| "grad_norm": 6.65625, |
| "learning_rate": 7.919597989949749e-08, |
| "loss": 0.9456, |
| "step": 21200 |
| }, |
| { |
| "epoch": 0.53, |
| "grad_norm": 6.5, |
| "learning_rate": 7.909547738693467e-08, |
| "loss": 0.9637, |
| "step": 21300 |
| }, |
| { |
| "epoch": 0.54, |
| "grad_norm": 6.84375, |
| "learning_rate": 7.899497487437186e-08, |
| "loss": 1.0326, |
| "step": 21400 |
| }, |
| { |
| "epoch": 0.54, |
| "grad_norm": 10.5625, |
| "learning_rate": 7.889447236180905e-08, |
| "loss": 0.972, |
| "step": 21500 |
| }, |
| { |
| "epoch": 0.54, |
| "grad_norm": 8.75, |
| "learning_rate": 7.879396984924624e-08, |
| "loss": 0.9322, |
| "step": 21600 |
| }, |
| { |
| "epoch": 0.54, |
| "grad_norm": 6.84375, |
| "learning_rate": 7.869346733668342e-08, |
| "loss": 1.0037, |
| "step": 21700 |
| }, |
| { |
| "epoch": 0.55, |
| "grad_norm": 7.46875, |
| "learning_rate": 7.859296482412061e-08, |
| "loss": 1.0428, |
| "step": 21800 |
| }, |
| { |
| "epoch": 0.55, |
| "grad_norm": 5.46875, |
| "learning_rate": 7.84924623115578e-08, |
| "loss": 0.8998, |
| "step": 21900 |
| }, |
| { |
| "epoch": 0.55, |
| "grad_norm": 4.875, |
| "learning_rate": 7.839195979899498e-08, |
| "loss": 1.0292, |
| "step": 22000 |
| }, |
| { |
| "epoch": 0.55, |
| "grad_norm": 6.0, |
| "learning_rate": 7.829145728643217e-08, |
| "loss": 0.9347, |
| "step": 22100 |
| }, |
| { |
| "epoch": 0.56, |
| "grad_norm": 6.53125, |
| "learning_rate": 7.819095477386935e-08, |
| "loss": 0.9662, |
| "step": 22200 |
| }, |
| { |
| "epoch": 0.56, |
| "grad_norm": 8.5625, |
| "learning_rate": 7.809045226130653e-08, |
| "loss": 0.9424, |
| "step": 22300 |
| }, |
| { |
| "epoch": 0.56, |
| "grad_norm": 6.6875, |
| "learning_rate": 7.798994974874372e-08, |
| "loss": 0.972, |
| "step": 22400 |
| }, |
| { |
| "epoch": 0.56, |
| "grad_norm": 5.59375, |
| "learning_rate": 7.788944723618091e-08, |
| "loss": 0.9037, |
| "step": 22500 |
| }, |
| { |
| "epoch": 0.56, |
| "grad_norm": 10.25, |
| "learning_rate": 7.77889447236181e-08, |
| "loss": 0.9267, |
| "step": 22600 |
| }, |
| { |
| "epoch": 0.57, |
| "grad_norm": 7.21875, |
| "learning_rate": 7.768844221105528e-08, |
| "loss": 0.941, |
| "step": 22700 |
| }, |
| { |
| "epoch": 0.57, |
| "grad_norm": 5.46875, |
| "learning_rate": 7.758793969849247e-08, |
| "loss": 0.9825, |
| "step": 22800 |
| }, |
| { |
| "epoch": 0.57, |
| "grad_norm": 7.53125, |
| "learning_rate": 7.748743718592966e-08, |
| "loss": 0.9415, |
| "step": 22900 |
| }, |
| { |
| "epoch": 0.57, |
| "grad_norm": 6.75, |
| "learning_rate": 7.738693467336684e-08, |
| "loss": 0.9295, |
| "step": 23000 |
| }, |
| { |
| "epoch": 0.58, |
| "grad_norm": 5.71875, |
| "learning_rate": 7.728643216080402e-08, |
| "loss": 0.9659, |
| "step": 23100 |
| }, |
| { |
| "epoch": 0.58, |
| "grad_norm": 6.46875, |
| "learning_rate": 7.71859296482412e-08, |
| "loss": 0.9845, |
| "step": 23200 |
| }, |
| { |
| "epoch": 0.58, |
| "grad_norm": 8.625, |
| "learning_rate": 7.708542713567839e-08, |
| "loss": 0.9109, |
| "step": 23300 |
| }, |
| { |
| "epoch": 0.58, |
| "grad_norm": 7.34375, |
| "learning_rate": 7.698492462311558e-08, |
| "loss": 0.9265, |
| "step": 23400 |
| }, |
| { |
| "epoch": 0.59, |
| "grad_norm": 4.9375, |
| "learning_rate": 7.688442211055277e-08, |
| "loss": 0.8953, |
| "step": 23500 |
| }, |
| { |
| "epoch": 0.59, |
| "grad_norm": 4.96875, |
| "learning_rate": 7.678391959798995e-08, |
| "loss": 0.8979, |
| "step": 23600 |
| }, |
| { |
| "epoch": 0.59, |
| "grad_norm": 5.21875, |
| "learning_rate": 7.668341708542714e-08, |
| "loss": 0.9755, |
| "step": 23700 |
| }, |
| { |
| "epoch": 0.59, |
| "grad_norm": 5.875, |
| "learning_rate": 7.658291457286433e-08, |
| "loss": 0.9359, |
| "step": 23800 |
| }, |
| { |
| "epoch": 0.6, |
| "grad_norm": 6.9375, |
| "learning_rate": 7.648241206030151e-08, |
| "loss": 0.982, |
| "step": 23900 |
| }, |
| { |
| "epoch": 0.6, |
| "grad_norm": 8.375, |
| "learning_rate": 7.63819095477387e-08, |
| "loss": 0.9888, |
| "step": 24000 |
| }, |
| { |
| "epoch": 0.6, |
| "grad_norm": 2.734375, |
| "learning_rate": 7.628140703517588e-08, |
| "loss": 0.9529, |
| "step": 24100 |
| }, |
| { |
| "epoch": 0.6, |
| "grad_norm": 5.96875, |
| "learning_rate": 7.618090452261306e-08, |
| "loss": 0.9906, |
| "step": 24200 |
| }, |
| { |
| "epoch": 0.61, |
| "grad_norm": 6.125, |
| "learning_rate": 7.608040201005025e-08, |
| "loss": 0.9653, |
| "step": 24300 |
| }, |
| { |
| "epoch": 0.61, |
| "grad_norm": 7.78125, |
| "learning_rate": 7.597989949748744e-08, |
| "loss": 0.9666, |
| "step": 24400 |
| }, |
| { |
| "epoch": 0.61, |
| "grad_norm": 6.15625, |
| "learning_rate": 7.587939698492462e-08, |
| "loss": 1.0016, |
| "step": 24500 |
| }, |
| { |
| "epoch": 0.61, |
| "grad_norm": 6.0, |
| "learning_rate": 7.577889447236181e-08, |
| "loss": 0.9509, |
| "step": 24600 |
| }, |
| { |
| "epoch": 0.62, |
| "grad_norm": 6.59375, |
| "learning_rate": 7.5678391959799e-08, |
| "loss": 0.9285, |
| "step": 24700 |
| }, |
| { |
| "epoch": 0.62, |
| "grad_norm": 8.1875, |
| "learning_rate": 7.557788944723619e-08, |
| "loss": 0.9239, |
| "step": 24800 |
| }, |
| { |
| "epoch": 0.62, |
| "grad_norm": 7.3125, |
| "learning_rate": 7.547738693467337e-08, |
| "loss": 0.9311, |
| "step": 24900 |
| }, |
| { |
| "epoch": 0.62, |
| "grad_norm": 5.3125, |
| "learning_rate": 7.537688442211056e-08, |
| "loss": 0.905, |
| "step": 25000 |
| }, |
| { |
| "epoch": 0.63, |
| "grad_norm": 6.53125, |
| "learning_rate": 7.527638190954774e-08, |
| "loss": 0.9557, |
| "step": 25100 |
| }, |
| { |
| "epoch": 0.63, |
| "grad_norm": 7.25, |
| "learning_rate": 7.517587939698492e-08, |
| "loss": 0.9289, |
| "step": 25200 |
| }, |
| { |
| "epoch": 0.63, |
| "grad_norm": 7.21875, |
| "learning_rate": 7.507537688442211e-08, |
| "loss": 0.9892, |
| "step": 25300 |
| }, |
| { |
| "epoch": 0.64, |
| "grad_norm": 6.84375, |
| "learning_rate": 7.49748743718593e-08, |
| "loss": 0.9056, |
| "step": 25400 |
| }, |
| { |
| "epoch": 0.64, |
| "grad_norm": 5.03125, |
| "learning_rate": 7.487437185929648e-08, |
| "loss": 0.9216, |
| "step": 25500 |
| }, |
| { |
| "epoch": 0.64, |
| "grad_norm": 7.1875, |
| "learning_rate": 7.477386934673367e-08, |
| "loss": 1.0327, |
| "step": 25600 |
| }, |
| { |
| "epoch": 0.64, |
| "grad_norm": 5.53125, |
| "learning_rate": 7.467336683417086e-08, |
| "loss": 0.9853, |
| "step": 25700 |
| }, |
| { |
| "epoch": 0.65, |
| "grad_norm": 6.96875, |
| "learning_rate": 7.457286432160805e-08, |
| "loss": 0.9082, |
| "step": 25800 |
| }, |
| { |
| "epoch": 0.65, |
| "grad_norm": 4.59375, |
| "learning_rate": 7.447236180904523e-08, |
| "loss": 0.967, |
| "step": 25900 |
| }, |
| { |
| "epoch": 0.65, |
| "grad_norm": 4.3125, |
| "learning_rate": 7.437185929648241e-08, |
| "loss": 0.9624, |
| "step": 26000 |
| }, |
| { |
| "epoch": 0.65, |
| "grad_norm": 5.90625, |
| "learning_rate": 7.42713567839196e-08, |
| "loss": 0.9153, |
| "step": 26100 |
| }, |
| { |
| "epoch": 0.66, |
| "grad_norm": 7.15625, |
| "learning_rate": 7.417085427135678e-08, |
| "loss": 0.9392, |
| "step": 26200 |
| }, |
| { |
| "epoch": 0.66, |
| "grad_norm": 4.40625, |
| "learning_rate": 7.407035175879397e-08, |
| "loss": 0.9845, |
| "step": 26300 |
| }, |
| { |
| "epoch": 0.66, |
| "grad_norm": 5.03125, |
| "learning_rate": 7.396984924623116e-08, |
| "loss": 0.9416, |
| "step": 26400 |
| }, |
| { |
| "epoch": 0.66, |
| "grad_norm": 5.78125, |
| "learning_rate": 7.386934673366834e-08, |
| "loss": 1.0113, |
| "step": 26500 |
| }, |
| { |
| "epoch": 0.67, |
| "grad_norm": 7.96875, |
| "learning_rate": 7.376884422110553e-08, |
| "loss": 0.9729, |
| "step": 26600 |
| }, |
| { |
| "epoch": 0.67, |
| "grad_norm": 7.15625, |
| "learning_rate": 7.366834170854272e-08, |
| "loss": 0.9909, |
| "step": 26700 |
| }, |
| { |
| "epoch": 0.67, |
| "grad_norm": 5.25, |
| "learning_rate": 7.35678391959799e-08, |
| "loss": 0.9247, |
| "step": 26800 |
| }, |
| { |
| "epoch": 0.67, |
| "grad_norm": 5.1875, |
| "learning_rate": 7.346733668341709e-08, |
| "loss": 0.977, |
| "step": 26900 |
| }, |
| { |
| "epoch": 0.68, |
| "grad_norm": 7.125, |
| "learning_rate": 7.336683417085427e-08, |
| "loss": 0.9593, |
| "step": 27000 |
| }, |
| { |
| "epoch": 0.68, |
| "grad_norm": 6.625, |
| "learning_rate": 7.326633165829145e-08, |
| "loss": 0.9694, |
| "step": 27100 |
| }, |
| { |
| "epoch": 0.68, |
| "grad_norm": 4.375, |
| "learning_rate": 7.316582914572864e-08, |
| "loss": 0.9794, |
| "step": 27200 |
| }, |
| { |
| "epoch": 0.68, |
| "grad_norm": 6.25, |
| "learning_rate": 7.306532663316583e-08, |
| "loss": 0.9583, |
| "step": 27300 |
| }, |
| { |
| "epoch": 0.69, |
| "grad_norm": 4.3125, |
| "learning_rate": 7.296482412060301e-08, |
| "loss": 0.9174, |
| "step": 27400 |
| }, |
| { |
| "epoch": 0.69, |
| "grad_norm": 8.125, |
| "learning_rate": 7.28643216080402e-08, |
| "loss": 0.9515, |
| "step": 27500 |
| }, |
| { |
| "epoch": 0.69, |
| "grad_norm": 6.21875, |
| "learning_rate": 7.276381909547739e-08, |
| "loss": 0.9606, |
| "step": 27600 |
| }, |
| { |
| "epoch": 0.69, |
| "grad_norm": 6.8125, |
| "learning_rate": 7.266331658291458e-08, |
| "loss": 1.0048, |
| "step": 27700 |
| }, |
| { |
| "epoch": 0.69, |
| "grad_norm": 4.4375, |
| "learning_rate": 7.256281407035176e-08, |
| "loss": 0.9901, |
| "step": 27800 |
| }, |
| { |
| "epoch": 0.7, |
| "grad_norm": 4.875, |
| "learning_rate": 7.246231155778894e-08, |
| "loss": 0.9554, |
| "step": 27900 |
| }, |
| { |
| "epoch": 0.7, |
| "grad_norm": 6.5625, |
| "learning_rate": 7.236180904522612e-08, |
| "loss": 0.9625, |
| "step": 28000 |
| }, |
| { |
| "epoch": 0.7, |
| "grad_norm": 4.84375, |
| "learning_rate": 7.226130653266331e-08, |
| "loss": 0.9843, |
| "step": 28100 |
| }, |
| { |
| "epoch": 0.7, |
| "grad_norm": 3.828125, |
| "learning_rate": 7.21608040201005e-08, |
| "loss": 0.9848, |
| "step": 28200 |
| }, |
| { |
| "epoch": 0.71, |
| "grad_norm": 4.40625, |
| "learning_rate": 7.206030150753769e-08, |
| "loss": 0.9629, |
| "step": 28300 |
| }, |
| { |
| "epoch": 0.71, |
| "grad_norm": 9.625, |
| "learning_rate": 7.195979899497487e-08, |
| "loss": 0.9431, |
| "step": 28400 |
| }, |
| { |
| "epoch": 0.71, |
| "grad_norm": 8.6875, |
| "learning_rate": 7.185929648241206e-08, |
| "loss": 1.0016, |
| "step": 28500 |
| }, |
| { |
| "epoch": 0.71, |
| "grad_norm": 5.84375, |
| "learning_rate": 7.175879396984925e-08, |
| "loss": 1.0209, |
| "step": 28600 |
| }, |
| { |
| "epoch": 0.72, |
| "grad_norm": 4.1875, |
| "learning_rate": 7.165829145728644e-08, |
| "loss": 1.0161, |
| "step": 28700 |
| }, |
| { |
| "epoch": 0.72, |
| "grad_norm": 9.0, |
| "learning_rate": 7.155778894472362e-08, |
| "loss": 0.9406, |
| "step": 28800 |
| }, |
| { |
| "epoch": 0.72, |
| "grad_norm": 4.5, |
| "learning_rate": 7.14572864321608e-08, |
| "loss": 0.9077, |
| "step": 28900 |
| }, |
| { |
| "epoch": 0.72, |
| "grad_norm": 6.09375, |
| "learning_rate": 7.135678391959798e-08, |
| "loss": 0.9576, |
| "step": 29000 |
| }, |
| { |
| "epoch": 0.73, |
| "grad_norm": 7.625, |
| "learning_rate": 7.125628140703517e-08, |
| "loss": 0.9693, |
| "step": 29100 |
| }, |
| { |
| "epoch": 0.73, |
| "grad_norm": 5.96875, |
| "learning_rate": 7.115577889447236e-08, |
| "loss": 0.8866, |
| "step": 29200 |
| }, |
| { |
| "epoch": 0.73, |
| "grad_norm": 6.78125, |
| "learning_rate": 7.105527638190955e-08, |
| "loss": 1.0084, |
| "step": 29300 |
| }, |
| { |
| "epoch": 0.73, |
| "grad_norm": 6.25, |
| "learning_rate": 7.095477386934673e-08, |
| "loss": 0.9381, |
| "step": 29400 |
| }, |
| { |
| "epoch": 0.74, |
| "grad_norm": 4.53125, |
| "learning_rate": 7.085427135678392e-08, |
| "loss": 0.9975, |
| "step": 29500 |
| }, |
| { |
| "epoch": 0.74, |
| "grad_norm": 6.09375, |
| "learning_rate": 7.075376884422111e-08, |
| "loss": 0.9602, |
| "step": 29600 |
| }, |
| { |
| "epoch": 0.74, |
| "grad_norm": 7.53125, |
| "learning_rate": 7.06532663316583e-08, |
| "loss": 0.941, |
| "step": 29700 |
| }, |
| { |
| "epoch": 0.74, |
| "grad_norm": 5.75, |
| "learning_rate": 7.055276381909547e-08, |
| "loss": 1.0162, |
| "step": 29800 |
| }, |
| { |
| "epoch": 0.75, |
| "grad_norm": 7.65625, |
| "learning_rate": 7.045226130653266e-08, |
| "loss": 0.9812, |
| "step": 29900 |
| }, |
| { |
| "epoch": 0.75, |
| "grad_norm": 6.25, |
| "learning_rate": 7.035175879396984e-08, |
| "loss": 0.9399, |
| "step": 30000 |
| }, |
| { |
| "epoch": 0.75, |
| "eval_loss": 1.0255589485168457, |
| "eval_runtime": 60.091, |
| "eval_samples_per_second": 16.641, |
| "eval_steps_per_second": 4.16, |
| "step": 30000 |
| }, |
| { |
| "epoch": 0.75, |
| "grad_norm": 5.90625, |
| "learning_rate": 7.025125628140703e-08, |
| "loss": 0.9587, |
| "step": 30100 |
| }, |
| { |
| "epoch": 0.76, |
| "grad_norm": 5.71875, |
| "learning_rate": 7.015075376884422e-08, |
| "loss": 0.9292, |
| "step": 30200 |
| }, |
| { |
| "epoch": 0.76, |
| "grad_norm": 5.71875, |
| "learning_rate": 7.00502512562814e-08, |
| "loss": 0.9821, |
| "step": 30300 |
| }, |
| { |
| "epoch": 0.76, |
| "grad_norm": 5.03125, |
| "learning_rate": 6.994974874371859e-08, |
| "loss": 0.9572, |
| "step": 30400 |
| }, |
| { |
| "epoch": 0.76, |
| "grad_norm": 7.96875, |
| "learning_rate": 6.984924623115578e-08, |
| "loss": 0.9498, |
| "step": 30500 |
| }, |
| { |
| "epoch": 0.77, |
| "grad_norm": 4.375, |
| "learning_rate": 6.974874371859297e-08, |
| "loss": 0.8748, |
| "step": 30600 |
| }, |
| { |
| "epoch": 0.77, |
| "grad_norm": 7.28125, |
| "learning_rate": 6.964824120603015e-08, |
| "loss": 0.9525, |
| "step": 30700 |
| }, |
| { |
| "epoch": 0.77, |
| "grad_norm": 7.625, |
| "learning_rate": 6.954773869346733e-08, |
| "loss": 0.9893, |
| "step": 30800 |
| }, |
| { |
| "epoch": 0.77, |
| "grad_norm": 6.8125, |
| "learning_rate": 6.944723618090451e-08, |
| "loss": 0.9862, |
| "step": 30900 |
| }, |
| { |
| "epoch": 0.78, |
| "grad_norm": 8.375, |
| "learning_rate": 6.93467336683417e-08, |
| "loss": 1.0285, |
| "step": 31000 |
| }, |
| { |
| "epoch": 0.78, |
| "grad_norm": 7.96875, |
| "learning_rate": 6.924623115577889e-08, |
| "loss": 0.9803, |
| "step": 31100 |
| }, |
| { |
| "epoch": 0.78, |
| "grad_norm": 6.40625, |
| "learning_rate": 6.914572864321608e-08, |
| "loss": 0.9453, |
| "step": 31200 |
| }, |
| { |
| "epoch": 0.78, |
| "grad_norm": 4.1875, |
| "learning_rate": 6.904522613065328e-08, |
| "loss": 0.9543, |
| "step": 31300 |
| }, |
| { |
| "epoch": 0.79, |
| "grad_norm": 6.40625, |
| "learning_rate": 6.894472361809046e-08, |
| "loss": 0.941, |
| "step": 31400 |
| }, |
| { |
| "epoch": 0.79, |
| "grad_norm": 8.125, |
| "learning_rate": 6.884422110552765e-08, |
| "loss": 0.9058, |
| "step": 31500 |
| }, |
| { |
| "epoch": 0.79, |
| "grad_norm": 8.625, |
| "learning_rate": 6.874371859296483e-08, |
| "loss": 0.9443, |
| "step": 31600 |
| }, |
| { |
| "epoch": 0.79, |
| "grad_norm": 7.84375, |
| "learning_rate": 6.864321608040201e-08, |
| "loss": 0.924, |
| "step": 31700 |
| }, |
| { |
| "epoch": 0.8, |
| "grad_norm": 5.65625, |
| "learning_rate": 6.85427135678392e-08, |
| "loss": 0.9291, |
| "step": 31800 |
| }, |
| { |
| "epoch": 0.8, |
| "grad_norm": 8.9375, |
| "learning_rate": 6.844221105527639e-08, |
| "loss": 0.9687, |
| "step": 31900 |
| }, |
| { |
| "epoch": 0.8, |
| "grad_norm": 5.8125, |
| "learning_rate": 6.834170854271357e-08, |
| "loss": 0.9915, |
| "step": 32000 |
| }, |
| { |
| "epoch": 0.8, |
| "grad_norm": 20.25, |
| "learning_rate": 6.824120603015076e-08, |
| "loss": 0.9376, |
| "step": 32100 |
| }, |
| { |
| "epoch": 0.81, |
| "grad_norm": 11.0625, |
| "learning_rate": 6.814070351758795e-08, |
| "loss": 0.9131, |
| "step": 32200 |
| }, |
| { |
| "epoch": 0.81, |
| "grad_norm": 7.53125, |
| "learning_rate": 6.804020100502514e-08, |
| "loss": 0.9248, |
| "step": 32300 |
| }, |
| { |
| "epoch": 0.81, |
| "grad_norm": 6.375, |
| "learning_rate": 6.793969849246232e-08, |
| "loss": 0.9706, |
| "step": 32400 |
| }, |
| { |
| "epoch": 0.81, |
| "grad_norm": 10.125, |
| "learning_rate": 6.78391959798995e-08, |
| "loss": 0.9587, |
| "step": 32500 |
| }, |
| { |
| "epoch": 0.81, |
| "grad_norm": 7.25, |
| "learning_rate": 6.773869346733668e-08, |
| "loss": 1.0018, |
| "step": 32600 |
| }, |
| { |
| "epoch": 0.82, |
| "grad_norm": 6.78125, |
| "learning_rate": 6.763819095477387e-08, |
| "loss": 0.9087, |
| "step": 32700 |
| }, |
| { |
| "epoch": 0.82, |
| "grad_norm": 5.6875, |
| "learning_rate": 6.753768844221106e-08, |
| "loss": 1.0081, |
| "step": 32800 |
| }, |
| { |
| "epoch": 0.82, |
| "grad_norm": 7.4375, |
| "learning_rate": 6.743718592964825e-08, |
| "loss": 0.9636, |
| "step": 32900 |
| }, |
| { |
| "epoch": 0.82, |
| "grad_norm": 5.1875, |
| "learning_rate": 6.733668341708543e-08, |
| "loss": 0.8546, |
| "step": 33000 |
| }, |
| { |
| "epoch": 0.83, |
| "grad_norm": 6.0625, |
| "learning_rate": 6.723618090452262e-08, |
| "loss": 0.9708, |
| "step": 33100 |
| }, |
| { |
| "epoch": 0.83, |
| "grad_norm": 5.15625, |
| "learning_rate": 6.713567839195981e-08, |
| "loss": 0.967, |
| "step": 33200 |
| }, |
| { |
| "epoch": 0.83, |
| "grad_norm": 5.4375, |
| "learning_rate": 6.7035175879397e-08, |
| "loss": 0.9658, |
| "step": 33300 |
| }, |
| { |
| "epoch": 0.83, |
| "grad_norm": 7.65625, |
| "learning_rate": 6.693467336683418e-08, |
| "loss": 0.8818, |
| "step": 33400 |
| }, |
| { |
| "epoch": 0.84, |
| "grad_norm": 3.578125, |
| "learning_rate": 6.683417085427136e-08, |
| "loss": 0.9882, |
| "step": 33500 |
| }, |
| { |
| "epoch": 0.84, |
| "grad_norm": 5.75, |
| "learning_rate": 6.673366834170854e-08, |
| "loss": 0.9726, |
| "step": 33600 |
| }, |
| { |
| "epoch": 0.84, |
| "grad_norm": 7.03125, |
| "learning_rate": 6.663316582914573e-08, |
| "loss": 0.9468, |
| "step": 33700 |
| }, |
| { |
| "epoch": 0.84, |
| "grad_norm": 6.78125, |
| "learning_rate": 6.653266331658292e-08, |
| "loss": 0.9443, |
| "step": 33800 |
| }, |
| { |
| "epoch": 0.85, |
| "grad_norm": 6.15625, |
| "learning_rate": 6.64321608040201e-08, |
| "loss": 1.0086, |
| "step": 33900 |
| }, |
| { |
| "epoch": 0.85, |
| "grad_norm": 6.5625, |
| "learning_rate": 6.633165829145729e-08, |
| "loss": 0.9398, |
| "step": 34000 |
| }, |
| { |
| "epoch": 0.85, |
| "grad_norm": 7.5, |
| "learning_rate": 6.623115577889448e-08, |
| "loss": 0.9407, |
| "step": 34100 |
| }, |
| { |
| "epoch": 0.85, |
| "grad_norm": 4.9375, |
| "learning_rate": 6.613065326633167e-08, |
| "loss": 0.9193, |
| "step": 34200 |
| }, |
| { |
| "epoch": 0.86, |
| "grad_norm": 6.875, |
| "learning_rate": 6.603015075376885e-08, |
| "loss": 0.918, |
| "step": 34300 |
| }, |
| { |
| "epoch": 0.86, |
| "grad_norm": 6.28125, |
| "learning_rate": 6.592964824120603e-08, |
| "loss": 0.9085, |
| "step": 34400 |
| }, |
| { |
| "epoch": 0.86, |
| "grad_norm": 11.5625, |
| "learning_rate": 6.582914572864322e-08, |
| "loss": 0.9345, |
| "step": 34500 |
| }, |
| { |
| "epoch": 0.86, |
| "grad_norm": 9.1875, |
| "learning_rate": 6.57286432160804e-08, |
| "loss": 1.0002, |
| "step": 34600 |
| }, |
| { |
| "epoch": 0.87, |
| "grad_norm": 6.3125, |
| "learning_rate": 6.562814070351759e-08, |
| "loss": 0.9694, |
| "step": 34700 |
| }, |
| { |
| "epoch": 0.87, |
| "grad_norm": 9.75, |
| "learning_rate": 6.552763819095478e-08, |
| "loss": 0.9525, |
| "step": 34800 |
| }, |
| { |
| "epoch": 0.87, |
| "grad_norm": 6.0, |
| "learning_rate": 6.542713567839196e-08, |
| "loss": 0.9514, |
| "step": 34900 |
| }, |
| { |
| "epoch": 0.88, |
| "grad_norm": 7.8125, |
| "learning_rate": 6.532663316582915e-08, |
| "loss": 0.9394, |
| "step": 35000 |
| }, |
| { |
| "epoch": 0.88, |
| "grad_norm": 7.78125, |
| "learning_rate": 6.522613065326634e-08, |
| "loss": 0.9558, |
| "step": 35100 |
| }, |
| { |
| "epoch": 0.88, |
| "grad_norm": 6.9375, |
| "learning_rate": 6.512562814070353e-08, |
| "loss": 0.9644, |
| "step": 35200 |
| }, |
| { |
| "epoch": 0.88, |
| "grad_norm": 5.46875, |
| "learning_rate": 6.502512562814071e-08, |
| "loss": 0.9093, |
| "step": 35300 |
| }, |
| { |
| "epoch": 0.89, |
| "grad_norm": 8.6875, |
| "learning_rate": 6.492462311557789e-08, |
| "loss": 0.9008, |
| "step": 35400 |
| }, |
| { |
| "epoch": 0.89, |
| "grad_norm": 10.0625, |
| "learning_rate": 6.482412060301507e-08, |
| "loss": 0.9712, |
| "step": 35500 |
| }, |
| { |
| "epoch": 0.89, |
| "grad_norm": 8.1875, |
| "learning_rate": 6.472361809045226e-08, |
| "loss": 0.9111, |
| "step": 35600 |
| }, |
| { |
| "epoch": 0.89, |
| "grad_norm": 8.3125, |
| "learning_rate": 6.462311557788945e-08, |
| "loss": 0.9768, |
| "step": 35700 |
| }, |
| { |
| "epoch": 0.9, |
| "grad_norm": 4.03125, |
| "learning_rate": 6.452261306532664e-08, |
| "loss": 0.9247, |
| "step": 35800 |
| }, |
| { |
| "epoch": 0.9, |
| "grad_norm": 5.09375, |
| "learning_rate": 6.442211055276382e-08, |
| "loss": 0.9598, |
| "step": 35900 |
| }, |
| { |
| "epoch": 0.9, |
| "grad_norm": 4.75, |
| "learning_rate": 6.432160804020101e-08, |
| "loss": 0.9472, |
| "step": 36000 |
| }, |
| { |
| "epoch": 0.9, |
| "grad_norm": 7.5, |
| "learning_rate": 6.42211055276382e-08, |
| "loss": 0.976, |
| "step": 36100 |
| }, |
| { |
| "epoch": 0.91, |
| "grad_norm": 6.46875, |
| "learning_rate": 6.412060301507538e-08, |
| "loss": 0.9229, |
| "step": 36200 |
| }, |
| { |
| "epoch": 0.91, |
| "grad_norm": 6.4375, |
| "learning_rate": 6.402010050251257e-08, |
| "loss": 0.916, |
| "step": 36300 |
| }, |
| { |
| "epoch": 0.91, |
| "grad_norm": 7.90625, |
| "learning_rate": 6.391959798994975e-08, |
| "loss": 0.9456, |
| "step": 36400 |
| }, |
| { |
| "epoch": 0.91, |
| "grad_norm": 9.0, |
| "learning_rate": 6.381909547738693e-08, |
| "loss": 0.952, |
| "step": 36500 |
| }, |
| { |
| "epoch": 0.92, |
| "grad_norm": 7.09375, |
| "learning_rate": 6.371859296482412e-08, |
| "loss": 0.9805, |
| "step": 36600 |
| }, |
| { |
| "epoch": 0.92, |
| "grad_norm": 7.46875, |
| "learning_rate": 6.361809045226131e-08, |
| "loss": 0.9899, |
| "step": 36700 |
| }, |
| { |
| "epoch": 0.92, |
| "grad_norm": 5.125, |
| "learning_rate": 6.35175879396985e-08, |
| "loss": 0.9339, |
| "step": 36800 |
| }, |
| { |
| "epoch": 0.92, |
| "grad_norm": 9.5625, |
| "learning_rate": 6.341708542713568e-08, |
| "loss": 0.9171, |
| "step": 36900 |
| }, |
| { |
| "epoch": 0.93, |
| "grad_norm": 8.6875, |
| "learning_rate": 6.331658291457287e-08, |
| "loss": 0.9853, |
| "step": 37000 |
| }, |
| { |
| "epoch": 0.93, |
| "grad_norm": 5.5, |
| "learning_rate": 6.321608040201006e-08, |
| "loss": 0.9881, |
| "step": 37100 |
| }, |
| { |
| "epoch": 0.93, |
| "grad_norm": 6.78125, |
| "learning_rate": 6.311557788944724e-08, |
| "loss": 0.9987, |
| "step": 37200 |
| }, |
| { |
| "epoch": 0.93, |
| "grad_norm": 4.59375, |
| "learning_rate": 6.301507537688442e-08, |
| "loss": 1.0295, |
| "step": 37300 |
| }, |
| { |
| "epoch": 0.94, |
| "grad_norm": 5.9375, |
| "learning_rate": 6.29145728643216e-08, |
| "loss": 0.9171, |
| "step": 37400 |
| }, |
| { |
| "epoch": 0.94, |
| "grad_norm": 6.3125, |
| "learning_rate": 6.281407035175879e-08, |
| "loss": 0.9337, |
| "step": 37500 |
| }, |
| { |
| "epoch": 0.94, |
| "grad_norm": 5.21875, |
| "learning_rate": 6.271356783919598e-08, |
| "loss": 0.9479, |
| "step": 37600 |
| }, |
| { |
| "epoch": 0.94, |
| "grad_norm": 5.59375, |
| "learning_rate": 6.261306532663317e-08, |
| "loss": 0.9544, |
| "step": 37700 |
| }, |
| { |
| "epoch": 0.94, |
| "grad_norm": 6.46875, |
| "learning_rate": 6.251256281407035e-08, |
| "loss": 0.9719, |
| "step": 37800 |
| }, |
| { |
| "epoch": 0.95, |
| "grad_norm": 5.15625, |
| "learning_rate": 6.241206030150754e-08, |
| "loss": 0.9901, |
| "step": 37900 |
| }, |
| { |
| "epoch": 0.95, |
| "grad_norm": 6.28125, |
| "learning_rate": 6.231155778894473e-08, |
| "loss": 0.9377, |
| "step": 38000 |
| }, |
| { |
| "epoch": 0.95, |
| "grad_norm": 6.875, |
| "learning_rate": 6.221105527638192e-08, |
| "loss": 0.9521, |
| "step": 38100 |
| }, |
| { |
| "epoch": 0.95, |
| "grad_norm": 5.9375, |
| "learning_rate": 6.21105527638191e-08, |
| "loss": 0.9793, |
| "step": 38200 |
| }, |
| { |
| "epoch": 0.96, |
| "grad_norm": 12.875, |
| "learning_rate": 6.201005025125628e-08, |
| "loss": 1.0009, |
| "step": 38300 |
| }, |
| { |
| "epoch": 0.96, |
| "grad_norm": 5.75, |
| "learning_rate": 6.190954773869346e-08, |
| "loss": 0.9451, |
| "step": 38400 |
| }, |
| { |
| "epoch": 0.96, |
| "grad_norm": 6.46875, |
| "learning_rate": 6.180904522613065e-08, |
| "loss": 0.9397, |
| "step": 38500 |
| }, |
| { |
| "epoch": 0.96, |
| "grad_norm": 4.84375, |
| "learning_rate": 6.170854271356784e-08, |
| "loss": 0.9011, |
| "step": 38600 |
| }, |
| { |
| "epoch": 0.97, |
| "grad_norm": 5.59375, |
| "learning_rate": 6.160804020100503e-08, |
| "loss": 0.963, |
| "step": 38700 |
| }, |
| { |
| "epoch": 0.97, |
| "grad_norm": 9.375, |
| "learning_rate": 6.150753768844221e-08, |
| "loss": 0.9328, |
| "step": 38800 |
| }, |
| { |
| "epoch": 0.97, |
| "grad_norm": 6.90625, |
| "learning_rate": 6.14070351758794e-08, |
| "loss": 0.9225, |
| "step": 38900 |
| }, |
| { |
| "epoch": 0.97, |
| "grad_norm": 6.71875, |
| "learning_rate": 6.130653266331659e-08, |
| "loss": 0.9526, |
| "step": 39000 |
| }, |
| { |
| "epoch": 0.98, |
| "grad_norm": 4.53125, |
| "learning_rate": 6.120603015075377e-08, |
| "loss": 0.9599, |
| "step": 39100 |
| }, |
| { |
| "epoch": 0.98, |
| "grad_norm": 6.1875, |
| "learning_rate": 6.110552763819095e-08, |
| "loss": 0.939, |
| "step": 39200 |
| }, |
| { |
| "epoch": 0.98, |
| "grad_norm": 7.5625, |
| "learning_rate": 6.100502512562814e-08, |
| "loss": 0.8991, |
| "step": 39300 |
| }, |
| { |
| "epoch": 0.98, |
| "grad_norm": 7.6875, |
| "learning_rate": 6.090452261306532e-08, |
| "loss": 0.9519, |
| "step": 39400 |
| }, |
| { |
| "epoch": 0.99, |
| "grad_norm": 6.46875, |
| "learning_rate": 6.080402010050251e-08, |
| "loss": 0.9196, |
| "step": 39500 |
| }, |
| { |
| "epoch": 0.99, |
| "grad_norm": 5.59375, |
| "learning_rate": 6.07035175879397e-08, |
| "loss": 0.9775, |
| "step": 39600 |
| }, |
| { |
| "epoch": 0.99, |
| "grad_norm": 9.125, |
| "learning_rate": 6.060301507537688e-08, |
| "loss": 0.9111, |
| "step": 39700 |
| }, |
| { |
| "epoch": 0.99, |
| "grad_norm": 8.5, |
| "learning_rate": 6.050251256281407e-08, |
| "loss": 0.9614, |
| "step": 39800 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 3.90625, |
| "learning_rate": 6.040201005025126e-08, |
| "loss": 0.9762, |
| "step": 39900 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 6.125, |
| "learning_rate": 6.030150753768845e-08, |
| "loss": 0.8843, |
| "step": 40000 |
| }, |
| { |
| "epoch": 1.0, |
| "eval_loss": 1.0208626985549927, |
| "eval_runtime": 60.0329, |
| "eval_samples_per_second": 16.658, |
| "eval_steps_per_second": 4.164, |
| "step": 40000 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 8.5625, |
| "learning_rate": 6.020100502512563e-08, |
| "loss": 1.0214, |
| "step": 40100 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 13.0625, |
| "learning_rate": 6.010050251256281e-08, |
| "loss": 0.9954, |
| "step": 40200 |
| }, |
| { |
| "epoch": 1.01, |
| "grad_norm": 6.90625, |
| "learning_rate": 6e-08, |
| "loss": 0.9723, |
| "step": 40300 |
| }, |
| { |
| "epoch": 1.01, |
| "grad_norm": 4.125, |
| "learning_rate": 5.989949748743718e-08, |
| "loss": 0.9679, |
| "step": 40400 |
| }, |
| { |
| "epoch": 1.01, |
| "grad_norm": 6.3125, |
| "learning_rate": 5.979899497487437e-08, |
| "loss": 0.9618, |
| "step": 40500 |
| }, |
| { |
| "epoch": 1.01, |
| "grad_norm": 7.5625, |
| "learning_rate": 5.969849246231156e-08, |
| "loss": 0.9921, |
| "step": 40600 |
| }, |
| { |
| "epoch": 1.02, |
| "grad_norm": 5.09375, |
| "learning_rate": 5.9597989949748744e-08, |
| "loss": 0.9605, |
| "step": 40700 |
| }, |
| { |
| "epoch": 1.02, |
| "grad_norm": 5.53125, |
| "learning_rate": 5.949748743718593e-08, |
| "loss": 0.9766, |
| "step": 40800 |
| }, |
| { |
| "epoch": 1.02, |
| "grad_norm": 11.5625, |
| "learning_rate": 5.939698492462311e-08, |
| "loss": 0.963, |
| "step": 40900 |
| }, |
| { |
| "epoch": 1.02, |
| "grad_norm": 6.8125, |
| "learning_rate": 5.92964824120603e-08, |
| "loss": 0.9791, |
| "step": 41000 |
| }, |
| { |
| "epoch": 1.03, |
| "grad_norm": 4.84375, |
| "learning_rate": 5.9195979899497486e-08, |
| "loss": 0.9815, |
| "step": 41100 |
| }, |
| { |
| "epoch": 1.03, |
| "grad_norm": 4.59375, |
| "learning_rate": 5.909547738693467e-08, |
| "loss": 0.9632, |
| "step": 41200 |
| }, |
| { |
| "epoch": 1.03, |
| "grad_norm": 5.84375, |
| "learning_rate": 5.899497487437186e-08, |
| "loss": 0.9738, |
| "step": 41300 |
| }, |
| { |
| "epoch": 1.03, |
| "grad_norm": 13.125, |
| "learning_rate": 5.889447236180904e-08, |
| "loss": 0.956, |
| "step": 41400 |
| }, |
| { |
| "epoch": 1.04, |
| "grad_norm": 6.9375, |
| "learning_rate": 5.879396984924623e-08, |
| "loss": 0.9617, |
| "step": 41500 |
| }, |
| { |
| "epoch": 1.04, |
| "grad_norm": 6.34375, |
| "learning_rate": 5.8693467336683415e-08, |
| "loss": 0.9972, |
| "step": 41600 |
| }, |
| { |
| "epoch": 1.04, |
| "grad_norm": 9.1875, |
| "learning_rate": 5.85929648241206e-08, |
| "loss": 0.9284, |
| "step": 41700 |
| }, |
| { |
| "epoch": 1.04, |
| "grad_norm": 6.0, |
| "learning_rate": 5.849246231155778e-08, |
| "loss": 0.982, |
| "step": 41800 |
| }, |
| { |
| "epoch": 1.05, |
| "grad_norm": 5.65625, |
| "learning_rate": 5.839195979899497e-08, |
| "loss": 0.9013, |
| "step": 41900 |
| }, |
| { |
| "epoch": 1.05, |
| "grad_norm": 6.21875, |
| "learning_rate": 5.829145728643216e-08, |
| "loss": 0.994, |
| "step": 42000 |
| }, |
| { |
| "epoch": 1.05, |
| "grad_norm": 4.6875, |
| "learning_rate": 5.8190954773869345e-08, |
| "loss": 0.9819, |
| "step": 42100 |
| }, |
| { |
| "epoch": 1.05, |
| "grad_norm": 7.09375, |
| "learning_rate": 5.809045226130653e-08, |
| "loss": 0.961, |
| "step": 42200 |
| }, |
| { |
| "epoch": 1.06, |
| "grad_norm": 11.625, |
| "learning_rate": 5.798994974874371e-08, |
| "loss": 0.9869, |
| "step": 42300 |
| }, |
| { |
| "epoch": 1.06, |
| "grad_norm": 7.625, |
| "learning_rate": 5.78894472361809e-08, |
| "loss": 1.0082, |
| "step": 42400 |
| }, |
| { |
| "epoch": 1.06, |
| "grad_norm": 9.5, |
| "learning_rate": 5.778894472361809e-08, |
| "loss": 0.9392, |
| "step": 42500 |
| }, |
| { |
| "epoch": 1.06, |
| "grad_norm": 6.34375, |
| "learning_rate": 5.7688442211055274e-08, |
| "loss": 0.9234, |
| "step": 42600 |
| }, |
| { |
| "epoch": 1.07, |
| "grad_norm": 5.4375, |
| "learning_rate": 5.758793969849246e-08, |
| "loss": 0.978, |
| "step": 42700 |
| }, |
| { |
| "epoch": 1.07, |
| "grad_norm": 8.25, |
| "learning_rate": 5.748743718592964e-08, |
| "loss": 0.9698, |
| "step": 42800 |
| }, |
| { |
| "epoch": 1.07, |
| "grad_norm": 8.1875, |
| "learning_rate": 5.738693467336683e-08, |
| "loss": 0.9794, |
| "step": 42900 |
| }, |
| { |
| "epoch": 1.07, |
| "grad_norm": 7.625, |
| "learning_rate": 5.7286432160804017e-08, |
| "loss": 0.9655, |
| "step": 43000 |
| }, |
| { |
| "epoch": 1.08, |
| "grad_norm": 8.375, |
| "learning_rate": 5.7185929648241204e-08, |
| "loss": 0.9743, |
| "step": 43100 |
| }, |
| { |
| "epoch": 1.08, |
| "grad_norm": 4.375, |
| "learning_rate": 5.708542713567839e-08, |
| "loss": 0.9515, |
| "step": 43200 |
| }, |
| { |
| "epoch": 1.08, |
| "grad_norm": 6.96875, |
| "learning_rate": 5.698492462311557e-08, |
| "loss": 1.0257, |
| "step": 43300 |
| }, |
| { |
| "epoch": 1.08, |
| "grad_norm": 8.5625, |
| "learning_rate": 5.688442211055276e-08, |
| "loss": 0.9759, |
| "step": 43400 |
| }, |
| { |
| "epoch": 1.09, |
| "grad_norm": 5.90625, |
| "learning_rate": 5.6783919597989946e-08, |
| "loss": 0.9961, |
| "step": 43500 |
| }, |
| { |
| "epoch": 1.09, |
| "grad_norm": 7.5, |
| "learning_rate": 5.6683417085427133e-08, |
| "loss": 0.9859, |
| "step": 43600 |
| }, |
| { |
| "epoch": 1.09, |
| "grad_norm": 6.90625, |
| "learning_rate": 5.658291457286432e-08, |
| "loss": 0.986, |
| "step": 43700 |
| }, |
| { |
| "epoch": 1.09, |
| "grad_norm": 11.75, |
| "learning_rate": 5.6482412060301514e-08, |
| "loss": 0.9469, |
| "step": 43800 |
| }, |
| { |
| "epoch": 1.1, |
| "grad_norm": 6.5, |
| "learning_rate": 5.63819095477387e-08, |
| "loss": 0.9295, |
| "step": 43900 |
| }, |
| { |
| "epoch": 1.1, |
| "grad_norm": 6.71875, |
| "learning_rate": 5.628140703517589e-08, |
| "loss": 1.0052, |
| "step": 44000 |
| }, |
| { |
| "epoch": 1.1, |
| "grad_norm": 5.40625, |
| "learning_rate": 5.618090452261307e-08, |
| "loss": 1.018, |
| "step": 44100 |
| }, |
| { |
| "epoch": 1.1, |
| "grad_norm": 8.625, |
| "learning_rate": 5.6080402010050257e-08, |
| "loss": 0.9492, |
| "step": 44200 |
| }, |
| { |
| "epoch": 1.11, |
| "grad_norm": 8.3125, |
| "learning_rate": 5.5979899497487444e-08, |
| "loss": 0.9821, |
| "step": 44300 |
| }, |
| { |
| "epoch": 1.11, |
| "grad_norm": 6.34375, |
| "learning_rate": 5.587939698492463e-08, |
| "loss": 0.9606, |
| "step": 44400 |
| }, |
| { |
| "epoch": 1.11, |
| "grad_norm": 6.15625, |
| "learning_rate": 5.577889447236182e-08, |
| "loss": 0.9933, |
| "step": 44500 |
| }, |
| { |
| "epoch": 1.11, |
| "grad_norm": 8.0625, |
| "learning_rate": 5.5678391959799e-08, |
| "loss": 0.9813, |
| "step": 44600 |
| }, |
| { |
| "epoch": 1.12, |
| "grad_norm": 7.09375, |
| "learning_rate": 5.5577889447236186e-08, |
| "loss": 1.0051, |
| "step": 44700 |
| }, |
| { |
| "epoch": 1.12, |
| "grad_norm": 9.5625, |
| "learning_rate": 5.5477386934673373e-08, |
| "loss": 0.9277, |
| "step": 44800 |
| }, |
| { |
| "epoch": 1.12, |
| "grad_norm": 10.6875, |
| "learning_rate": 5.537688442211056e-08, |
| "loss": 0.9432, |
| "step": 44900 |
| }, |
| { |
| "epoch": 1.12, |
| "grad_norm": 4.84375, |
| "learning_rate": 5.527638190954774e-08, |
| "loss": 0.9998, |
| "step": 45000 |
| }, |
| { |
| "epoch": 1.13, |
| "grad_norm": 7.15625, |
| "learning_rate": 5.517587939698493e-08, |
| "loss": 1.0521, |
| "step": 45100 |
| }, |
| { |
| "epoch": 1.13, |
| "grad_norm": 6.0625, |
| "learning_rate": 5.5075376884422116e-08, |
| "loss": 1.0189, |
| "step": 45200 |
| }, |
| { |
| "epoch": 1.13, |
| "grad_norm": 6.5625, |
| "learning_rate": 5.49748743718593e-08, |
| "loss": 0.97, |
| "step": 45300 |
| }, |
| { |
| "epoch": 1.14, |
| "grad_norm": 6.03125, |
| "learning_rate": 5.487437185929649e-08, |
| "loss": 0.9906, |
| "step": 45400 |
| }, |
| { |
| "epoch": 1.14, |
| "grad_norm": 5.1875, |
| "learning_rate": 5.477386934673367e-08, |
| "loss": 0.9627, |
| "step": 45500 |
| }, |
| { |
| "epoch": 1.14, |
| "grad_norm": 5.0, |
| "learning_rate": 5.467336683417086e-08, |
| "loss": 0.9621, |
| "step": 45600 |
| }, |
| { |
| "epoch": 1.14, |
| "grad_norm": 9.0, |
| "learning_rate": 5.4572864321608045e-08, |
| "loss": 0.9875, |
| "step": 45700 |
| }, |
| { |
| "epoch": 1.15, |
| "grad_norm": 8.25, |
| "learning_rate": 5.447236180904523e-08, |
| "loss": 0.9537, |
| "step": 45800 |
| }, |
| { |
| "epoch": 1.15, |
| "grad_norm": 5.125, |
| "learning_rate": 5.437185929648242e-08, |
| "loss": 0.938, |
| "step": 45900 |
| }, |
| { |
| "epoch": 1.15, |
| "grad_norm": 10.25, |
| "learning_rate": 5.42713567839196e-08, |
| "loss": 0.9981, |
| "step": 46000 |
| }, |
| { |
| "epoch": 1.15, |
| "grad_norm": 7.5, |
| "learning_rate": 5.417085427135679e-08, |
| "loss": 0.9967, |
| "step": 46100 |
| }, |
| { |
| "epoch": 1.16, |
| "grad_norm": 6.75, |
| "learning_rate": 5.4070351758793975e-08, |
| "loss": 0.95, |
| "step": 46200 |
| }, |
| { |
| "epoch": 1.16, |
| "grad_norm": 5.5625, |
| "learning_rate": 5.396984924623116e-08, |
| "loss": 0.9748, |
| "step": 46300 |
| }, |
| { |
| "epoch": 1.16, |
| "grad_norm": 8.875, |
| "learning_rate": 5.386934673366835e-08, |
| "loss": 0.9643, |
| "step": 46400 |
| }, |
| { |
| "epoch": 1.16, |
| "grad_norm": 10.6875, |
| "learning_rate": 5.376884422110553e-08, |
| "loss": 0.9516, |
| "step": 46500 |
| }, |
| { |
| "epoch": 1.17, |
| "grad_norm": 5.9375, |
| "learning_rate": 5.366834170854272e-08, |
| "loss": 0.9789, |
| "step": 46600 |
| }, |
| { |
| "epoch": 1.17, |
| "grad_norm": 5.59375, |
| "learning_rate": 5.3567839195979904e-08, |
| "loss": 0.9469, |
| "step": 46700 |
| }, |
| { |
| "epoch": 1.17, |
| "grad_norm": 9.5, |
| "learning_rate": 5.346733668341709e-08, |
| "loss": 0.9931, |
| "step": 46800 |
| }, |
| { |
| "epoch": 1.17, |
| "grad_norm": 8.0, |
| "learning_rate": 5.336683417085427e-08, |
| "loss": 0.9775, |
| "step": 46900 |
| }, |
| { |
| "epoch": 1.18, |
| "grad_norm": 6.6875, |
| "learning_rate": 5.326633165829146e-08, |
| "loss": 0.9502, |
| "step": 47000 |
| }, |
| { |
| "epoch": 1.18, |
| "grad_norm": 6.96875, |
| "learning_rate": 5.3165829145728646e-08, |
| "loss": 0.9498, |
| "step": 47100 |
| }, |
| { |
| "epoch": 1.18, |
| "grad_norm": 4.96875, |
| "learning_rate": 5.3065326633165834e-08, |
| "loss": 0.9383, |
| "step": 47200 |
| }, |
| { |
| "epoch": 1.18, |
| "grad_norm": 4.875, |
| "learning_rate": 5.296482412060302e-08, |
| "loss": 0.9807, |
| "step": 47300 |
| }, |
| { |
| "epoch": 1.19, |
| "grad_norm": 7.84375, |
| "learning_rate": 5.28643216080402e-08, |
| "loss": 0.9163, |
| "step": 47400 |
| }, |
| { |
| "epoch": 1.19, |
| "grad_norm": 5.53125, |
| "learning_rate": 5.276381909547739e-08, |
| "loss": 0.9943, |
| "step": 47500 |
| }, |
| { |
| "epoch": 1.19, |
| "grad_norm": 7.03125, |
| "learning_rate": 5.2663316582914576e-08, |
| "loss": 0.9654, |
| "step": 47600 |
| }, |
| { |
| "epoch": 1.19, |
| "grad_norm": 7.40625, |
| "learning_rate": 5.256281407035176e-08, |
| "loss": 0.9992, |
| "step": 47700 |
| }, |
| { |
| "epoch": 1.2, |
| "grad_norm": 4.1875, |
| "learning_rate": 5.246231155778895e-08, |
| "loss": 0.9386, |
| "step": 47800 |
| }, |
| { |
| "epoch": 1.2, |
| "grad_norm": 8.3125, |
| "learning_rate": 5.236180904522613e-08, |
| "loss": 0.9572, |
| "step": 47900 |
| }, |
| { |
| "epoch": 1.2, |
| "grad_norm": 5.71875, |
| "learning_rate": 5.226130653266332e-08, |
| "loss": 0.9433, |
| "step": 48000 |
| }, |
| { |
| "epoch": 1.2, |
| "grad_norm": 5.46875, |
| "learning_rate": 5.2160804020100505e-08, |
| "loss": 0.9092, |
| "step": 48100 |
| }, |
| { |
| "epoch": 1.21, |
| "grad_norm": 7.46875, |
| "learning_rate": 5.206030150753769e-08, |
| "loss": 0.9815, |
| "step": 48200 |
| }, |
| { |
| "epoch": 1.21, |
| "grad_norm": 8.0625, |
| "learning_rate": 5.195979899497488e-08, |
| "loss": 0.9606, |
| "step": 48300 |
| }, |
| { |
| "epoch": 1.21, |
| "grad_norm": 6.40625, |
| "learning_rate": 5.185929648241206e-08, |
| "loss": 1.0209, |
| "step": 48400 |
| }, |
| { |
| "epoch": 1.21, |
| "grad_norm": 5.90625, |
| "learning_rate": 5.175879396984925e-08, |
| "loss": 0.9392, |
| "step": 48500 |
| }, |
| { |
| "epoch": 1.22, |
| "grad_norm": 12.1875, |
| "learning_rate": 5.1658291457286435e-08, |
| "loss": 0.9491, |
| "step": 48600 |
| }, |
| { |
| "epoch": 1.22, |
| "grad_norm": 6.6875, |
| "learning_rate": 5.155778894472362e-08, |
| "loss": 0.9291, |
| "step": 48700 |
| }, |
| { |
| "epoch": 1.22, |
| "grad_norm": 7.0, |
| "learning_rate": 5.145728643216081e-08, |
| "loss": 0.97, |
| "step": 48800 |
| }, |
| { |
| "epoch": 1.22, |
| "grad_norm": 5.03125, |
| "learning_rate": 5.135678391959799e-08, |
| "loss": 0.9782, |
| "step": 48900 |
| }, |
| { |
| "epoch": 1.23, |
| "grad_norm": 6.5, |
| "learning_rate": 5.125628140703518e-08, |
| "loss": 0.9603, |
| "step": 49000 |
| }, |
| { |
| "epoch": 1.23, |
| "grad_norm": 7.15625, |
| "learning_rate": 5.1155778894472365e-08, |
| "loss": 0.9882, |
| "step": 49100 |
| }, |
| { |
| "epoch": 1.23, |
| "grad_norm": 10.5, |
| "learning_rate": 5.105527638190955e-08, |
| "loss": 0.9778, |
| "step": 49200 |
| }, |
| { |
| "epoch": 1.23, |
| "grad_norm": 6.40625, |
| "learning_rate": 5.095477386934673e-08, |
| "loss": 0.9749, |
| "step": 49300 |
| }, |
| { |
| "epoch": 1.23, |
| "grad_norm": 5.0625, |
| "learning_rate": 5.085427135678392e-08, |
| "loss": 0.9316, |
| "step": 49400 |
| }, |
| { |
| "epoch": 1.24, |
| "grad_norm": 5.6875, |
| "learning_rate": 5.075376884422111e-08, |
| "loss": 0.9238, |
| "step": 49500 |
| }, |
| { |
| "epoch": 1.24, |
| "grad_norm": 6.1875, |
| "learning_rate": 5.0653266331658294e-08, |
| "loss": 0.9979, |
| "step": 49600 |
| }, |
| { |
| "epoch": 1.24, |
| "grad_norm": 7.96875, |
| "learning_rate": 5.055276381909548e-08, |
| "loss": 0.936, |
| "step": 49700 |
| }, |
| { |
| "epoch": 1.25, |
| "grad_norm": 6.65625, |
| "learning_rate": 5.045226130653266e-08, |
| "loss": 0.9642, |
| "step": 49800 |
| }, |
| { |
| "epoch": 1.25, |
| "grad_norm": 7.9375, |
| "learning_rate": 5.035175879396985e-08, |
| "loss": 0.9363, |
| "step": 49900 |
| }, |
| { |
| "epoch": 1.25, |
| "grad_norm": 4.28125, |
| "learning_rate": 5.0251256281407036e-08, |
| "loss": 0.9519, |
| "step": 50000 |
| }, |
| { |
| "epoch": 1.25, |
| "eval_loss": 1.057003140449524, |
| "eval_runtime": 60.0619, |
| "eval_samples_per_second": 16.649, |
| "eval_steps_per_second": 4.162, |
| "step": 50000 |
| }, |
| { |
| "epoch": 1.25, |
| "grad_norm": 7.8125, |
| "learning_rate": 5.0150753768844224e-08, |
| "loss": 0.962, |
| "step": 50100 |
| }, |
| { |
| "epoch": 1.25, |
| "grad_norm": 5.34375, |
| "learning_rate": 5.005025125628141e-08, |
| "loss": 0.9802, |
| "step": 50200 |
| }, |
| { |
| "epoch": 1.26, |
| "grad_norm": 6.40625, |
| "learning_rate": 4.994974874371859e-08, |
| "loss": 0.9832, |
| "step": 50300 |
| }, |
| { |
| "epoch": 1.26, |
| "grad_norm": 7.9375, |
| "learning_rate": 4.984924623115578e-08, |
| "loss": 0.9983, |
| "step": 50400 |
| }, |
| { |
| "epoch": 1.26, |
| "grad_norm": 7.28125, |
| "learning_rate": 4.9748743718592966e-08, |
| "loss": 0.98, |
| "step": 50500 |
| }, |
| { |
| "epoch": 1.27, |
| "grad_norm": 4.34375, |
| "learning_rate": 4.964824120603015e-08, |
| "loss": 1.0038, |
| "step": 50600 |
| }, |
| { |
| "epoch": 1.27, |
| "grad_norm": 5.625, |
| "learning_rate": 4.954773869346734e-08, |
| "loss": 0.9949, |
| "step": 50700 |
| }, |
| { |
| "epoch": 1.27, |
| "grad_norm": 3.671875, |
| "learning_rate": 4.944723618090452e-08, |
| "loss": 0.9518, |
| "step": 50800 |
| }, |
| { |
| "epoch": 1.27, |
| "grad_norm": 5.9375, |
| "learning_rate": 4.934673366834171e-08, |
| "loss": 0.9327, |
| "step": 50900 |
| }, |
| { |
| "epoch": 1.27, |
| "grad_norm": 7.71875, |
| "learning_rate": 4.9246231155778895e-08, |
| "loss": 0.9931, |
| "step": 51000 |
| }, |
| { |
| "epoch": 1.28, |
| "grad_norm": 11.0, |
| "learning_rate": 4.914572864321608e-08, |
| "loss": 0.981, |
| "step": 51100 |
| }, |
| { |
| "epoch": 1.28, |
| "grad_norm": 7.25, |
| "learning_rate": 4.904522613065326e-08, |
| "loss": 0.9865, |
| "step": 51200 |
| }, |
| { |
| "epoch": 1.28, |
| "grad_norm": 7.0, |
| "learning_rate": 4.894472361809045e-08, |
| "loss": 0.9617, |
| "step": 51300 |
| }, |
| { |
| "epoch": 1.28, |
| "grad_norm": 5.96875, |
| "learning_rate": 4.884422110552764e-08, |
| "loss": 0.9184, |
| "step": 51400 |
| }, |
| { |
| "epoch": 1.29, |
| "grad_norm": 4.125, |
| "learning_rate": 4.8743718592964825e-08, |
| "loss": 0.9074, |
| "step": 51500 |
| }, |
| { |
| "epoch": 1.29, |
| "grad_norm": 7.03125, |
| "learning_rate": 4.864321608040201e-08, |
| "loss": 0.9717, |
| "step": 51600 |
| }, |
| { |
| "epoch": 1.29, |
| "grad_norm": 8.5625, |
| "learning_rate": 4.854271356783919e-08, |
| "loss": 0.964, |
| "step": 51700 |
| }, |
| { |
| "epoch": 1.29, |
| "grad_norm": 3.671875, |
| "learning_rate": 4.844221105527638e-08, |
| "loss": 0.9795, |
| "step": 51800 |
| }, |
| { |
| "epoch": 1.3, |
| "grad_norm": 7.375, |
| "learning_rate": 4.834170854271357e-08, |
| "loss": 0.9763, |
| "step": 51900 |
| }, |
| { |
| "epoch": 1.3, |
| "grad_norm": 5.9375, |
| "learning_rate": 4.8241206030150754e-08, |
| "loss": 0.931, |
| "step": 52000 |
| }, |
| { |
| "epoch": 1.3, |
| "grad_norm": 8.375, |
| "learning_rate": 4.814070351758794e-08, |
| "loss": 0.9088, |
| "step": 52100 |
| }, |
| { |
| "epoch": 1.3, |
| "grad_norm": 6.5, |
| "learning_rate": 4.804020100502512e-08, |
| "loss": 0.9834, |
| "step": 52200 |
| }, |
| { |
| "epoch": 1.31, |
| "grad_norm": 7.96875, |
| "learning_rate": 4.793969849246231e-08, |
| "loss": 0.9953, |
| "step": 52300 |
| }, |
| { |
| "epoch": 1.31, |
| "grad_norm": 6.25, |
| "learning_rate": 4.7839195979899497e-08, |
| "loss": 1.0036, |
| "step": 52400 |
| }, |
| { |
| "epoch": 1.31, |
| "grad_norm": 8.625, |
| "learning_rate": 4.7738693467336684e-08, |
| "loss": 0.8907, |
| "step": 52500 |
| }, |
| { |
| "epoch": 1.31, |
| "grad_norm": 10.75, |
| "learning_rate": 4.763819095477387e-08, |
| "loss": 0.9583, |
| "step": 52600 |
| }, |
| { |
| "epoch": 1.32, |
| "grad_norm": 8.1875, |
| "learning_rate": 4.753768844221105e-08, |
| "loss": 0.977, |
| "step": 52700 |
| }, |
| { |
| "epoch": 1.32, |
| "grad_norm": 7.96875, |
| "learning_rate": 4.743718592964824e-08, |
| "loss": 0.9456, |
| "step": 52800 |
| }, |
| { |
| "epoch": 1.32, |
| "grad_norm": 8.9375, |
| "learning_rate": 4.7336683417085426e-08, |
| "loss": 0.9952, |
| "step": 52900 |
| }, |
| { |
| "epoch": 1.32, |
| "grad_norm": 5.8125, |
| "learning_rate": 4.7236180904522613e-08, |
| "loss": 0.9916, |
| "step": 53000 |
| }, |
| { |
| "epoch": 1.33, |
| "grad_norm": 8.9375, |
| "learning_rate": 4.71356783919598e-08, |
| "loss": 1.0113, |
| "step": 53100 |
| }, |
| { |
| "epoch": 1.33, |
| "grad_norm": 7.15625, |
| "learning_rate": 4.703517587939699e-08, |
| "loss": 0.8996, |
| "step": 53200 |
| }, |
| { |
| "epoch": 1.33, |
| "grad_norm": 6.0, |
| "learning_rate": 4.6934673366834175e-08, |
| "loss": 0.9295, |
| "step": 53300 |
| }, |
| { |
| "epoch": 1.33, |
| "grad_norm": 8.625, |
| "learning_rate": 4.683417085427136e-08, |
| "loss": 0.9844, |
| "step": 53400 |
| }, |
| { |
| "epoch": 1.34, |
| "grad_norm": 4.9375, |
| "learning_rate": 4.673366834170855e-08, |
| "loss": 1.0053, |
| "step": 53500 |
| }, |
| { |
| "epoch": 1.34, |
| "grad_norm": 6.8125, |
| "learning_rate": 4.663316582914573e-08, |
| "loss": 0.9091, |
| "step": 53600 |
| }, |
| { |
| "epoch": 1.34, |
| "grad_norm": 9.5, |
| "learning_rate": 4.653266331658292e-08, |
| "loss": 0.9683, |
| "step": 53700 |
| }, |
| { |
| "epoch": 1.34, |
| "grad_norm": 6.0625, |
| "learning_rate": 4.6432160804020105e-08, |
| "loss": 0.9715, |
| "step": 53800 |
| }, |
| { |
| "epoch": 1.35, |
| "grad_norm": 8.0625, |
| "learning_rate": 4.633165829145729e-08, |
| "loss": 0.9531, |
| "step": 53900 |
| }, |
| { |
| "epoch": 1.35, |
| "grad_norm": 6.0, |
| "learning_rate": 4.623115577889447e-08, |
| "loss": 0.956, |
| "step": 54000 |
| }, |
| { |
| "epoch": 1.35, |
| "grad_norm": 7.03125, |
| "learning_rate": 4.613065326633166e-08, |
| "loss": 0.9348, |
| "step": 54100 |
| }, |
| { |
| "epoch": 1.35, |
| "grad_norm": 9.1875, |
| "learning_rate": 4.603015075376885e-08, |
| "loss": 0.9732, |
| "step": 54200 |
| }, |
| { |
| "epoch": 1.36, |
| "grad_norm": 8.375, |
| "learning_rate": 4.5929648241206034e-08, |
| "loss": 0.9213, |
| "step": 54300 |
| }, |
| { |
| "epoch": 1.36, |
| "grad_norm": 4.125, |
| "learning_rate": 4.582914572864322e-08, |
| "loss": 0.9606, |
| "step": 54400 |
| }, |
| { |
| "epoch": 1.36, |
| "grad_norm": 6.6875, |
| "learning_rate": 4.57286432160804e-08, |
| "loss": 0.9223, |
| "step": 54500 |
| }, |
| { |
| "epoch": 1.36, |
| "grad_norm": 5.21875, |
| "learning_rate": 4.562814070351759e-08, |
| "loss": 0.9549, |
| "step": 54600 |
| }, |
| { |
| "epoch": 1.37, |
| "grad_norm": 5.9375, |
| "learning_rate": 4.5527638190954776e-08, |
| "loss": 0.9934, |
| "step": 54700 |
| }, |
| { |
| "epoch": 1.37, |
| "grad_norm": 5.6875, |
| "learning_rate": 4.5427135678391964e-08, |
| "loss": 0.962, |
| "step": 54800 |
| }, |
| { |
| "epoch": 1.37, |
| "grad_norm": 8.5, |
| "learning_rate": 4.532663316582915e-08, |
| "loss": 1.0155, |
| "step": 54900 |
| }, |
| { |
| "epoch": 1.38, |
| "grad_norm": 7.0625, |
| "learning_rate": 4.522613065326633e-08, |
| "loss": 0.964, |
| "step": 55000 |
| }, |
| { |
| "epoch": 1.38, |
| "grad_norm": 5.40625, |
| "learning_rate": 4.512562814070352e-08, |
| "loss": 0.956, |
| "step": 55100 |
| }, |
| { |
| "epoch": 1.38, |
| "grad_norm": 5.84375, |
| "learning_rate": 4.5025125628140706e-08, |
| "loss": 1.0025, |
| "step": 55200 |
| }, |
| { |
| "epoch": 1.38, |
| "grad_norm": 6.0625, |
| "learning_rate": 4.492462311557789e-08, |
| "loss": 0.9457, |
| "step": 55300 |
| }, |
| { |
| "epoch": 1.39, |
| "grad_norm": 5.40625, |
| "learning_rate": 4.482412060301508e-08, |
| "loss": 0.9152, |
| "step": 55400 |
| }, |
| { |
| "epoch": 1.39, |
| "grad_norm": 3.8125, |
| "learning_rate": 4.472361809045226e-08, |
| "loss": 1.0083, |
| "step": 55500 |
| }, |
| { |
| "epoch": 1.39, |
| "grad_norm": 5.53125, |
| "learning_rate": 4.462311557788945e-08, |
| "loss": 0.937, |
| "step": 55600 |
| }, |
| { |
| "epoch": 1.39, |
| "grad_norm": 6.65625, |
| "learning_rate": 4.4522613065326635e-08, |
| "loss": 0.9472, |
| "step": 55700 |
| }, |
| { |
| "epoch": 1.4, |
| "grad_norm": 5.0, |
| "learning_rate": 4.442211055276382e-08, |
| "loss": 1.013, |
| "step": 55800 |
| }, |
| { |
| "epoch": 1.4, |
| "grad_norm": 7.46875, |
| "learning_rate": 4.4321608040201e-08, |
| "loss": 0.9848, |
| "step": 55900 |
| }, |
| { |
| "epoch": 1.4, |
| "grad_norm": 3.4375, |
| "learning_rate": 4.422110552763819e-08, |
| "loss": 0.9673, |
| "step": 56000 |
| }, |
| { |
| "epoch": 1.4, |
| "grad_norm": 5.8125, |
| "learning_rate": 4.412060301507538e-08, |
| "loss": 0.9647, |
| "step": 56100 |
| }, |
| { |
| "epoch": 1.41, |
| "grad_norm": 7.34375, |
| "learning_rate": 4.4020100502512565e-08, |
| "loss": 0.9551, |
| "step": 56200 |
| }, |
| { |
| "epoch": 1.41, |
| "grad_norm": 6.625, |
| "learning_rate": 4.391959798994975e-08, |
| "loss": 1.0394, |
| "step": 56300 |
| }, |
| { |
| "epoch": 1.41, |
| "grad_norm": 9.25, |
| "learning_rate": 4.381909547738693e-08, |
| "loss": 0.9819, |
| "step": 56400 |
| }, |
| { |
| "epoch": 1.41, |
| "grad_norm": 6.4375, |
| "learning_rate": 4.371859296482412e-08, |
| "loss": 1.0049, |
| "step": 56500 |
| }, |
| { |
| "epoch": 1.42, |
| "grad_norm": 7.8125, |
| "learning_rate": 4.361809045226131e-08, |
| "loss": 0.9815, |
| "step": 56600 |
| }, |
| { |
| "epoch": 1.42, |
| "grad_norm": 6.375, |
| "learning_rate": 4.3517587939698494e-08, |
| "loss": 0.9417, |
| "step": 56700 |
| }, |
| { |
| "epoch": 1.42, |
| "grad_norm": 5.9375, |
| "learning_rate": 4.341708542713568e-08, |
| "loss": 0.9617, |
| "step": 56800 |
| }, |
| { |
| "epoch": 1.42, |
| "grad_norm": 6.96875, |
| "learning_rate": 4.331658291457286e-08, |
| "loss": 0.9694, |
| "step": 56900 |
| }, |
| { |
| "epoch": 1.43, |
| "grad_norm": 4.40625, |
| "learning_rate": 4.321608040201005e-08, |
| "loss": 0.9392, |
| "step": 57000 |
| }, |
| { |
| "epoch": 1.43, |
| "grad_norm": 2.28125, |
| "learning_rate": 4.3115577889447237e-08, |
| "loss": 0.939, |
| "step": 57100 |
| }, |
| { |
| "epoch": 1.43, |
| "grad_norm": 6.78125, |
| "learning_rate": 4.3015075376884424e-08, |
| "loss": 0.9567, |
| "step": 57200 |
| }, |
| { |
| "epoch": 1.43, |
| "grad_norm": 6.90625, |
| "learning_rate": 4.291457286432161e-08, |
| "loss": 0.9173, |
| "step": 57300 |
| }, |
| { |
| "epoch": 1.44, |
| "grad_norm": 7.6875, |
| "learning_rate": 4.281407035175879e-08, |
| "loss": 0.9893, |
| "step": 57400 |
| }, |
| { |
| "epoch": 1.44, |
| "grad_norm": 10.0, |
| "learning_rate": 4.271356783919598e-08, |
| "loss": 0.9424, |
| "step": 57500 |
| }, |
| { |
| "epoch": 1.44, |
| "grad_norm": 5.21875, |
| "learning_rate": 4.2613065326633166e-08, |
| "loss": 0.9104, |
| "step": 57600 |
| }, |
| { |
| "epoch": 1.44, |
| "grad_norm": 6.59375, |
| "learning_rate": 4.2512562814070353e-08, |
| "loss": 1.004, |
| "step": 57700 |
| }, |
| { |
| "epoch": 1.45, |
| "grad_norm": 5.21875, |
| "learning_rate": 4.241206030150754e-08, |
| "loss": 0.975, |
| "step": 57800 |
| }, |
| { |
| "epoch": 1.45, |
| "grad_norm": 7.71875, |
| "learning_rate": 4.231155778894472e-08, |
| "loss": 0.9487, |
| "step": 57900 |
| }, |
| { |
| "epoch": 1.45, |
| "grad_norm": 5.625, |
| "learning_rate": 4.221105527638191e-08, |
| "loss": 0.973, |
| "step": 58000 |
| }, |
| { |
| "epoch": 1.45, |
| "grad_norm": 7.6875, |
| "learning_rate": 4.2110552763819096e-08, |
| "loss": 0.9233, |
| "step": 58100 |
| }, |
| { |
| "epoch": 1.46, |
| "grad_norm": 6.0, |
| "learning_rate": 4.201005025125628e-08, |
| "loss": 1.0018, |
| "step": 58200 |
| }, |
| { |
| "epoch": 1.46, |
| "grad_norm": 7.34375, |
| "learning_rate": 4.1909547738693463e-08, |
| "loss": 0.9226, |
| "step": 58300 |
| }, |
| { |
| "epoch": 1.46, |
| "grad_norm": 4.65625, |
| "learning_rate": 4.180904522613065e-08, |
| "loss": 0.9206, |
| "step": 58400 |
| }, |
| { |
| "epoch": 1.46, |
| "grad_norm": 13.6875, |
| "learning_rate": 4.170854271356784e-08, |
| "loss": 1.0329, |
| "step": 58500 |
| }, |
| { |
| "epoch": 1.47, |
| "grad_norm": 6.6875, |
| "learning_rate": 4.1608040201005025e-08, |
| "loss": 0.9998, |
| "step": 58600 |
| }, |
| { |
| "epoch": 1.47, |
| "grad_norm": 6.71875, |
| "learning_rate": 4.150753768844221e-08, |
| "loss": 0.9444, |
| "step": 58700 |
| }, |
| { |
| "epoch": 1.47, |
| "grad_norm": 6.90625, |
| "learning_rate": 4.140703517587939e-08, |
| "loss": 0.9434, |
| "step": 58800 |
| }, |
| { |
| "epoch": 1.47, |
| "grad_norm": 5.5625, |
| "learning_rate": 4.130653266331658e-08, |
| "loss": 0.8801, |
| "step": 58900 |
| }, |
| { |
| "epoch": 1.48, |
| "grad_norm": 5.21875, |
| "learning_rate": 4.120603015075377e-08, |
| "loss": 0.9564, |
| "step": 59000 |
| }, |
| { |
| "epoch": 1.48, |
| "grad_norm": 5.375, |
| "learning_rate": 4.1105527638190955e-08, |
| "loss": 0.9745, |
| "step": 59100 |
| }, |
| { |
| "epoch": 1.48, |
| "grad_norm": 6.875, |
| "learning_rate": 4.100502512562814e-08, |
| "loss": 0.9509, |
| "step": 59200 |
| }, |
| { |
| "epoch": 1.48, |
| "grad_norm": 5.96875, |
| "learning_rate": 4.090452261306532e-08, |
| "loss": 0.8941, |
| "step": 59300 |
| }, |
| { |
| "epoch": 1.48, |
| "grad_norm": 6.75, |
| "learning_rate": 4.0804020100502516e-08, |
| "loss": 0.9393, |
| "step": 59400 |
| }, |
| { |
| "epoch": 1.49, |
| "grad_norm": 5.375, |
| "learning_rate": 4.0703517587939704e-08, |
| "loss": 0.924, |
| "step": 59500 |
| }, |
| { |
| "epoch": 1.49, |
| "grad_norm": 4.84375, |
| "learning_rate": 4.060301507537689e-08, |
| "loss": 0.9013, |
| "step": 59600 |
| }, |
| { |
| "epoch": 1.49, |
| "grad_norm": 6.125, |
| "learning_rate": 4.050251256281407e-08, |
| "loss": 0.951, |
| "step": 59700 |
| }, |
| { |
| "epoch": 1.5, |
| "grad_norm": 6.5, |
| "learning_rate": 4.040201005025126e-08, |
| "loss": 0.9097, |
| "step": 59800 |
| }, |
| { |
| "epoch": 1.5, |
| "grad_norm": 4.0625, |
| "learning_rate": 4.0301507537688446e-08, |
| "loss": 0.9846, |
| "step": 59900 |
| }, |
| { |
| "epoch": 1.5, |
| "grad_norm": 4.3125, |
| "learning_rate": 4.020100502512563e-08, |
| "loss": 1.001, |
| "step": 60000 |
| }, |
| { |
| "epoch": 1.5, |
| "eval_loss": 1.0297139883041382, |
| "eval_runtime": 60.0085, |
| "eval_samples_per_second": 16.664, |
| "eval_steps_per_second": 4.166, |
| "step": 60000 |
| }, |
| { |
| "epoch": 1.5, |
| "grad_norm": 5.65625, |
| "learning_rate": 4.010050251256282e-08, |
| "loss": 0.9566, |
| "step": 60100 |
| }, |
| { |
| "epoch": 1.5, |
| "grad_norm": 5.75, |
| "learning_rate": 4e-08, |
| "loss": 0.9717, |
| "step": 60200 |
| }, |
| { |
| "epoch": 1.51, |
| "grad_norm": 3.28125, |
| "learning_rate": 3.989949748743719e-08, |
| "loss": 0.9496, |
| "step": 60300 |
| }, |
| { |
| "epoch": 1.51, |
| "grad_norm": 8.625, |
| "learning_rate": 3.9798994974874375e-08, |
| "loss": 0.9947, |
| "step": 60400 |
| }, |
| { |
| "epoch": 1.51, |
| "grad_norm": 6.71875, |
| "learning_rate": 3.969849246231156e-08, |
| "loss": 0.9912, |
| "step": 60500 |
| }, |
| { |
| "epoch": 1.52, |
| "grad_norm": 9.875, |
| "learning_rate": 3.959798994974874e-08, |
| "loss": 0.9177, |
| "step": 60600 |
| }, |
| { |
| "epoch": 1.52, |
| "grad_norm": 8.5, |
| "learning_rate": 3.949748743718593e-08, |
| "loss": 1.026, |
| "step": 60700 |
| }, |
| { |
| "epoch": 1.52, |
| "grad_norm": 6.375, |
| "learning_rate": 3.939698492462312e-08, |
| "loss": 0.9714, |
| "step": 60800 |
| }, |
| { |
| "epoch": 1.52, |
| "grad_norm": 6.8125, |
| "learning_rate": 3.9296482412060305e-08, |
| "loss": 0.9195, |
| "step": 60900 |
| }, |
| { |
| "epoch": 1.52, |
| "grad_norm": 4.46875, |
| "learning_rate": 3.919597989949749e-08, |
| "loss": 0.9541, |
| "step": 61000 |
| }, |
| { |
| "epoch": 1.53, |
| "grad_norm": 6.09375, |
| "learning_rate": 3.909547738693467e-08, |
| "loss": 0.9479, |
| "step": 61100 |
| }, |
| { |
| "epoch": 1.53, |
| "grad_norm": 4.03125, |
| "learning_rate": 3.899497487437186e-08, |
| "loss": 0.9016, |
| "step": 61200 |
| }, |
| { |
| "epoch": 1.53, |
| "grad_norm": 4.5625, |
| "learning_rate": 3.889447236180905e-08, |
| "loss": 1.0165, |
| "step": 61300 |
| }, |
| { |
| "epoch": 1.54, |
| "grad_norm": 5.875, |
| "learning_rate": 3.8793969849246234e-08, |
| "loss": 0.9437, |
| "step": 61400 |
| }, |
| { |
| "epoch": 1.54, |
| "grad_norm": 5.90625, |
| "learning_rate": 3.869346733668342e-08, |
| "loss": 0.9593, |
| "step": 61500 |
| }, |
| { |
| "epoch": 1.54, |
| "grad_norm": 6.28125, |
| "learning_rate": 3.85929648241206e-08, |
| "loss": 0.9451, |
| "step": 61600 |
| }, |
| { |
| "epoch": 1.54, |
| "grad_norm": 8.5, |
| "learning_rate": 3.849246231155779e-08, |
| "loss": 0.9821, |
| "step": 61700 |
| }, |
| { |
| "epoch": 1.54, |
| "grad_norm": 4.15625, |
| "learning_rate": 3.8391959798994977e-08, |
| "loss": 0.9196, |
| "step": 61800 |
| }, |
| { |
| "epoch": 1.55, |
| "grad_norm": 6.65625, |
| "learning_rate": 3.8291457286432164e-08, |
| "loss": 0.9607, |
| "step": 61900 |
| }, |
| { |
| "epoch": 1.55, |
| "grad_norm": 7.0, |
| "learning_rate": 3.819095477386935e-08, |
| "loss": 0.9695, |
| "step": 62000 |
| }, |
| { |
| "epoch": 1.55, |
| "grad_norm": 8.4375, |
| "learning_rate": 3.809045226130653e-08, |
| "loss": 0.9509, |
| "step": 62100 |
| }, |
| { |
| "epoch": 1.56, |
| "grad_norm": 8.3125, |
| "learning_rate": 3.798994974874372e-08, |
| "loss": 0.9513, |
| "step": 62200 |
| }, |
| { |
| "epoch": 1.56, |
| "grad_norm": 7.875, |
| "learning_rate": 3.7889447236180906e-08, |
| "loss": 0.9426, |
| "step": 62300 |
| }, |
| { |
| "epoch": 1.56, |
| "grad_norm": 7.40625, |
| "learning_rate": 3.778894472361809e-08, |
| "loss": 0.9296, |
| "step": 62400 |
| }, |
| { |
| "epoch": 1.56, |
| "grad_norm": 6.53125, |
| "learning_rate": 3.768844221105528e-08, |
| "loss": 0.9573, |
| "step": 62500 |
| }, |
| { |
| "epoch": 1.56, |
| "grad_norm": 6.625, |
| "learning_rate": 3.758793969849246e-08, |
| "loss": 0.9514, |
| "step": 62600 |
| }, |
| { |
| "epoch": 1.57, |
| "grad_norm": 5.75, |
| "learning_rate": 3.748743718592965e-08, |
| "loss": 0.9511, |
| "step": 62700 |
| }, |
| { |
| "epoch": 1.57, |
| "grad_norm": 6.09375, |
| "learning_rate": 3.7386934673366836e-08, |
| "loss": 0.9011, |
| "step": 62800 |
| }, |
| { |
| "epoch": 1.57, |
| "grad_norm": 7.15625, |
| "learning_rate": 3.728643216080402e-08, |
| "loss": 0.9259, |
| "step": 62900 |
| }, |
| { |
| "epoch": 1.57, |
| "grad_norm": 4.53125, |
| "learning_rate": 3.7185929648241203e-08, |
| "loss": 0.9448, |
| "step": 63000 |
| }, |
| { |
| "epoch": 1.58, |
| "grad_norm": 5.53125, |
| "learning_rate": 3.708542713567839e-08, |
| "loss": 0.9775, |
| "step": 63100 |
| }, |
| { |
| "epoch": 1.58, |
| "grad_norm": 8.25, |
| "learning_rate": 3.698492462311558e-08, |
| "loss": 0.9529, |
| "step": 63200 |
| }, |
| { |
| "epoch": 1.58, |
| "grad_norm": 5.84375, |
| "learning_rate": 3.6884422110552765e-08, |
| "loss": 0.9213, |
| "step": 63300 |
| }, |
| { |
| "epoch": 1.58, |
| "grad_norm": 5.03125, |
| "learning_rate": 3.678391959798995e-08, |
| "loss": 0.9599, |
| "step": 63400 |
| }, |
| { |
| "epoch": 1.59, |
| "grad_norm": 4.9375, |
| "learning_rate": 3.668341708542713e-08, |
| "loss": 0.9787, |
| "step": 63500 |
| }, |
| { |
| "epoch": 1.59, |
| "grad_norm": 11.4375, |
| "learning_rate": 3.658291457286432e-08, |
| "loss": 0.9525, |
| "step": 63600 |
| }, |
| { |
| "epoch": 1.59, |
| "grad_norm": 5.3125, |
| "learning_rate": 3.648241206030151e-08, |
| "loss": 0.9541, |
| "step": 63700 |
| }, |
| { |
| "epoch": 1.59, |
| "grad_norm": 7.0625, |
| "learning_rate": 3.6381909547738695e-08, |
| "loss": 1.0157, |
| "step": 63800 |
| }, |
| { |
| "epoch": 1.6, |
| "grad_norm": 6.8125, |
| "learning_rate": 3.628140703517588e-08, |
| "loss": 0.9937, |
| "step": 63900 |
| }, |
| { |
| "epoch": 1.6, |
| "grad_norm": 7.21875, |
| "learning_rate": 3.618090452261306e-08, |
| "loss": 1.0143, |
| "step": 64000 |
| }, |
| { |
| "epoch": 1.6, |
| "grad_norm": 5.625, |
| "learning_rate": 3.608040201005025e-08, |
| "loss": 0.9805, |
| "step": 64100 |
| }, |
| { |
| "epoch": 1.6, |
| "grad_norm": 6.875, |
| "learning_rate": 3.597989949748744e-08, |
| "loss": 0.9366, |
| "step": 64200 |
| }, |
| { |
| "epoch": 1.61, |
| "grad_norm": 4.875, |
| "learning_rate": 3.5879396984924624e-08, |
| "loss": 0.981, |
| "step": 64300 |
| }, |
| { |
| "epoch": 1.61, |
| "grad_norm": 9.25, |
| "learning_rate": 3.577889447236181e-08, |
| "loss": 0.9633, |
| "step": 64400 |
| }, |
| { |
| "epoch": 1.61, |
| "grad_norm": 5.625, |
| "learning_rate": 3.567839195979899e-08, |
| "loss": 0.9892, |
| "step": 64500 |
| }, |
| { |
| "epoch": 1.61, |
| "grad_norm": 6.21875, |
| "learning_rate": 3.557788944723618e-08, |
| "loss": 0.9169, |
| "step": 64600 |
| }, |
| { |
| "epoch": 1.62, |
| "grad_norm": 6.75, |
| "learning_rate": 3.5477386934673366e-08, |
| "loss": 1.0164, |
| "step": 64700 |
| }, |
| { |
| "epoch": 1.62, |
| "grad_norm": 7.28125, |
| "learning_rate": 3.5376884422110554e-08, |
| "loss": 0.9512, |
| "step": 64800 |
| }, |
| { |
| "epoch": 1.62, |
| "grad_norm": 9.1875, |
| "learning_rate": 3.5276381909547734e-08, |
| "loss": 0.9667, |
| "step": 64900 |
| }, |
| { |
| "epoch": 1.62, |
| "grad_norm": 3.984375, |
| "learning_rate": 3.517587939698492e-08, |
| "loss": 0.9844, |
| "step": 65000 |
| }, |
| { |
| "epoch": 1.63, |
| "grad_norm": 6.6875, |
| "learning_rate": 3.507537688442211e-08, |
| "loss": 0.9412, |
| "step": 65100 |
| }, |
| { |
| "epoch": 1.63, |
| "grad_norm": 5.90625, |
| "learning_rate": 3.4974874371859296e-08, |
| "loss": 0.9439, |
| "step": 65200 |
| }, |
| { |
| "epoch": 1.63, |
| "grad_norm": 6.875, |
| "learning_rate": 3.487437185929648e-08, |
| "loss": 0.9834, |
| "step": 65300 |
| }, |
| { |
| "epoch": 1.64, |
| "grad_norm": 7.5, |
| "learning_rate": 3.4773869346733664e-08, |
| "loss": 0.8961, |
| "step": 65400 |
| }, |
| { |
| "epoch": 1.64, |
| "grad_norm": 6.6875, |
| "learning_rate": 3.467336683417085e-08, |
| "loss": 0.9536, |
| "step": 65500 |
| }, |
| { |
| "epoch": 1.64, |
| "grad_norm": 6.46875, |
| "learning_rate": 3.457286432160804e-08, |
| "loss": 0.9665, |
| "step": 65600 |
| }, |
| { |
| "epoch": 1.64, |
| "grad_norm": 5.40625, |
| "learning_rate": 3.447236180904523e-08, |
| "loss": 0.9056, |
| "step": 65700 |
| }, |
| { |
| "epoch": 1.65, |
| "grad_norm": 6.375, |
| "learning_rate": 3.437185929648241e-08, |
| "loss": 0.9048, |
| "step": 65800 |
| }, |
| { |
| "epoch": 1.65, |
| "grad_norm": 7.46875, |
| "learning_rate": 3.42713567839196e-08, |
| "loss": 0.9802, |
| "step": 65900 |
| }, |
| { |
| "epoch": 1.65, |
| "grad_norm": 5.875, |
| "learning_rate": 3.417085427135679e-08, |
| "loss": 0.9669, |
| "step": 66000 |
| }, |
| { |
| "epoch": 1.65, |
| "grad_norm": 6.75, |
| "learning_rate": 3.4070351758793974e-08, |
| "loss": 0.9424, |
| "step": 66100 |
| }, |
| { |
| "epoch": 1.66, |
| "grad_norm": 7.0, |
| "learning_rate": 3.396984924623116e-08, |
| "loss": 0.9306, |
| "step": 66200 |
| }, |
| { |
| "epoch": 1.66, |
| "grad_norm": 6.1875, |
| "learning_rate": 3.386934673366834e-08, |
| "loss": 0.9764, |
| "step": 66300 |
| }, |
| { |
| "epoch": 1.66, |
| "grad_norm": 9.4375, |
| "learning_rate": 3.376884422110553e-08, |
| "loss": 1.0015, |
| "step": 66400 |
| }, |
| { |
| "epoch": 1.66, |
| "grad_norm": 7.8125, |
| "learning_rate": 3.3668341708542717e-08, |
| "loss": 0.9376, |
| "step": 66500 |
| }, |
| { |
| "epoch": 1.67, |
| "grad_norm": 4.71875, |
| "learning_rate": 3.3567839195979904e-08, |
| "loss": 0.9352, |
| "step": 66600 |
| }, |
| { |
| "epoch": 1.67, |
| "grad_norm": 5.90625, |
| "learning_rate": 3.346733668341709e-08, |
| "loss": 0.895, |
| "step": 66700 |
| }, |
| { |
| "epoch": 1.67, |
| "grad_norm": 6.21875, |
| "learning_rate": 3.336683417085427e-08, |
| "loss": 0.9482, |
| "step": 66800 |
| }, |
| { |
| "epoch": 1.67, |
| "grad_norm": 3.859375, |
| "learning_rate": 3.326633165829146e-08, |
| "loss": 1.0106, |
| "step": 66900 |
| }, |
| { |
| "epoch": 1.68, |
| "grad_norm": 4.90625, |
| "learning_rate": 3.3165829145728646e-08, |
| "loss": 0.9115, |
| "step": 67000 |
| }, |
| { |
| "epoch": 1.68, |
| "grad_norm": 5.03125, |
| "learning_rate": 3.306532663316583e-08, |
| "loss": 0.9384, |
| "step": 67100 |
| }, |
| { |
| "epoch": 1.68, |
| "grad_norm": 8.375, |
| "learning_rate": 3.2964824120603014e-08, |
| "loss": 0.9182, |
| "step": 67200 |
| }, |
| { |
| "epoch": 1.68, |
| "grad_norm": 4.9375, |
| "learning_rate": 3.28643216080402e-08, |
| "loss": 0.9324, |
| "step": 67300 |
| }, |
| { |
| "epoch": 1.69, |
| "grad_norm": 5.96875, |
| "learning_rate": 3.276381909547739e-08, |
| "loss": 0.9554, |
| "step": 67400 |
| }, |
| { |
| "epoch": 1.69, |
| "grad_norm": 3.46875, |
| "learning_rate": 3.2663316582914576e-08, |
| "loss": 0.9322, |
| "step": 67500 |
| }, |
| { |
| "epoch": 1.69, |
| "grad_norm": 5.75, |
| "learning_rate": 3.256281407035176e-08, |
| "loss": 1.0046, |
| "step": 67600 |
| }, |
| { |
| "epoch": 1.69, |
| "grad_norm": 7.625, |
| "learning_rate": 3.2462311557788943e-08, |
| "loss": 0.9466, |
| "step": 67700 |
| }, |
| { |
| "epoch": 1.69, |
| "grad_norm": 4.46875, |
| "learning_rate": 3.236180904522613e-08, |
| "loss": 0.9562, |
| "step": 67800 |
| }, |
| { |
| "epoch": 1.7, |
| "grad_norm": 8.9375, |
| "learning_rate": 3.226130653266332e-08, |
| "loss": 0.9675, |
| "step": 67900 |
| }, |
| { |
| "epoch": 1.7, |
| "grad_norm": 5.3125, |
| "learning_rate": 3.2160804020100505e-08, |
| "loss": 0.9889, |
| "step": 68000 |
| }, |
| { |
| "epoch": 1.7, |
| "grad_norm": 7.1875, |
| "learning_rate": 3.206030150753769e-08, |
| "loss": 0.9679, |
| "step": 68100 |
| }, |
| { |
| "epoch": 1.71, |
| "grad_norm": 5.0625, |
| "learning_rate": 3.195979899497487e-08, |
| "loss": 0.9824, |
| "step": 68200 |
| }, |
| { |
| "epoch": 1.71, |
| "grad_norm": 9.9375, |
| "learning_rate": 3.185929648241206e-08, |
| "loss": 0.9237, |
| "step": 68300 |
| }, |
| { |
| "epoch": 1.71, |
| "grad_norm": 6.28125, |
| "learning_rate": 3.175879396984925e-08, |
| "loss": 0.9529, |
| "step": 68400 |
| }, |
| { |
| "epoch": 1.71, |
| "grad_norm": 5.96875, |
| "learning_rate": 3.1658291457286435e-08, |
| "loss": 0.9364, |
| "step": 68500 |
| }, |
| { |
| "epoch": 1.71, |
| "grad_norm": 4.84375, |
| "learning_rate": 3.155778894472362e-08, |
| "loss": 0.9004, |
| "step": 68600 |
| }, |
| { |
| "epoch": 1.72, |
| "grad_norm": 4.3125, |
| "learning_rate": 3.14572864321608e-08, |
| "loss": 0.9296, |
| "step": 68700 |
| }, |
| { |
| "epoch": 1.72, |
| "grad_norm": 5.0625, |
| "learning_rate": 3.135678391959799e-08, |
| "loss": 0.9691, |
| "step": 68800 |
| }, |
| { |
| "epoch": 1.72, |
| "grad_norm": 6.4375, |
| "learning_rate": 3.125628140703518e-08, |
| "loss": 0.9076, |
| "step": 68900 |
| }, |
| { |
| "epoch": 1.73, |
| "grad_norm": 5.34375, |
| "learning_rate": 3.1155778894472364e-08, |
| "loss": 0.8654, |
| "step": 69000 |
| }, |
| { |
| "epoch": 1.73, |
| "grad_norm": 6.5625, |
| "learning_rate": 3.105527638190955e-08, |
| "loss": 0.9213, |
| "step": 69100 |
| }, |
| { |
| "epoch": 1.73, |
| "grad_norm": 8.375, |
| "learning_rate": 3.095477386934673e-08, |
| "loss": 0.9055, |
| "step": 69200 |
| }, |
| { |
| "epoch": 1.73, |
| "grad_norm": 7.53125, |
| "learning_rate": 3.085427135678392e-08, |
| "loss": 0.9347, |
| "step": 69300 |
| }, |
| { |
| "epoch": 1.73, |
| "grad_norm": 5.9375, |
| "learning_rate": 3.0753768844221106e-08, |
| "loss": 1.0156, |
| "step": 69400 |
| }, |
| { |
| "epoch": 1.74, |
| "grad_norm": 6.15625, |
| "learning_rate": 3.0653266331658294e-08, |
| "loss": 0.9217, |
| "step": 69500 |
| }, |
| { |
| "epoch": 1.74, |
| "grad_norm": 6.6875, |
| "learning_rate": 3.0552763819095474e-08, |
| "loss": 0.9123, |
| "step": 69600 |
| }, |
| { |
| "epoch": 1.74, |
| "grad_norm": 10.8125, |
| "learning_rate": 3.045226130653266e-08, |
| "loss": 0.9781, |
| "step": 69700 |
| }, |
| { |
| "epoch": 1.75, |
| "grad_norm": 4.96875, |
| "learning_rate": 3.035175879396985e-08, |
| "loss": 0.9449, |
| "step": 69800 |
| }, |
| { |
| "epoch": 1.75, |
| "grad_norm": 4.75, |
| "learning_rate": 3.0251256281407036e-08, |
| "loss": 1.0215, |
| "step": 69900 |
| }, |
| { |
| "epoch": 1.75, |
| "grad_norm": 6.21875, |
| "learning_rate": 3.015075376884422e-08, |
| "loss": 0.9357, |
| "step": 70000 |
| }, |
| { |
| "epoch": 1.75, |
| "eval_loss": 0.9999986290931702, |
| "eval_runtime": 60.0527, |
| "eval_samples_per_second": 16.652, |
| "eval_steps_per_second": 4.163, |
| "step": 70000 |
| }, |
| { |
| "epoch": 1.75, |
| "grad_norm": 7.78125, |
| "learning_rate": 3.0050251256281404e-08, |
| "loss": 0.9616, |
| "step": 70100 |
| }, |
| { |
| "epoch": 1.75, |
| "grad_norm": 5.46875, |
| "learning_rate": 2.994974874371859e-08, |
| "loss": 0.9282, |
| "step": 70200 |
| }, |
| { |
| "epoch": 1.76, |
| "grad_norm": 7.0625, |
| "learning_rate": 2.984924623115578e-08, |
| "loss": 0.9819, |
| "step": 70300 |
| }, |
| { |
| "epoch": 1.76, |
| "grad_norm": 11.6875, |
| "learning_rate": 2.9748743718592965e-08, |
| "loss": 1.0126, |
| "step": 70400 |
| }, |
| { |
| "epoch": 1.76, |
| "grad_norm": 6.53125, |
| "learning_rate": 2.964824120603015e-08, |
| "loss": 0.952, |
| "step": 70500 |
| }, |
| { |
| "epoch": 1.77, |
| "grad_norm": 6.71875, |
| "learning_rate": 2.9547738693467337e-08, |
| "loss": 0.9167, |
| "step": 70600 |
| }, |
| { |
| "epoch": 1.77, |
| "grad_norm": 4.875, |
| "learning_rate": 2.944723618090452e-08, |
| "loss": 0.959, |
| "step": 70700 |
| }, |
| { |
| "epoch": 1.77, |
| "grad_norm": 6.78125, |
| "learning_rate": 2.9346733668341708e-08, |
| "loss": 0.9443, |
| "step": 70800 |
| }, |
| { |
| "epoch": 1.77, |
| "grad_norm": 6.125, |
| "learning_rate": 2.924623115577889e-08, |
| "loss": 0.9041, |
| "step": 70900 |
| }, |
| { |
| "epoch": 1.77, |
| "grad_norm": 10.25, |
| "learning_rate": 2.914572864321608e-08, |
| "loss": 0.9468, |
| "step": 71000 |
| }, |
| { |
| "epoch": 1.78, |
| "grad_norm": 5.65625, |
| "learning_rate": 2.9045226130653266e-08, |
| "loss": 0.9685, |
| "step": 71100 |
| }, |
| { |
| "epoch": 1.78, |
| "grad_norm": 9.0625, |
| "learning_rate": 2.894472361809045e-08, |
| "loss": 0.9723, |
| "step": 71200 |
| }, |
| { |
| "epoch": 1.78, |
| "grad_norm": 10.1875, |
| "learning_rate": 2.8844221105527637e-08, |
| "loss": 0.9101, |
| "step": 71300 |
| }, |
| { |
| "epoch": 1.79, |
| "grad_norm": 7.0625, |
| "learning_rate": 2.874371859296482e-08, |
| "loss": 0.9806, |
| "step": 71400 |
| }, |
| { |
| "epoch": 1.79, |
| "grad_norm": 11.4375, |
| "learning_rate": 2.8643216080402008e-08, |
| "loss": 1.0012, |
| "step": 71500 |
| }, |
| { |
| "epoch": 1.79, |
| "grad_norm": 6.6875, |
| "learning_rate": 2.8542713567839196e-08, |
| "loss": 0.9899, |
| "step": 71600 |
| }, |
| { |
| "epoch": 1.79, |
| "grad_norm": 6.6875, |
| "learning_rate": 2.844221105527638e-08, |
| "loss": 0.9397, |
| "step": 71700 |
| }, |
| { |
| "epoch": 1.79, |
| "grad_norm": 5.46875, |
| "learning_rate": 2.8341708542713567e-08, |
| "loss": 0.982, |
| "step": 71800 |
| }, |
| { |
| "epoch": 1.8, |
| "grad_norm": 6.9375, |
| "learning_rate": 2.8241206030150757e-08, |
| "loss": 0.9602, |
| "step": 71900 |
| }, |
| { |
| "epoch": 1.8, |
| "grad_norm": 6.90625, |
| "learning_rate": 2.8140703517587944e-08, |
| "loss": 0.9691, |
| "step": 72000 |
| }, |
| { |
| "epoch": 1.8, |
| "grad_norm": 5.375, |
| "learning_rate": 2.8040201005025128e-08, |
| "loss": 0.9466, |
| "step": 72100 |
| }, |
| { |
| "epoch": 1.81, |
| "grad_norm": 7.375, |
| "learning_rate": 2.7939698492462316e-08, |
| "loss": 1.0198, |
| "step": 72200 |
| }, |
| { |
| "epoch": 1.81, |
| "grad_norm": 6.5625, |
| "learning_rate": 2.78391959798995e-08, |
| "loss": 0.9378, |
| "step": 72300 |
| }, |
| { |
| "epoch": 1.81, |
| "grad_norm": 5.34375, |
| "learning_rate": 2.7738693467336687e-08, |
| "loss": 0.9785, |
| "step": 72400 |
| }, |
| { |
| "epoch": 1.81, |
| "grad_norm": 5.1875, |
| "learning_rate": 2.763819095477387e-08, |
| "loss": 0.9804, |
| "step": 72500 |
| }, |
| { |
| "epoch": 1.81, |
| "grad_norm": 9.3125, |
| "learning_rate": 2.7537688442211058e-08, |
| "loss": 0.9293, |
| "step": 72600 |
| }, |
| { |
| "epoch": 1.82, |
| "grad_norm": 5.71875, |
| "learning_rate": 2.7437185929648245e-08, |
| "loss": 0.8982, |
| "step": 72700 |
| }, |
| { |
| "epoch": 1.82, |
| "grad_norm": 7.34375, |
| "learning_rate": 2.733668341708543e-08, |
| "loss": 0.9553, |
| "step": 72800 |
| }, |
| { |
| "epoch": 1.82, |
| "grad_norm": 5.6875, |
| "learning_rate": 2.7236180904522616e-08, |
| "loss": 0.9614, |
| "step": 72900 |
| }, |
| { |
| "epoch": 1.82, |
| "grad_norm": 5.9375, |
| "learning_rate": 2.71356783919598e-08, |
| "loss": 0.9744, |
| "step": 73000 |
| }, |
| { |
| "epoch": 1.83, |
| "grad_norm": 4.40625, |
| "learning_rate": 2.7035175879396987e-08, |
| "loss": 0.9284, |
| "step": 73100 |
| }, |
| { |
| "epoch": 1.83, |
| "grad_norm": 4.40625, |
| "learning_rate": 2.6934673366834175e-08, |
| "loss": 0.9075, |
| "step": 73200 |
| }, |
| { |
| "epoch": 1.83, |
| "grad_norm": 8.1875, |
| "learning_rate": 2.683417085427136e-08, |
| "loss": 0.9448, |
| "step": 73300 |
| }, |
| { |
| "epoch": 1.83, |
| "grad_norm": 4.9375, |
| "learning_rate": 2.6733668341708546e-08, |
| "loss": 0.9215, |
| "step": 73400 |
| }, |
| { |
| "epoch": 1.84, |
| "grad_norm": 5.46875, |
| "learning_rate": 2.663316582914573e-08, |
| "loss": 0.9815, |
| "step": 73500 |
| }, |
| { |
| "epoch": 1.84, |
| "grad_norm": 3.90625, |
| "learning_rate": 2.6532663316582917e-08, |
| "loss": 0.9832, |
| "step": 73600 |
| }, |
| { |
| "epoch": 1.84, |
| "grad_norm": 8.625, |
| "learning_rate": 2.64321608040201e-08, |
| "loss": 0.9929, |
| "step": 73700 |
| }, |
| { |
| "epoch": 1.84, |
| "grad_norm": 10.125, |
| "learning_rate": 2.6331658291457288e-08, |
| "loss": 0.9937, |
| "step": 73800 |
| }, |
| { |
| "epoch": 1.85, |
| "grad_norm": 6.40625, |
| "learning_rate": 2.6231155778894475e-08, |
| "loss": 0.9443, |
| "step": 73900 |
| }, |
| { |
| "epoch": 1.85, |
| "grad_norm": 6.3125, |
| "learning_rate": 2.613065326633166e-08, |
| "loss": 0.9822, |
| "step": 74000 |
| }, |
| { |
| "epoch": 1.85, |
| "grad_norm": 6.65625, |
| "learning_rate": 2.6030150753768846e-08, |
| "loss": 0.9814, |
| "step": 74100 |
| }, |
| { |
| "epoch": 1.85, |
| "grad_norm": 8.125, |
| "learning_rate": 2.592964824120603e-08, |
| "loss": 0.9854, |
| "step": 74200 |
| }, |
| { |
| "epoch": 1.86, |
| "grad_norm": 7.40625, |
| "learning_rate": 2.5829145728643218e-08, |
| "loss": 0.9702, |
| "step": 74300 |
| }, |
| { |
| "epoch": 1.86, |
| "grad_norm": 9.25, |
| "learning_rate": 2.5728643216080405e-08, |
| "loss": 1.0065, |
| "step": 74400 |
| }, |
| { |
| "epoch": 1.86, |
| "grad_norm": 12.0625, |
| "learning_rate": 2.562814070351759e-08, |
| "loss": 0.9264, |
| "step": 74500 |
| }, |
| { |
| "epoch": 1.86, |
| "grad_norm": 7.0625, |
| "learning_rate": 2.5527638190954776e-08, |
| "loss": 0.9612, |
| "step": 74600 |
| }, |
| { |
| "epoch": 1.87, |
| "grad_norm": 5.125, |
| "learning_rate": 2.542713567839196e-08, |
| "loss": 0.9377, |
| "step": 74700 |
| }, |
| { |
| "epoch": 1.87, |
| "grad_norm": 4.6875, |
| "learning_rate": 2.5326633165829147e-08, |
| "loss": 0.9855, |
| "step": 74800 |
| }, |
| { |
| "epoch": 1.87, |
| "grad_norm": 7.46875, |
| "learning_rate": 2.522613065326633e-08, |
| "loss": 0.9767, |
| "step": 74900 |
| }, |
| { |
| "epoch": 1.88, |
| "grad_norm": 8.375, |
| "learning_rate": 2.5125628140703518e-08, |
| "loss": 0.913, |
| "step": 75000 |
| }, |
| { |
| "epoch": 1.88, |
| "grad_norm": 5.78125, |
| "learning_rate": 2.5025125628140705e-08, |
| "loss": 0.957, |
| "step": 75100 |
| }, |
| { |
| "epoch": 1.88, |
| "grad_norm": 7.71875, |
| "learning_rate": 2.492462311557789e-08, |
| "loss": 0.9567, |
| "step": 75200 |
| }, |
| { |
| "epoch": 1.88, |
| "grad_norm": 5.0, |
| "learning_rate": 2.4824120603015077e-08, |
| "loss": 0.9379, |
| "step": 75300 |
| }, |
| { |
| "epoch": 1.89, |
| "grad_norm": 6.5625, |
| "learning_rate": 2.472361809045226e-08, |
| "loss": 0.978, |
| "step": 75400 |
| }, |
| { |
| "epoch": 1.89, |
| "grad_norm": 5.625, |
| "learning_rate": 2.4623115577889448e-08, |
| "loss": 0.9688, |
| "step": 75500 |
| }, |
| { |
| "epoch": 1.89, |
| "grad_norm": 6.71875, |
| "learning_rate": 2.452261306532663e-08, |
| "loss": 0.9691, |
| "step": 75600 |
| }, |
| { |
| "epoch": 1.89, |
| "grad_norm": 5.09375, |
| "learning_rate": 2.442211055276382e-08, |
| "loss": 0.9414, |
| "step": 75700 |
| }, |
| { |
| "epoch": 1.9, |
| "grad_norm": 9.6875, |
| "learning_rate": 2.4321608040201006e-08, |
| "loss": 0.8994, |
| "step": 75800 |
| }, |
| { |
| "epoch": 1.9, |
| "grad_norm": 6.5625, |
| "learning_rate": 2.422110552763819e-08, |
| "loss": 0.8903, |
| "step": 75900 |
| }, |
| { |
| "epoch": 1.9, |
| "grad_norm": 6.65625, |
| "learning_rate": 2.4120603015075377e-08, |
| "loss": 0.9246, |
| "step": 76000 |
| }, |
| { |
| "epoch": 1.9, |
| "grad_norm": 5.90625, |
| "learning_rate": 2.402010050251256e-08, |
| "loss": 0.9351, |
| "step": 76100 |
| }, |
| { |
| "epoch": 1.91, |
| "grad_norm": 6.3125, |
| "learning_rate": 2.3919597989949748e-08, |
| "loss": 0.9671, |
| "step": 76200 |
| }, |
| { |
| "epoch": 1.91, |
| "grad_norm": 5.53125, |
| "learning_rate": 2.3819095477386936e-08, |
| "loss": 0.9706, |
| "step": 76300 |
| }, |
| { |
| "epoch": 1.91, |
| "grad_norm": 5.5625, |
| "learning_rate": 2.371859296482412e-08, |
| "loss": 0.9419, |
| "step": 76400 |
| }, |
| { |
| "epoch": 1.91, |
| "grad_norm": 5.09375, |
| "learning_rate": 2.3618090452261307e-08, |
| "loss": 0.9739, |
| "step": 76500 |
| }, |
| { |
| "epoch": 1.92, |
| "grad_norm": 5.375, |
| "learning_rate": 2.3517587939698494e-08, |
| "loss": 0.9722, |
| "step": 76600 |
| }, |
| { |
| "epoch": 1.92, |
| "grad_norm": 5.71875, |
| "learning_rate": 2.341708542713568e-08, |
| "loss": 0.9212, |
| "step": 76700 |
| }, |
| { |
| "epoch": 1.92, |
| "grad_norm": 5.8125, |
| "learning_rate": 2.3316582914572865e-08, |
| "loss": 0.8863, |
| "step": 76800 |
| }, |
| { |
| "epoch": 1.92, |
| "grad_norm": 9.0, |
| "learning_rate": 2.3216080402010052e-08, |
| "loss": 0.9425, |
| "step": 76900 |
| }, |
| { |
| "epoch": 1.93, |
| "grad_norm": 6.65625, |
| "learning_rate": 2.3115577889447236e-08, |
| "loss": 0.9497, |
| "step": 77000 |
| }, |
| { |
| "epoch": 1.93, |
| "grad_norm": 5.28125, |
| "learning_rate": 2.3015075376884423e-08, |
| "loss": 0.933, |
| "step": 77100 |
| }, |
| { |
| "epoch": 1.93, |
| "grad_norm": 11.875, |
| "learning_rate": 2.291457286432161e-08, |
| "loss": 0.9661, |
| "step": 77200 |
| }, |
| { |
| "epoch": 1.93, |
| "grad_norm": 7.875, |
| "learning_rate": 2.2814070351758795e-08, |
| "loss": 0.9015, |
| "step": 77300 |
| }, |
| { |
| "epoch": 1.94, |
| "grad_norm": 6.59375, |
| "learning_rate": 2.2713567839195982e-08, |
| "loss": 0.9095, |
| "step": 77400 |
| }, |
| { |
| "epoch": 1.94, |
| "grad_norm": 7.21875, |
| "learning_rate": 2.2613065326633166e-08, |
| "loss": 0.9207, |
| "step": 77500 |
| }, |
| { |
| "epoch": 1.94, |
| "grad_norm": 4.375, |
| "learning_rate": 2.2512562814070353e-08, |
| "loss": 0.9315, |
| "step": 77600 |
| }, |
| { |
| "epoch": 1.94, |
| "grad_norm": 11.0, |
| "learning_rate": 2.241206030150754e-08, |
| "loss": 0.9018, |
| "step": 77700 |
| }, |
| { |
| "epoch": 1.94, |
| "grad_norm": 8.4375, |
| "learning_rate": 2.2311557788944724e-08, |
| "loss": 0.9298, |
| "step": 77800 |
| }, |
| { |
| "epoch": 1.95, |
| "grad_norm": 6.65625, |
| "learning_rate": 2.221105527638191e-08, |
| "loss": 0.9506, |
| "step": 77900 |
| }, |
| { |
| "epoch": 1.95, |
| "grad_norm": 9.3125, |
| "learning_rate": 2.2110552763819095e-08, |
| "loss": 0.9076, |
| "step": 78000 |
| }, |
| { |
| "epoch": 1.95, |
| "grad_norm": 7.5625, |
| "learning_rate": 2.2010050251256282e-08, |
| "loss": 1.0066, |
| "step": 78100 |
| }, |
| { |
| "epoch": 1.96, |
| "grad_norm": 5.78125, |
| "learning_rate": 2.1909547738693466e-08, |
| "loss": 0.972, |
| "step": 78200 |
| }, |
| { |
| "epoch": 1.96, |
| "grad_norm": 5.25, |
| "learning_rate": 2.1809045226130654e-08, |
| "loss": 0.9311, |
| "step": 78300 |
| }, |
| { |
| "epoch": 1.96, |
| "grad_norm": 4.84375, |
| "learning_rate": 2.170854271356784e-08, |
| "loss": 0.9604, |
| "step": 78400 |
| }, |
| { |
| "epoch": 1.96, |
| "grad_norm": 2.796875, |
| "learning_rate": 2.1608040201005025e-08, |
| "loss": 0.8835, |
| "step": 78500 |
| }, |
| { |
| "epoch": 1.96, |
| "grad_norm": 7.15625, |
| "learning_rate": 2.1507537688442212e-08, |
| "loss": 0.9863, |
| "step": 78600 |
| }, |
| { |
| "epoch": 1.97, |
| "grad_norm": 6.28125, |
| "learning_rate": 2.1407035175879396e-08, |
| "loss": 0.9428, |
| "step": 78700 |
| }, |
| { |
| "epoch": 1.97, |
| "grad_norm": 3.921875, |
| "learning_rate": 2.1306532663316583e-08, |
| "loss": 0.9335, |
| "step": 78800 |
| }, |
| { |
| "epoch": 1.97, |
| "grad_norm": 5.03125, |
| "learning_rate": 2.120603015075377e-08, |
| "loss": 1.0154, |
| "step": 78900 |
| }, |
| { |
| "epoch": 1.98, |
| "grad_norm": 6.15625, |
| "learning_rate": 2.1105527638190954e-08, |
| "loss": 0.9613, |
| "step": 79000 |
| }, |
| { |
| "epoch": 1.98, |
| "grad_norm": 5.21875, |
| "learning_rate": 2.100502512562814e-08, |
| "loss": 0.9645, |
| "step": 79100 |
| }, |
| { |
| "epoch": 1.98, |
| "grad_norm": 9.875, |
| "learning_rate": 2.0904522613065325e-08, |
| "loss": 0.9791, |
| "step": 79200 |
| }, |
| { |
| "epoch": 1.98, |
| "grad_norm": 10.3125, |
| "learning_rate": 2.0804020100502513e-08, |
| "loss": 0.9765, |
| "step": 79300 |
| }, |
| { |
| "epoch": 1.98, |
| "grad_norm": 6.28125, |
| "learning_rate": 2.0703517587939696e-08, |
| "loss": 0.9877, |
| "step": 79400 |
| }, |
| { |
| "epoch": 1.99, |
| "grad_norm": 6.71875, |
| "learning_rate": 2.0603015075376884e-08, |
| "loss": 0.9844, |
| "step": 79500 |
| }, |
| { |
| "epoch": 1.99, |
| "grad_norm": 7.21875, |
| "learning_rate": 2.050251256281407e-08, |
| "loss": 0.9206, |
| "step": 79600 |
| }, |
| { |
| "epoch": 1.99, |
| "grad_norm": 7.34375, |
| "learning_rate": 2.0402010050251258e-08, |
| "loss": 0.9705, |
| "step": 79700 |
| }, |
| { |
| "epoch": 2.0, |
| "grad_norm": 6.9375, |
| "learning_rate": 2.0301507537688445e-08, |
| "loss": 0.972, |
| "step": 79800 |
| }, |
| { |
| "epoch": 2.0, |
| "grad_norm": 4.5625, |
| "learning_rate": 2.020100502512563e-08, |
| "loss": 0.9654, |
| "step": 79900 |
| }, |
| { |
| "epoch": 2.0, |
| "grad_norm": 7.0625, |
| "learning_rate": 2.0100502512562817e-08, |
| "loss": 0.9442, |
| "step": 80000 |
| }, |
| { |
| "epoch": 2.0, |
| "eval_loss": 1.016454815864563, |
| "eval_runtime": 60.0142, |
| "eval_samples_per_second": 16.663, |
| "eval_steps_per_second": 4.166, |
| "step": 80000 |
| }, |
| { |
| "epoch": 2.0, |
| "grad_norm": 6.28125, |
| "learning_rate": 2e-08, |
| "loss": 0.9476, |
| "step": 80100 |
| }, |
| { |
| "epoch": 2.0, |
| "grad_norm": 5.0625, |
| "learning_rate": 1.9899497487437188e-08, |
| "loss": 0.9094, |
| "step": 80200 |
| }, |
| { |
| "epoch": 2.01, |
| "grad_norm": 6.1875, |
| "learning_rate": 1.979899497487437e-08, |
| "loss": 0.955, |
| "step": 80300 |
| }, |
| { |
| "epoch": 2.01, |
| "grad_norm": 6.125, |
| "learning_rate": 1.969849246231156e-08, |
| "loss": 0.9935, |
| "step": 80400 |
| }, |
| { |
| "epoch": 2.01, |
| "grad_norm": 8.125, |
| "learning_rate": 1.9597989949748746e-08, |
| "loss": 0.9063, |
| "step": 80500 |
| }, |
| { |
| "epoch": 2.02, |
| "grad_norm": 6.5625, |
| "learning_rate": 1.949748743718593e-08, |
| "loss": 0.9381, |
| "step": 80600 |
| }, |
| { |
| "epoch": 2.02, |
| "grad_norm": 10.125, |
| "learning_rate": 1.9396984924623117e-08, |
| "loss": 0.9802, |
| "step": 80700 |
| }, |
| { |
| "epoch": 2.02, |
| "grad_norm": 5.5, |
| "learning_rate": 1.92964824120603e-08, |
| "loss": 0.9377, |
| "step": 80800 |
| }, |
| { |
| "epoch": 2.02, |
| "grad_norm": 6.09375, |
| "learning_rate": 1.9195979899497488e-08, |
| "loss": 0.8899, |
| "step": 80900 |
| }, |
| { |
| "epoch": 2.02, |
| "grad_norm": 5.75, |
| "learning_rate": 1.9095477386934676e-08, |
| "loss": 0.8906, |
| "step": 81000 |
| }, |
| { |
| "epoch": 2.03, |
| "grad_norm": 4.90625, |
| "learning_rate": 1.899497487437186e-08, |
| "loss": 0.9292, |
| "step": 81100 |
| }, |
| { |
| "epoch": 2.03, |
| "grad_norm": 9.8125, |
| "learning_rate": 1.8894472361809047e-08, |
| "loss": 0.8881, |
| "step": 81200 |
| }, |
| { |
| "epoch": 2.03, |
| "grad_norm": 7.3125, |
| "learning_rate": 1.879396984924623e-08, |
| "loss": 0.9303, |
| "step": 81300 |
| }, |
| { |
| "epoch": 2.04, |
| "grad_norm": 4.25, |
| "learning_rate": 1.8693467336683418e-08, |
| "loss": 1.0181, |
| "step": 81400 |
| }, |
| { |
| "epoch": 2.04, |
| "grad_norm": 4.9375, |
| "learning_rate": 1.8592964824120602e-08, |
| "loss": 0.9946, |
| "step": 81500 |
| }, |
| { |
| "epoch": 2.04, |
| "grad_norm": 5.59375, |
| "learning_rate": 1.849246231155779e-08, |
| "loss": 0.9528, |
| "step": 81600 |
| }, |
| { |
| "epoch": 2.04, |
| "grad_norm": 6.75, |
| "learning_rate": 1.8391959798994976e-08, |
| "loss": 0.9546, |
| "step": 81700 |
| }, |
| { |
| "epoch": 2.04, |
| "grad_norm": 3.875, |
| "learning_rate": 1.829145728643216e-08, |
| "loss": 0.9268, |
| "step": 81800 |
| }, |
| { |
| "epoch": 2.05, |
| "grad_norm": 4.34375, |
| "learning_rate": 1.8190954773869347e-08, |
| "loss": 0.9261, |
| "step": 81900 |
| }, |
| { |
| "epoch": 2.05, |
| "grad_norm": 10.9375, |
| "learning_rate": 1.809045226130653e-08, |
| "loss": 1.0338, |
| "step": 82000 |
| }, |
| { |
| "epoch": 2.05, |
| "grad_norm": 6.21875, |
| "learning_rate": 1.798994974874372e-08, |
| "loss": 1.0056, |
| "step": 82100 |
| }, |
| { |
| "epoch": 2.06, |
| "grad_norm": 4.15625, |
| "learning_rate": 1.7889447236180906e-08, |
| "loss": 0.9034, |
| "step": 82200 |
| }, |
| { |
| "epoch": 2.06, |
| "grad_norm": 9.9375, |
| "learning_rate": 1.778894472361809e-08, |
| "loss": 0.9524, |
| "step": 82300 |
| }, |
| { |
| "epoch": 2.06, |
| "grad_norm": 8.625, |
| "learning_rate": 1.7688442211055277e-08, |
| "loss": 0.9035, |
| "step": 82400 |
| }, |
| { |
| "epoch": 2.06, |
| "grad_norm": 5.78125, |
| "learning_rate": 1.758793969849246e-08, |
| "loss": 0.9823, |
| "step": 82500 |
| }, |
| { |
| "epoch": 2.06, |
| "grad_norm": 6.5, |
| "learning_rate": 1.7487437185929648e-08, |
| "loss": 0.9044, |
| "step": 82600 |
| }, |
| { |
| "epoch": 2.07, |
| "grad_norm": 4.9375, |
| "learning_rate": 1.7386934673366832e-08, |
| "loss": 0.8919, |
| "step": 82700 |
| }, |
| { |
| "epoch": 2.07, |
| "grad_norm": 4.625, |
| "learning_rate": 1.728643216080402e-08, |
| "loss": 0.9323, |
| "step": 82800 |
| }, |
| { |
| "epoch": 2.07, |
| "grad_norm": 4.875, |
| "learning_rate": 1.7185929648241206e-08, |
| "loss": 0.9792, |
| "step": 82900 |
| }, |
| { |
| "epoch": 2.08, |
| "grad_norm": 6.03125, |
| "learning_rate": 1.7085427135678394e-08, |
| "loss": 0.9269, |
| "step": 83000 |
| }, |
| { |
| "epoch": 2.08, |
| "grad_norm": 4.5, |
| "learning_rate": 1.698492462311558e-08, |
| "loss": 0.9527, |
| "step": 83100 |
| }, |
| { |
| "epoch": 2.08, |
| "grad_norm": 6.5625, |
| "learning_rate": 1.6884422110552765e-08, |
| "loss": 0.9316, |
| "step": 83200 |
| }, |
| { |
| "epoch": 2.08, |
| "grad_norm": 7.09375, |
| "learning_rate": 1.6783919597989952e-08, |
| "loss": 0.8942, |
| "step": 83300 |
| }, |
| { |
| "epoch": 2.08, |
| "grad_norm": 5.34375, |
| "learning_rate": 1.6683417085427136e-08, |
| "loss": 0.9274, |
| "step": 83400 |
| }, |
| { |
| "epoch": 2.09, |
| "grad_norm": 5.25, |
| "learning_rate": 1.6582914572864323e-08, |
| "loss": 0.9712, |
| "step": 83500 |
| }, |
| { |
| "epoch": 2.09, |
| "grad_norm": 7.96875, |
| "learning_rate": 1.6482412060301507e-08, |
| "loss": 0.931, |
| "step": 83600 |
| }, |
| { |
| "epoch": 2.09, |
| "grad_norm": 4.125, |
| "learning_rate": 1.6381909547738694e-08, |
| "loss": 0.8989, |
| "step": 83700 |
| }, |
| { |
| "epoch": 2.1, |
| "grad_norm": 6.4375, |
| "learning_rate": 1.628140703517588e-08, |
| "loss": 0.9929, |
| "step": 83800 |
| }, |
| { |
| "epoch": 2.1, |
| "grad_norm": 7.65625, |
| "learning_rate": 1.6180904522613065e-08, |
| "loss": 0.9448, |
| "step": 83900 |
| }, |
| { |
| "epoch": 2.1, |
| "grad_norm": 7.40625, |
| "learning_rate": 1.6080402010050253e-08, |
| "loss": 0.9364, |
| "step": 84000 |
| }, |
| { |
| "epoch": 2.1, |
| "grad_norm": 7.59375, |
| "learning_rate": 1.5979899497487436e-08, |
| "loss": 0.9086, |
| "step": 84100 |
| }, |
| { |
| "epoch": 2.1, |
| "grad_norm": 5.71875, |
| "learning_rate": 1.5879396984924624e-08, |
| "loss": 0.9385, |
| "step": 84200 |
| }, |
| { |
| "epoch": 2.11, |
| "grad_norm": 5.375, |
| "learning_rate": 1.577889447236181e-08, |
| "loss": 0.9125, |
| "step": 84300 |
| }, |
| { |
| "epoch": 2.11, |
| "grad_norm": 6.5, |
| "learning_rate": 1.5678391959798995e-08, |
| "loss": 0.9316, |
| "step": 84400 |
| }, |
| { |
| "epoch": 2.11, |
| "grad_norm": 5.28125, |
| "learning_rate": 1.5577889447236182e-08, |
| "loss": 0.9358, |
| "step": 84500 |
| }, |
| { |
| "epoch": 2.12, |
| "grad_norm": 7.625, |
| "learning_rate": 1.5477386934673366e-08, |
| "loss": 0.9307, |
| "step": 84600 |
| }, |
| { |
| "epoch": 2.12, |
| "grad_norm": 5.96875, |
| "learning_rate": 1.5376884422110553e-08, |
| "loss": 0.913, |
| "step": 84700 |
| }, |
| { |
| "epoch": 2.12, |
| "grad_norm": 6.71875, |
| "learning_rate": 1.5276381909547737e-08, |
| "loss": 0.8759, |
| "step": 84800 |
| }, |
| { |
| "epoch": 2.12, |
| "grad_norm": 6.03125, |
| "learning_rate": 1.5175879396984924e-08, |
| "loss": 0.9576, |
| "step": 84900 |
| }, |
| { |
| "epoch": 2.12, |
| "grad_norm": 7.28125, |
| "learning_rate": 1.507537688442211e-08, |
| "loss": 0.9269, |
| "step": 85000 |
| }, |
| { |
| "epoch": 2.13, |
| "grad_norm": 5.625, |
| "learning_rate": 1.4974874371859295e-08, |
| "loss": 0.9792, |
| "step": 85100 |
| }, |
| { |
| "epoch": 2.13, |
| "grad_norm": 6.21875, |
| "learning_rate": 1.4874371859296483e-08, |
| "loss": 0.9723, |
| "step": 85200 |
| }, |
| { |
| "epoch": 2.13, |
| "grad_norm": 6.71875, |
| "learning_rate": 1.4773869346733668e-08, |
| "loss": 0.9035, |
| "step": 85300 |
| }, |
| { |
| "epoch": 2.13, |
| "grad_norm": 6.5625, |
| "learning_rate": 1.4673366834170854e-08, |
| "loss": 0.9527, |
| "step": 85400 |
| }, |
| { |
| "epoch": 2.14, |
| "grad_norm": 5.75, |
| "learning_rate": 1.457286432160804e-08, |
| "loss": 0.9798, |
| "step": 85500 |
| }, |
| { |
| "epoch": 2.14, |
| "grad_norm": 5.3125, |
| "learning_rate": 1.4472361809045225e-08, |
| "loss": 0.9221, |
| "step": 85600 |
| }, |
| { |
| "epoch": 2.14, |
| "grad_norm": 6.9375, |
| "learning_rate": 1.437185929648241e-08, |
| "loss": 0.9874, |
| "step": 85700 |
| }, |
| { |
| "epoch": 2.15, |
| "grad_norm": 8.1875, |
| "learning_rate": 1.4271356783919598e-08, |
| "loss": 0.93, |
| "step": 85800 |
| }, |
| { |
| "epoch": 2.15, |
| "grad_norm": 5.6875, |
| "learning_rate": 1.4170854271356783e-08, |
| "loss": 0.8976, |
| "step": 85900 |
| }, |
| { |
| "epoch": 2.15, |
| "grad_norm": 9.1875, |
| "learning_rate": 1.4070351758793972e-08, |
| "loss": 0.9174, |
| "step": 86000 |
| }, |
| { |
| "epoch": 2.15, |
| "grad_norm": 6.96875, |
| "learning_rate": 1.3969849246231158e-08, |
| "loss": 0.9544, |
| "step": 86100 |
| }, |
| { |
| "epoch": 2.15, |
| "grad_norm": 8.875, |
| "learning_rate": 1.3869346733668343e-08, |
| "loss": 0.9221, |
| "step": 86200 |
| }, |
| { |
| "epoch": 2.16, |
| "grad_norm": 7.21875, |
| "learning_rate": 1.3768844221105529e-08, |
| "loss": 0.972, |
| "step": 86300 |
| }, |
| { |
| "epoch": 2.16, |
| "grad_norm": 7.1875, |
| "learning_rate": 1.3668341708542714e-08, |
| "loss": 0.9527, |
| "step": 86400 |
| }, |
| { |
| "epoch": 2.16, |
| "grad_norm": 6.46875, |
| "learning_rate": 1.35678391959799e-08, |
| "loss": 0.9785, |
| "step": 86500 |
| }, |
| { |
| "epoch": 2.17, |
| "grad_norm": 5.53125, |
| "learning_rate": 1.3467336683417087e-08, |
| "loss": 0.9369, |
| "step": 86600 |
| }, |
| { |
| "epoch": 2.17, |
| "grad_norm": 5.21875, |
| "learning_rate": 1.3366834170854273e-08, |
| "loss": 0.9371, |
| "step": 86700 |
| }, |
| { |
| "epoch": 2.17, |
| "grad_norm": 5.5625, |
| "learning_rate": 1.3266331658291458e-08, |
| "loss": 0.9412, |
| "step": 86800 |
| }, |
| { |
| "epoch": 2.17, |
| "grad_norm": 6.8125, |
| "learning_rate": 1.3165829145728644e-08, |
| "loss": 0.9501, |
| "step": 86900 |
| }, |
| { |
| "epoch": 2.17, |
| "grad_norm": 6.75, |
| "learning_rate": 1.306532663316583e-08, |
| "loss": 0.8898, |
| "step": 87000 |
| }, |
| { |
| "epoch": 2.18, |
| "grad_norm": 5.84375, |
| "learning_rate": 1.2964824120603015e-08, |
| "loss": 0.9686, |
| "step": 87100 |
| }, |
| { |
| "epoch": 2.18, |
| "grad_norm": 7.53125, |
| "learning_rate": 1.2864321608040202e-08, |
| "loss": 0.8871, |
| "step": 87200 |
| }, |
| { |
| "epoch": 2.18, |
| "grad_norm": 5.25, |
| "learning_rate": 1.2763819095477388e-08, |
| "loss": 0.9987, |
| "step": 87300 |
| }, |
| { |
| "epoch": 2.19, |
| "grad_norm": 8.0, |
| "learning_rate": 1.2663316582914574e-08, |
| "loss": 0.9803, |
| "step": 87400 |
| }, |
| { |
| "epoch": 2.19, |
| "grad_norm": 4.1875, |
| "learning_rate": 1.2562814070351759e-08, |
| "loss": 0.8711, |
| "step": 87500 |
| }, |
| { |
| "epoch": 2.19, |
| "grad_norm": 4.71875, |
| "learning_rate": 1.2462311557788945e-08, |
| "loss": 0.985, |
| "step": 87600 |
| }, |
| { |
| "epoch": 2.19, |
| "grad_norm": 5.53125, |
| "learning_rate": 1.236180904522613e-08, |
| "loss": 1.0226, |
| "step": 87700 |
| }, |
| { |
| "epoch": 2.19, |
| "grad_norm": 6.5, |
| "learning_rate": 1.2261306532663316e-08, |
| "loss": 0.9716, |
| "step": 87800 |
| }, |
| { |
| "epoch": 2.2, |
| "grad_norm": 7.71875, |
| "learning_rate": 1.2160804020100503e-08, |
| "loss": 0.8701, |
| "step": 87900 |
| }, |
| { |
| "epoch": 2.2, |
| "grad_norm": 5.4375, |
| "learning_rate": 1.2060301507537689e-08, |
| "loss": 0.9632, |
| "step": 88000 |
| }, |
| { |
| "epoch": 2.2, |
| "grad_norm": 7.0, |
| "learning_rate": 1.1959798994974874e-08, |
| "loss": 0.9337, |
| "step": 88100 |
| }, |
| { |
| "epoch": 2.21, |
| "grad_norm": 5.65625, |
| "learning_rate": 1.185929648241206e-08, |
| "loss": 0.9862, |
| "step": 88200 |
| }, |
| { |
| "epoch": 2.21, |
| "grad_norm": 4.125, |
| "learning_rate": 1.1758793969849247e-08, |
| "loss": 0.9775, |
| "step": 88300 |
| }, |
| { |
| "epoch": 2.21, |
| "grad_norm": 6.34375, |
| "learning_rate": 1.1658291457286433e-08, |
| "loss": 0.9627, |
| "step": 88400 |
| }, |
| { |
| "epoch": 2.21, |
| "grad_norm": 9.0625, |
| "learning_rate": 1.1557788944723618e-08, |
| "loss": 0.9698, |
| "step": 88500 |
| }, |
| { |
| "epoch": 2.21, |
| "grad_norm": 4.625, |
| "learning_rate": 1.1457286432160805e-08, |
| "loss": 0.9363, |
| "step": 88600 |
| }, |
| { |
| "epoch": 2.22, |
| "grad_norm": 7.4375, |
| "learning_rate": 1.1356783919597991e-08, |
| "loss": 0.9233, |
| "step": 88700 |
| }, |
| { |
| "epoch": 2.22, |
| "grad_norm": 6.71875, |
| "learning_rate": 1.1256281407035176e-08, |
| "loss": 0.9362, |
| "step": 88800 |
| }, |
| { |
| "epoch": 2.22, |
| "grad_norm": 7.9375, |
| "learning_rate": 1.1155778894472362e-08, |
| "loss": 0.9636, |
| "step": 88900 |
| }, |
| { |
| "epoch": 2.23, |
| "grad_norm": 6.3125, |
| "learning_rate": 1.1055276381909548e-08, |
| "loss": 0.9665, |
| "step": 89000 |
| }, |
| { |
| "epoch": 2.23, |
| "grad_norm": 6.0625, |
| "learning_rate": 1.0954773869346733e-08, |
| "loss": 0.9535, |
| "step": 89100 |
| }, |
| { |
| "epoch": 2.23, |
| "grad_norm": 6.90625, |
| "learning_rate": 1.085427135678392e-08, |
| "loss": 0.9221, |
| "step": 89200 |
| }, |
| { |
| "epoch": 2.23, |
| "grad_norm": 5.9375, |
| "learning_rate": 1.0753768844221106e-08, |
| "loss": 0.9877, |
| "step": 89300 |
| }, |
| { |
| "epoch": 2.23, |
| "grad_norm": 5.09375, |
| "learning_rate": 1.0653266331658292e-08, |
| "loss": 0.9351, |
| "step": 89400 |
| }, |
| { |
| "epoch": 2.24, |
| "grad_norm": 5.25, |
| "learning_rate": 1.0552763819095477e-08, |
| "loss": 0.9609, |
| "step": 89500 |
| }, |
| { |
| "epoch": 2.24, |
| "grad_norm": 4.71875, |
| "learning_rate": 1.0452261306532663e-08, |
| "loss": 1.0024, |
| "step": 89600 |
| }, |
| { |
| "epoch": 2.24, |
| "grad_norm": 7.0625, |
| "learning_rate": 1.0351758793969848e-08, |
| "loss": 0.905, |
| "step": 89700 |
| }, |
| { |
| "epoch": 2.25, |
| "grad_norm": 6.34375, |
| "learning_rate": 1.0251256281407035e-08, |
| "loss": 0.9278, |
| "step": 89800 |
| }, |
| { |
| "epoch": 2.25, |
| "grad_norm": 3.5, |
| "learning_rate": 1.0150753768844223e-08, |
| "loss": 0.9558, |
| "step": 89900 |
| }, |
| { |
| "epoch": 2.25, |
| "grad_norm": 5.15625, |
| "learning_rate": 1.0050251256281408e-08, |
| "loss": 0.8837, |
| "step": 90000 |
| }, |
| { |
| "epoch": 2.25, |
| "eval_loss": 1.0259228944778442, |
| "eval_runtime": 60.132, |
| "eval_samples_per_second": 16.63, |
| "eval_steps_per_second": 4.158, |
| "step": 90000 |
| }, |
| { |
| "epoch": 2.25, |
| "grad_norm": 7.28125, |
| "learning_rate": 9.949748743718594e-09, |
| "loss": 0.952, |
| "step": 90100 |
| }, |
| { |
| "epoch": 2.25, |
| "grad_norm": 5.90625, |
| "learning_rate": 9.84924623115578e-09, |
| "loss": 0.9146, |
| "step": 90200 |
| }, |
| { |
| "epoch": 2.26, |
| "grad_norm": 7.375, |
| "learning_rate": 9.748743718592965e-09, |
| "loss": 0.9258, |
| "step": 90300 |
| }, |
| { |
| "epoch": 2.26, |
| "grad_norm": 5.78125, |
| "learning_rate": 9.64824120603015e-09, |
| "loss": 0.929, |
| "step": 90400 |
| }, |
| { |
| "epoch": 2.26, |
| "grad_norm": 5.09375, |
| "learning_rate": 9.547738693467338e-09, |
| "loss": 0.9446, |
| "step": 90500 |
| }, |
| { |
| "epoch": 2.27, |
| "grad_norm": 7.03125, |
| "learning_rate": 9.447236180904523e-09, |
| "loss": 0.9348, |
| "step": 90600 |
| }, |
| { |
| "epoch": 2.27, |
| "grad_norm": 5.4375, |
| "learning_rate": 9.346733668341709e-09, |
| "loss": 0.949, |
| "step": 90700 |
| }, |
| { |
| "epoch": 2.27, |
| "grad_norm": 6.65625, |
| "learning_rate": 9.246231155778894e-09, |
| "loss": 0.9297, |
| "step": 90800 |
| }, |
| { |
| "epoch": 2.27, |
| "grad_norm": 5.0625, |
| "learning_rate": 9.14572864321608e-09, |
| "loss": 0.987, |
| "step": 90900 |
| }, |
| { |
| "epoch": 2.27, |
| "grad_norm": 5.9375, |
| "learning_rate": 9.045226130653266e-09, |
| "loss": 0.9568, |
| "step": 91000 |
| }, |
| { |
| "epoch": 2.28, |
| "grad_norm": 8.1875, |
| "learning_rate": 8.944723618090453e-09, |
| "loss": 0.961, |
| "step": 91100 |
| }, |
| { |
| "epoch": 2.28, |
| "grad_norm": 8.125, |
| "learning_rate": 8.844221105527638e-09, |
| "loss": 0.996, |
| "step": 91200 |
| }, |
| { |
| "epoch": 2.28, |
| "grad_norm": 8.1875, |
| "learning_rate": 8.743718592964824e-09, |
| "loss": 0.9385, |
| "step": 91300 |
| }, |
| { |
| "epoch": 2.29, |
| "grad_norm": 5.96875, |
| "learning_rate": 8.64321608040201e-09, |
| "loss": 0.9266, |
| "step": 91400 |
| }, |
| { |
| "epoch": 2.29, |
| "grad_norm": 5.3125, |
| "learning_rate": 8.542713567839197e-09, |
| "loss": 0.9684, |
| "step": 91500 |
| }, |
| { |
| "epoch": 2.29, |
| "grad_norm": 5.28125, |
| "learning_rate": 8.442211055276382e-09, |
| "loss": 0.9791, |
| "step": 91600 |
| }, |
| { |
| "epoch": 2.29, |
| "grad_norm": 5.84375, |
| "learning_rate": 8.341708542713568e-09, |
| "loss": 0.9238, |
| "step": 91700 |
| }, |
| { |
| "epoch": 2.29, |
| "grad_norm": 4.84375, |
| "learning_rate": 8.241206030150753e-09, |
| "loss": 0.9493, |
| "step": 91800 |
| }, |
| { |
| "epoch": 2.3, |
| "grad_norm": 5.625, |
| "learning_rate": 8.14070351758794e-09, |
| "loss": 0.96, |
| "step": 91900 |
| }, |
| { |
| "epoch": 2.3, |
| "grad_norm": 5.25, |
| "learning_rate": 8.040201005025126e-09, |
| "loss": 0.9639, |
| "step": 92000 |
| }, |
| { |
| "epoch": 2.3, |
| "grad_norm": 6.59375, |
| "learning_rate": 7.939698492462312e-09, |
| "loss": 0.9787, |
| "step": 92100 |
| }, |
| { |
| "epoch": 2.31, |
| "grad_norm": 8.6875, |
| "learning_rate": 7.839195979899497e-09, |
| "loss": 1.002, |
| "step": 92200 |
| }, |
| { |
| "epoch": 2.31, |
| "grad_norm": 7.34375, |
| "learning_rate": 7.738693467336683e-09, |
| "loss": 0.9659, |
| "step": 92300 |
| }, |
| { |
| "epoch": 2.31, |
| "grad_norm": 7.0, |
| "learning_rate": 7.638190954773869e-09, |
| "loss": 0.926, |
| "step": 92400 |
| }, |
| { |
| "epoch": 2.31, |
| "grad_norm": 6.28125, |
| "learning_rate": 7.537688442211056e-09, |
| "loss": 1.0108, |
| "step": 92500 |
| }, |
| { |
| "epoch": 2.31, |
| "grad_norm": 5.09375, |
| "learning_rate": 7.437185929648241e-09, |
| "loss": 0.9151, |
| "step": 92600 |
| }, |
| { |
| "epoch": 2.32, |
| "grad_norm": 7.71875, |
| "learning_rate": 7.336683417085427e-09, |
| "loss": 0.9009, |
| "step": 92700 |
| }, |
| { |
| "epoch": 2.32, |
| "grad_norm": 11.5625, |
| "learning_rate": 7.2361809045226125e-09, |
| "loss": 0.9579, |
| "step": 92800 |
| }, |
| { |
| "epoch": 2.32, |
| "grad_norm": 6.78125, |
| "learning_rate": 7.135678391959799e-09, |
| "loss": 0.975, |
| "step": 92900 |
| }, |
| { |
| "epoch": 2.33, |
| "grad_norm": 6.84375, |
| "learning_rate": 7.035175879396986e-09, |
| "loss": 1.0061, |
| "step": 93000 |
| }, |
| { |
| "epoch": 2.33, |
| "grad_norm": 4.625, |
| "learning_rate": 6.934673366834172e-09, |
| "loss": 0.9624, |
| "step": 93100 |
| }, |
| { |
| "epoch": 2.33, |
| "grad_norm": 8.1875, |
| "learning_rate": 6.834170854271357e-09, |
| "loss": 0.9731, |
| "step": 93200 |
| }, |
| { |
| "epoch": 2.33, |
| "grad_norm": 9.375, |
| "learning_rate": 6.733668341708544e-09, |
| "loss": 0.952, |
| "step": 93300 |
| }, |
| { |
| "epoch": 2.33, |
| "grad_norm": 8.1875, |
| "learning_rate": 6.633165829145729e-09, |
| "loss": 0.9204, |
| "step": 93400 |
| }, |
| { |
| "epoch": 2.34, |
| "grad_norm": 6.6875, |
| "learning_rate": 6.532663316582915e-09, |
| "loss": 0.9009, |
| "step": 93500 |
| }, |
| { |
| "epoch": 2.34, |
| "grad_norm": 4.96875, |
| "learning_rate": 6.432160804020101e-09, |
| "loss": 1.0315, |
| "step": 93600 |
| }, |
| { |
| "epoch": 2.34, |
| "grad_norm": 5.5, |
| "learning_rate": 6.331658291457287e-09, |
| "loss": 0.972, |
| "step": 93700 |
| }, |
| { |
| "epoch": 2.34, |
| "grad_norm": 3.328125, |
| "learning_rate": 6.231155778894472e-09, |
| "loss": 0.9263, |
| "step": 93800 |
| }, |
| { |
| "epoch": 2.35, |
| "grad_norm": 6.8125, |
| "learning_rate": 6.130653266331658e-09, |
| "loss": 0.9461, |
| "step": 93900 |
| }, |
| { |
| "epoch": 2.35, |
| "grad_norm": 6.15625, |
| "learning_rate": 6.030150753768844e-09, |
| "loss": 0.9521, |
| "step": 94000 |
| }, |
| { |
| "epoch": 2.35, |
| "grad_norm": 5.4375, |
| "learning_rate": 5.92964824120603e-09, |
| "loss": 1.0075, |
| "step": 94100 |
| }, |
| { |
| "epoch": 2.35, |
| "grad_norm": 6.15625, |
| "learning_rate": 5.829145728643216e-09, |
| "loss": 1.0315, |
| "step": 94200 |
| }, |
| { |
| "epoch": 2.36, |
| "grad_norm": 5.09375, |
| "learning_rate": 5.728643216080403e-09, |
| "loss": 0.9098, |
| "step": 94300 |
| }, |
| { |
| "epoch": 2.36, |
| "grad_norm": 6.5, |
| "learning_rate": 5.628140703517588e-09, |
| "loss": 1.005, |
| "step": 94400 |
| }, |
| { |
| "epoch": 2.36, |
| "grad_norm": 4.71875, |
| "learning_rate": 5.527638190954774e-09, |
| "loss": 0.9818, |
| "step": 94500 |
| }, |
| { |
| "epoch": 2.37, |
| "grad_norm": 5.125, |
| "learning_rate": 5.42713567839196e-09, |
| "loss": 0.9769, |
| "step": 94600 |
| }, |
| { |
| "epoch": 2.37, |
| "grad_norm": 4.875, |
| "learning_rate": 5.326633165829146e-09, |
| "loss": 0.9299, |
| "step": 94700 |
| }, |
| { |
| "epoch": 2.37, |
| "grad_norm": 9.0625, |
| "learning_rate": 5.226130653266331e-09, |
| "loss": 0.9481, |
| "step": 94800 |
| }, |
| { |
| "epoch": 2.37, |
| "grad_norm": 6.9375, |
| "learning_rate": 5.125628140703518e-09, |
| "loss": 0.9804, |
| "step": 94900 |
| }, |
| { |
| "epoch": 2.38, |
| "grad_norm": 5.625, |
| "learning_rate": 5.025125628140704e-09, |
| "loss": 0.965, |
| "step": 95000 |
| }, |
| { |
| "epoch": 2.38, |
| "grad_norm": 5.75, |
| "learning_rate": 4.92462311557789e-09, |
| "loss": 0.926, |
| "step": 95100 |
| }, |
| { |
| "epoch": 2.38, |
| "grad_norm": 7.125, |
| "learning_rate": 4.824120603015075e-09, |
| "loss": 0.9498, |
| "step": 95200 |
| }, |
| { |
| "epoch": 2.38, |
| "grad_norm": 4.78125, |
| "learning_rate": 4.723618090452262e-09, |
| "loss": 0.9698, |
| "step": 95300 |
| }, |
| { |
| "epoch": 2.38, |
| "grad_norm": 7.8125, |
| "learning_rate": 4.623115577889447e-09, |
| "loss": 0.988, |
| "step": 95400 |
| }, |
| { |
| "epoch": 2.39, |
| "grad_norm": 6.0625, |
| "learning_rate": 4.522613065326633e-09, |
| "loss": 0.9923, |
| "step": 95500 |
| }, |
| { |
| "epoch": 2.39, |
| "grad_norm": 7.59375, |
| "learning_rate": 4.422110552763819e-09, |
| "loss": 0.9814, |
| "step": 95600 |
| }, |
| { |
| "epoch": 2.39, |
| "grad_norm": 6.53125, |
| "learning_rate": 4.321608040201005e-09, |
| "loss": 0.9604, |
| "step": 95700 |
| }, |
| { |
| "epoch": 2.4, |
| "grad_norm": 6.25, |
| "learning_rate": 4.221105527638191e-09, |
| "loss": 1.0047, |
| "step": 95800 |
| }, |
| { |
| "epoch": 2.4, |
| "grad_norm": 8.0, |
| "learning_rate": 4.120603015075377e-09, |
| "loss": 1.0077, |
| "step": 95900 |
| }, |
| { |
| "epoch": 2.4, |
| "grad_norm": 5.96875, |
| "learning_rate": 4.020100502512563e-09, |
| "loss": 0.899, |
| "step": 96000 |
| }, |
| { |
| "epoch": 2.4, |
| "grad_norm": 4.875, |
| "learning_rate": 3.919597989949749e-09, |
| "loss": 1.0192, |
| "step": 96100 |
| }, |
| { |
| "epoch": 2.41, |
| "grad_norm": 5.0, |
| "learning_rate": 3.819095477386934e-09, |
| "loss": 0.9151, |
| "step": 96200 |
| }, |
| { |
| "epoch": 2.41, |
| "grad_norm": 5.34375, |
| "learning_rate": 3.7185929648241207e-09, |
| "loss": 0.9238, |
| "step": 96300 |
| }, |
| { |
| "epoch": 2.41, |
| "grad_norm": 5.25, |
| "learning_rate": 3.6180904522613062e-09, |
| "loss": 0.9298, |
| "step": 96400 |
| }, |
| { |
| "epoch": 2.41, |
| "grad_norm": 6.65625, |
| "learning_rate": 3.517587939698493e-09, |
| "loss": 0.9592, |
| "step": 96500 |
| }, |
| { |
| "epoch": 2.42, |
| "grad_norm": 4.59375, |
| "learning_rate": 3.4170854271356786e-09, |
| "loss": 0.9678, |
| "step": 96600 |
| }, |
| { |
| "epoch": 2.42, |
| "grad_norm": 6.75, |
| "learning_rate": 3.3165829145728646e-09, |
| "loss": 0.9092, |
| "step": 96700 |
| }, |
| { |
| "epoch": 2.42, |
| "grad_norm": 5.28125, |
| "learning_rate": 3.2160804020100506e-09, |
| "loss": 0.9392, |
| "step": 96800 |
| }, |
| { |
| "epoch": 2.42, |
| "grad_norm": 5.34375, |
| "learning_rate": 3.115577889447236e-09, |
| "loss": 0.9643, |
| "step": 96900 |
| }, |
| { |
| "epoch": 2.42, |
| "grad_norm": 5.625, |
| "learning_rate": 3.015075376884422e-09, |
| "loss": 0.9835, |
| "step": 97000 |
| }, |
| { |
| "epoch": 2.43, |
| "grad_norm": 6.5, |
| "learning_rate": 2.914572864321608e-09, |
| "loss": 0.8944, |
| "step": 97100 |
| }, |
| { |
| "epoch": 2.43, |
| "grad_norm": 5.5625, |
| "learning_rate": 2.814070351758794e-09, |
| "loss": 1.0095, |
| "step": 97200 |
| }, |
| { |
| "epoch": 2.43, |
| "grad_norm": 10.0, |
| "learning_rate": 2.71356783919598e-09, |
| "loss": 0.9684, |
| "step": 97300 |
| }, |
| { |
| "epoch": 2.44, |
| "grad_norm": 6.59375, |
| "learning_rate": 2.6130653266331657e-09, |
| "loss": 0.9435, |
| "step": 97400 |
| }, |
| { |
| "epoch": 2.44, |
| "grad_norm": 5.0, |
| "learning_rate": 2.512562814070352e-09, |
| "loss": 0.9796, |
| "step": 97500 |
| }, |
| { |
| "epoch": 2.44, |
| "grad_norm": 5.84375, |
| "learning_rate": 2.4120603015075376e-09, |
| "loss": 0.9515, |
| "step": 97600 |
| }, |
| { |
| "epoch": 2.44, |
| "grad_norm": 7.09375, |
| "learning_rate": 2.3115577889447236e-09, |
| "loss": 0.9792, |
| "step": 97700 |
| }, |
| { |
| "epoch": 2.44, |
| "grad_norm": 5.28125, |
| "learning_rate": 2.2110552763819096e-09, |
| "loss": 0.931, |
| "step": 97800 |
| }, |
| { |
| "epoch": 2.45, |
| "grad_norm": 6.40625, |
| "learning_rate": 2.1105527638190956e-09, |
| "loss": 0.9806, |
| "step": 97900 |
| }, |
| { |
| "epoch": 2.45, |
| "grad_norm": 3.9375, |
| "learning_rate": 2.0100502512562816e-09, |
| "loss": 0.9843, |
| "step": 98000 |
| }, |
| { |
| "epoch": 2.45, |
| "grad_norm": 5.5, |
| "learning_rate": 1.909547738693467e-09, |
| "loss": 0.9771, |
| "step": 98100 |
| }, |
| { |
| "epoch": 2.46, |
| "grad_norm": 5.8125, |
| "learning_rate": 1.8090452261306531e-09, |
| "loss": 0.9804, |
| "step": 98200 |
| }, |
| { |
| "epoch": 2.46, |
| "grad_norm": 7.09375, |
| "learning_rate": 1.7085427135678393e-09, |
| "loss": 0.939, |
| "step": 98300 |
| }, |
| { |
| "epoch": 2.46, |
| "grad_norm": 7.78125, |
| "learning_rate": 1.6080402010050253e-09, |
| "loss": 0.9246, |
| "step": 98400 |
| }, |
| { |
| "epoch": 2.46, |
| "grad_norm": 10.5, |
| "learning_rate": 1.507537688442211e-09, |
| "loss": 0.9398, |
| "step": 98500 |
| }, |
| { |
| "epoch": 2.46, |
| "grad_norm": 4.5, |
| "learning_rate": 1.407035175879397e-09, |
| "loss": 0.9882, |
| "step": 98600 |
| }, |
| { |
| "epoch": 2.47, |
| "grad_norm": 9.6875, |
| "learning_rate": 1.3065326633165828e-09, |
| "loss": 1.0052, |
| "step": 98700 |
| }, |
| { |
| "epoch": 2.47, |
| "grad_norm": 5.96875, |
| "learning_rate": 1.2060301507537688e-09, |
| "loss": 0.9882, |
| "step": 98800 |
| }, |
| { |
| "epoch": 2.47, |
| "grad_norm": 4.34375, |
| "learning_rate": 1.1055276381909548e-09, |
| "loss": 0.9811, |
| "step": 98900 |
| }, |
| { |
| "epoch": 2.48, |
| "grad_norm": 5.3125, |
| "learning_rate": 1.0050251256281408e-09, |
| "loss": 0.9998, |
| "step": 99000 |
| }, |
| { |
| "epoch": 2.48, |
| "grad_norm": 6.0625, |
| "learning_rate": 9.045226130653266e-10, |
| "loss": 0.9428, |
| "step": 99100 |
| }, |
| { |
| "epoch": 2.48, |
| "grad_norm": 6.3125, |
| "learning_rate": 8.040201005025126e-10, |
| "loss": 1.0201, |
| "step": 99200 |
| }, |
| { |
| "epoch": 2.48, |
| "grad_norm": 6.6875, |
| "learning_rate": 7.035175879396985e-10, |
| "loss": 0.9774, |
| "step": 99300 |
| }, |
| { |
| "epoch": 2.48, |
| "grad_norm": 8.625, |
| "learning_rate": 6.030150753768844e-10, |
| "loss": 1.0282, |
| "step": 99400 |
| }, |
| { |
| "epoch": 2.49, |
| "grad_norm": 6.125, |
| "learning_rate": 5.025125628140704e-10, |
| "loss": 0.9233, |
| "step": 99500 |
| }, |
| { |
| "epoch": 2.49, |
| "grad_norm": 10.25, |
| "learning_rate": 4.020100502512563e-10, |
| "loss": 0.9259, |
| "step": 99600 |
| }, |
| { |
| "epoch": 2.49, |
| "grad_norm": 6.125, |
| "learning_rate": 3.015075376884422e-10, |
| "loss": 0.974, |
| "step": 99700 |
| }, |
| { |
| "epoch": 2.5, |
| "grad_norm": 7.40625, |
| "learning_rate": 2.0100502512562816e-10, |
| "loss": 0.9634, |
| "step": 99800 |
| }, |
| { |
| "epoch": 2.5, |
| "grad_norm": 7.65625, |
| "learning_rate": 1.0050251256281408e-10, |
| "loss": 0.98, |
| "step": 99900 |
| }, |
| { |
| "epoch": 2.5, |
| "grad_norm": 4.84375, |
| "learning_rate": 0.0, |
| "loss": 0.9418, |
| "step": 100000 |
| }, |
| { |
| "epoch": 2.5, |
| "eval_loss": 1.0053296089172363, |
| "eval_runtime": 60.0342, |
| "eval_samples_per_second": 16.657, |
| "eval_steps_per_second": 4.164, |
| "step": 100000 |
| } |
| ], |
| "logging_steps": 100, |
| "max_steps": 100000, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 3, |
| "save_steps": 10000, |
| "total_flos": 1.5733698330624e+18, |
| "train_batch_size": 4, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|