| { |
| "best_metric": 0.8897849462365591, |
| "best_model_checkpoint": "./results/checkpoint-465", |
| "epoch": 5.0, |
| "eval_steps": 500, |
| "global_step": 465, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.10752688172043011, |
| "grad_norm": 77.31194305419922, |
| "learning_rate": 1.5e-06, |
| "loss": 1.0774, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.21505376344086022, |
| "grad_norm": 21.743606567382812, |
| "learning_rate": 3e-06, |
| "loss": 1.205, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.3225806451612903, |
| "grad_norm": 46.53227233886719, |
| "learning_rate": 4.5e-06, |
| "loss": 0.9877, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.43010752688172044, |
| "grad_norm": 60.25495147705078, |
| "learning_rate": 6e-06, |
| "loss": 0.7626, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.5376344086021505, |
| "grad_norm": 42.291969299316406, |
| "learning_rate": 7.5e-06, |
| "loss": 0.6559, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.6451612903225806, |
| "grad_norm": 27.09429931640625, |
| "learning_rate": 9e-06, |
| "loss": 0.3539, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.7526881720430108, |
| "grad_norm": 29.117910385131836, |
| "learning_rate": 1.05e-05, |
| "loss": 0.516, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.8602150537634409, |
| "grad_norm": 35.026817321777344, |
| "learning_rate": 1.2e-05, |
| "loss": 0.5457, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.967741935483871, |
| "grad_norm": 17.851289749145508, |
| "learning_rate": 1.3500000000000001e-05, |
| "loss": 0.5821, |
| "step": 90 |
| }, |
| { |
| "epoch": 1.0, |
| "eval_accuracy": 0.8655913978494624, |
| "eval_f1": 0.8661634583786485, |
| "eval_loss": 0.3687117397785187, |
| "eval_runtime": 1.6159, |
| "eval_samples_per_second": 230.212, |
| "eval_steps_per_second": 3.713, |
| "step": 93 |
| }, |
| { |
| "epoch": 1.075268817204301, |
| "grad_norm": 13.748136520385742, |
| "learning_rate": 1.5e-05, |
| "loss": 0.307, |
| "step": 100 |
| }, |
| { |
| "epoch": 1.1827956989247312, |
| "grad_norm": 3.2135958671569824, |
| "learning_rate": 1.65e-05, |
| "loss": 0.4692, |
| "step": 110 |
| }, |
| { |
| "epoch": 1.2903225806451613, |
| "grad_norm": 11.485154151916504, |
| "learning_rate": 1.8e-05, |
| "loss": 0.3206, |
| "step": 120 |
| }, |
| { |
| "epoch": 1.3978494623655915, |
| "grad_norm": 7.797122478485107, |
| "learning_rate": 1.95e-05, |
| "loss": 0.3138, |
| "step": 130 |
| }, |
| { |
| "epoch": 1.5053763440860215, |
| "grad_norm": 6.313217639923096, |
| "learning_rate": 2.1e-05, |
| "loss": 0.3135, |
| "step": 140 |
| }, |
| { |
| "epoch": 1.6129032258064515, |
| "grad_norm": 18.319936752319336, |
| "learning_rate": 2.25e-05, |
| "loss": 0.2156, |
| "step": 150 |
| }, |
| { |
| "epoch": 1.7204301075268817, |
| "grad_norm": 21.939163208007812, |
| "learning_rate": 2.4e-05, |
| "loss": 0.2749, |
| "step": 160 |
| }, |
| { |
| "epoch": 1.827956989247312, |
| "grad_norm": 20.2148380279541, |
| "learning_rate": 2.55e-05, |
| "loss": 0.4502, |
| "step": 170 |
| }, |
| { |
| "epoch": 1.935483870967742, |
| "grad_norm": 22.044946670532227, |
| "learning_rate": 2.7000000000000002e-05, |
| "loss": 0.364, |
| "step": 180 |
| }, |
| { |
| "epoch": 2.0, |
| "eval_accuracy": 0.8763440860215054, |
| "eval_f1": 0.8742605847498066, |
| "eval_loss": 0.34187108278274536, |
| "eval_runtime": 1.6404, |
| "eval_samples_per_second": 226.774, |
| "eval_steps_per_second": 3.658, |
| "step": 186 |
| }, |
| { |
| "epoch": 2.043010752688172, |
| "grad_norm": 1.5786309242248535, |
| "learning_rate": 2.8499999999999998e-05, |
| "loss": 0.251, |
| "step": 190 |
| }, |
| { |
| "epoch": 2.150537634408602, |
| "grad_norm": 0.3100162446498871, |
| "learning_rate": 3e-05, |
| "loss": 0.0491, |
| "step": 200 |
| }, |
| { |
| "epoch": 2.258064516129032, |
| "grad_norm": 10.301804542541504, |
| "learning_rate": 2.8867924528301887e-05, |
| "loss": 0.0978, |
| "step": 210 |
| }, |
| { |
| "epoch": 2.3655913978494625, |
| "grad_norm": 0.5391497015953064, |
| "learning_rate": 2.7735849056603773e-05, |
| "loss": 0.2325, |
| "step": 220 |
| }, |
| { |
| "epoch": 2.4731182795698925, |
| "grad_norm": 33.04327392578125, |
| "learning_rate": 2.6603773584905663e-05, |
| "loss": 0.3053, |
| "step": 230 |
| }, |
| { |
| "epoch": 2.5806451612903225, |
| "grad_norm": 1.8449115753173828, |
| "learning_rate": 2.547169811320755e-05, |
| "loss": 0.1469, |
| "step": 240 |
| }, |
| { |
| "epoch": 2.688172043010753, |
| "grad_norm": 41.86624526977539, |
| "learning_rate": 2.4339622641509435e-05, |
| "loss": 0.0985, |
| "step": 250 |
| }, |
| { |
| "epoch": 2.795698924731183, |
| "grad_norm": 12.837655067443848, |
| "learning_rate": 2.320754716981132e-05, |
| "loss": 0.1707, |
| "step": 260 |
| }, |
| { |
| "epoch": 2.903225806451613, |
| "grad_norm": 7.691096305847168, |
| "learning_rate": 2.2075471698113208e-05, |
| "loss": 0.1889, |
| "step": 270 |
| }, |
| { |
| "epoch": 3.0, |
| "eval_accuracy": 0.8548387096774194, |
| "eval_f1": 0.8514191596326964, |
| "eval_loss": 0.5420617461204529, |
| "eval_runtime": 1.6403, |
| "eval_samples_per_second": 226.793, |
| "eval_steps_per_second": 3.658, |
| "step": 279 |
| }, |
| { |
| "epoch": 3.010752688172043, |
| "grad_norm": 24.820016860961914, |
| "learning_rate": 2.0943396226415094e-05, |
| "loss": 0.2798, |
| "step": 280 |
| }, |
| { |
| "epoch": 3.118279569892473, |
| "grad_norm": 0.1231377124786377, |
| "learning_rate": 1.981132075471698e-05, |
| "loss": 0.1421, |
| "step": 290 |
| }, |
| { |
| "epoch": 3.225806451612903, |
| "grad_norm": 22.27264404296875, |
| "learning_rate": 1.8679245283018867e-05, |
| "loss": 0.1455, |
| "step": 300 |
| }, |
| { |
| "epoch": 3.3333333333333335, |
| "grad_norm": 43.31319046020508, |
| "learning_rate": 1.7547169811320753e-05, |
| "loss": 0.1223, |
| "step": 310 |
| }, |
| { |
| "epoch": 3.4408602150537635, |
| "grad_norm": 0.16742512583732605, |
| "learning_rate": 1.6415094339622643e-05, |
| "loss": 0.0214, |
| "step": 320 |
| }, |
| { |
| "epoch": 3.5483870967741935, |
| "grad_norm": 44.84831619262695, |
| "learning_rate": 1.528301886792453e-05, |
| "loss": 0.0278, |
| "step": 330 |
| }, |
| { |
| "epoch": 3.6559139784946235, |
| "grad_norm": 16.63416862487793, |
| "learning_rate": 1.4150943396226415e-05, |
| "loss": 0.0238, |
| "step": 340 |
| }, |
| { |
| "epoch": 3.763440860215054, |
| "grad_norm": 0.019003387540578842, |
| "learning_rate": 1.3018867924528303e-05, |
| "loss": 0.0788, |
| "step": 350 |
| }, |
| { |
| "epoch": 3.870967741935484, |
| "grad_norm": 0.062045346945524216, |
| "learning_rate": 1.188679245283019e-05, |
| "loss": 0.0247, |
| "step": 360 |
| }, |
| { |
| "epoch": 3.978494623655914, |
| "grad_norm": 36.90345001220703, |
| "learning_rate": 1.0754716981132076e-05, |
| "loss": 0.1049, |
| "step": 370 |
| }, |
| { |
| "epoch": 4.0, |
| "eval_accuracy": 0.8844086021505376, |
| "eval_f1": 0.8851071788381587, |
| "eval_loss": 0.5284830927848816, |
| "eval_runtime": 1.6712, |
| "eval_samples_per_second": 222.589, |
| "eval_steps_per_second": 3.59, |
| "step": 372 |
| }, |
| { |
| "epoch": 4.086021505376344, |
| "grad_norm": 0.04606785252690315, |
| "learning_rate": 9.622641509433962e-06, |
| "loss": 0.0563, |
| "step": 380 |
| }, |
| { |
| "epoch": 4.193548387096774, |
| "grad_norm": 0.11379247903823853, |
| "learning_rate": 8.49056603773585e-06, |
| "loss": 0.0007, |
| "step": 390 |
| }, |
| { |
| "epoch": 4.301075268817204, |
| "grad_norm": 0.057167768478393555, |
| "learning_rate": 7.358490566037736e-06, |
| "loss": 0.0008, |
| "step": 400 |
| }, |
| { |
| "epoch": 4.408602150537634, |
| "grad_norm": 1.8930367231369019, |
| "learning_rate": 6.226415094339623e-06, |
| "loss": 0.0016, |
| "step": 410 |
| }, |
| { |
| "epoch": 4.516129032258064, |
| "grad_norm": 0.10506568104028702, |
| "learning_rate": 5.094339622641509e-06, |
| "loss": 0.0183, |
| "step": 420 |
| }, |
| { |
| "epoch": 4.623655913978495, |
| "grad_norm": 0.29419633746147156, |
| "learning_rate": 3.962264150943396e-06, |
| "loss": 0.0379, |
| "step": 430 |
| }, |
| { |
| "epoch": 4.731182795698925, |
| "grad_norm": 1.4433554410934448, |
| "learning_rate": 2.830188679245283e-06, |
| "loss": 0.0011, |
| "step": 440 |
| }, |
| { |
| "epoch": 4.838709677419355, |
| "grad_norm": 0.05199088156223297, |
| "learning_rate": 1.69811320754717e-06, |
| "loss": 0.0276, |
| "step": 450 |
| }, |
| { |
| "epoch": 4.946236559139785, |
| "grad_norm": 0.017505839467048645, |
| "learning_rate": 5.660377358490566e-07, |
| "loss": 0.0009, |
| "step": 460 |
| }, |
| { |
| "epoch": 5.0, |
| "eval_accuracy": 0.8897849462365591, |
| "eval_f1": 0.8902518525126154, |
| "eval_loss": 0.518719494342804, |
| "eval_runtime": 1.6655, |
| "eval_samples_per_second": 223.351, |
| "eval_steps_per_second": 3.602, |
| "step": 465 |
| }, |
| { |
| "epoch": 5.0, |
| "step": 465, |
| "total_flos": 312883992936900.0, |
| "train_loss": 0.26607618075144546, |
| "train_runtime": 161.9779, |
| "train_samples_per_second": 45.84, |
| "train_steps_per_second": 2.871 |
| }, |
| { |
| "epoch": 5.0, |
| "eval_accuracy": 0.8897849462365591, |
| "eval_f1": 0.8902518525126154, |
| "eval_loss": 0.518719494342804, |
| "eval_runtime": 1.6415, |
| "eval_samples_per_second": 226.619, |
| "eval_steps_per_second": 3.655, |
| "step": 465 |
| } |
| ], |
| "logging_steps": 10, |
| "max_steps": 465, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 5, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "EarlyStoppingCallback": { |
| "args": { |
| "early_stopping_patience": 3, |
| "early_stopping_threshold": 0.0 |
| }, |
| "attributes": { |
| "early_stopping_patience_counter": 0 |
| } |
| }, |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 312883992936900.0, |
| "train_batch_size": 16, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|