selmamalak's picture
End of training
af14dc8 verified
raw
history blame contribute delete
No virus
17.4 kB
{
"best_metric": 0.9751908396946565,
"best_model_checkpoint": "vit-base-patch16-224-in21k-finetuned-lora-medmnistv2/checkpoint-221",
"epoch": 9.898305084745763,
"eval_steps": 500,
"global_step": 730,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.13559322033898305,
"grad_norm": 0.5722031593322754,
"learning_rate": 0.004931506849315068,
"loss": 0.4736,
"step": 10
},
{
"epoch": 0.2711864406779661,
"grad_norm": 2.089559316635132,
"learning_rate": 0.0048630136986301375,
"loss": 0.4462,
"step": 20
},
{
"epoch": 0.4067796610169492,
"grad_norm": 3.0388331413269043,
"learning_rate": 0.004794520547945206,
"loss": 0.2969,
"step": 30
},
{
"epoch": 0.5423728813559322,
"grad_norm": 0.4622490406036377,
"learning_rate": 0.004726027397260274,
"loss": 0.309,
"step": 40
},
{
"epoch": 0.6779661016949152,
"grad_norm": 1.2164361476898193,
"learning_rate": 0.004657534246575342,
"loss": 0.2817,
"step": 50
},
{
"epoch": 0.8135593220338984,
"grad_norm": 0.30212166905403137,
"learning_rate": 0.004589041095890411,
"loss": 0.24,
"step": 60
},
{
"epoch": 0.9491525423728814,
"grad_norm": 0.9486951231956482,
"learning_rate": 0.00452054794520548,
"loss": 0.2447,
"step": 70
},
{
"epoch": 0.9898305084745763,
"eval_accuracy": 0.9561068702290076,
"eval_f1": 0.9445770389275943,
"eval_loss": 0.11798465996980667,
"eval_precision": 0.9313190730837789,
"eval_recall": 0.9607635913548509,
"eval_runtime": 2.3492,
"eval_samples_per_second": 223.055,
"eval_steps_per_second": 14.047,
"step": 73
},
{
"epoch": 1.0847457627118644,
"grad_norm": 1.1987316608428955,
"learning_rate": 0.004452054794520548,
"loss": 0.2017,
"step": 80
},
{
"epoch": 1.2203389830508475,
"grad_norm": 0.6783656477928162,
"learning_rate": 0.004383561643835616,
"loss": 0.2214,
"step": 90
},
{
"epoch": 1.3559322033898304,
"grad_norm": 1.073630452156067,
"learning_rate": 0.004315068493150685,
"loss": 0.1989,
"step": 100
},
{
"epoch": 1.4915254237288136,
"grad_norm": 0.3175281584262848,
"learning_rate": 0.0042465753424657535,
"loss": 0.1857,
"step": 110
},
{
"epoch": 1.6271186440677967,
"grad_norm": 0.6710169911384583,
"learning_rate": 0.004178082191780822,
"loss": 0.2305,
"step": 120
},
{
"epoch": 1.7627118644067796,
"grad_norm": 0.8543218374252319,
"learning_rate": 0.00410958904109589,
"loss": 0.2111,
"step": 130
},
{
"epoch": 1.8983050847457628,
"grad_norm": 0.6326661109924316,
"learning_rate": 0.004041095890410959,
"loss": 0.2136,
"step": 140
},
{
"epoch": 1.993220338983051,
"eval_accuracy": 0.9637404580152672,
"eval_f1": 0.9529434006853362,
"eval_loss": 0.10151813924312592,
"eval_precision": 0.9497822332357138,
"eval_recall": 0.9562315528896506,
"eval_runtime": 2.4266,
"eval_samples_per_second": 215.944,
"eval_steps_per_second": 13.6,
"step": 147
},
{
"epoch": 2.0338983050847457,
"grad_norm": 0.40849238634109497,
"learning_rate": 0.003972602739726027,
"loss": 0.2148,
"step": 150
},
{
"epoch": 2.169491525423729,
"grad_norm": 1.0935662984848022,
"learning_rate": 0.003904109589041096,
"loss": 0.2106,
"step": 160
},
{
"epoch": 2.305084745762712,
"grad_norm": 0.6485463380813599,
"learning_rate": 0.0038356164383561643,
"loss": 0.2205,
"step": 170
},
{
"epoch": 2.440677966101695,
"grad_norm": 0.5151948928833008,
"learning_rate": 0.003767123287671233,
"loss": 0.2018,
"step": 180
},
{
"epoch": 2.576271186440678,
"grad_norm": 0.5166621804237366,
"learning_rate": 0.0036986301369863013,
"loss": 0.1827,
"step": 190
},
{
"epoch": 2.711864406779661,
"grad_norm": 0.8355435729026794,
"learning_rate": 0.00363013698630137,
"loss": 0.2171,
"step": 200
},
{
"epoch": 2.847457627118644,
"grad_norm": 0.558962881565094,
"learning_rate": 0.003561643835616438,
"loss": 0.1845,
"step": 210
},
{
"epoch": 2.983050847457627,
"grad_norm": 0.6180588603019714,
"learning_rate": 0.003493150684931507,
"loss": 0.1431,
"step": 220
},
{
"epoch": 2.9966101694915253,
"eval_accuracy": 0.9751908396946565,
"eval_f1": 0.9671702932600785,
"eval_loss": 0.07291658967733383,
"eval_precision": 0.9731940648184303,
"eval_recall": 0.9615252784918595,
"eval_runtime": 2.3109,
"eval_samples_per_second": 226.748,
"eval_steps_per_second": 14.28,
"step": 221
},
{
"epoch": 3.1186440677966103,
"grad_norm": 0.4949386715888977,
"learning_rate": 0.003424657534246575,
"loss": 0.2148,
"step": 230
},
{
"epoch": 3.2542372881355934,
"grad_norm": 0.5671024918556213,
"learning_rate": 0.003356164383561644,
"loss": 0.172,
"step": 240
},
{
"epoch": 3.389830508474576,
"grad_norm": 0.515737771987915,
"learning_rate": 0.003287671232876712,
"loss": 0.1962,
"step": 250
},
{
"epoch": 3.5254237288135593,
"grad_norm": 0.42013198137283325,
"learning_rate": 0.0032191780821917808,
"loss": 0.1571,
"step": 260
},
{
"epoch": 3.6610169491525424,
"grad_norm": 0.41014379262924194,
"learning_rate": 0.003150684931506849,
"loss": 0.1905,
"step": 270
},
{
"epoch": 3.7966101694915255,
"grad_norm": 0.48967182636260986,
"learning_rate": 0.003082191780821918,
"loss": 0.1344,
"step": 280
},
{
"epoch": 3.9322033898305087,
"grad_norm": 0.3962344229221344,
"learning_rate": 0.0030136986301369864,
"loss": 0.1576,
"step": 290
},
{
"epoch": 4.0,
"eval_accuracy": 0.9637404580152672,
"eval_f1": 0.9531647653769258,
"eval_loss": 0.08732008934020996,
"eval_precision": 0.9480282738095238,
"eval_recall": 0.9586499095496526,
"eval_runtime": 2.3116,
"eval_samples_per_second": 226.687,
"eval_steps_per_second": 14.276,
"step": 295
},
{
"epoch": 4.067796610169491,
"grad_norm": 0.28995221853256226,
"learning_rate": 0.002945205479452055,
"loss": 0.1319,
"step": 300
},
{
"epoch": 4.203389830508475,
"grad_norm": 0.5315499901771545,
"learning_rate": 0.0028767123287671234,
"loss": 0.2781,
"step": 310
},
{
"epoch": 4.338983050847458,
"grad_norm": 0.24718379974365234,
"learning_rate": 0.002808219178082192,
"loss": 0.1896,
"step": 320
},
{
"epoch": 4.47457627118644,
"grad_norm": 0.404805988073349,
"learning_rate": 0.0027397260273972603,
"loss": 0.1829,
"step": 330
},
{
"epoch": 4.610169491525424,
"grad_norm": 0.5308801531791687,
"learning_rate": 0.002671232876712329,
"loss": 0.1665,
"step": 340
},
{
"epoch": 4.745762711864407,
"grad_norm": 0.7429054975509644,
"learning_rate": 0.0026027397260273972,
"loss": 0.1646,
"step": 350
},
{
"epoch": 4.88135593220339,
"grad_norm": 0.6361825466156006,
"learning_rate": 0.002534246575342466,
"loss": 0.2072,
"step": 360
},
{
"epoch": 4.989830508474577,
"eval_accuracy": 0.9713740458015268,
"eval_f1": 0.9626722135947228,
"eval_loss": 0.076077401638031,
"eval_precision": 0.9615676167374165,
"eval_recall": 0.9637912977244597,
"eval_runtime": 2.3751,
"eval_samples_per_second": 220.624,
"eval_steps_per_second": 13.894,
"step": 368
},
{
"epoch": 5.016949152542373,
"grad_norm": 0.6114002466201782,
"learning_rate": 0.002465753424657534,
"loss": 0.1943,
"step": 370
},
{
"epoch": 5.1525423728813555,
"grad_norm": 0.32443058490753174,
"learning_rate": 0.002397260273972603,
"loss": 0.1541,
"step": 380
},
{
"epoch": 5.288135593220339,
"grad_norm": 0.44808831810951233,
"learning_rate": 0.002328767123287671,
"loss": 0.1397,
"step": 390
},
{
"epoch": 5.423728813559322,
"grad_norm": 0.478524386882782,
"learning_rate": 0.00226027397260274,
"loss": 0.1647,
"step": 400
},
{
"epoch": 5.559322033898305,
"grad_norm": 0.8286917805671692,
"learning_rate": 0.002191780821917808,
"loss": 0.1532,
"step": 410
},
{
"epoch": 5.694915254237288,
"grad_norm": 0.7862728238105774,
"learning_rate": 0.0021232876712328768,
"loss": 0.1474,
"step": 420
},
{
"epoch": 5.830508474576272,
"grad_norm": 1.1601946353912354,
"learning_rate": 0.002054794520547945,
"loss": 0.2347,
"step": 430
},
{
"epoch": 5.966101694915254,
"grad_norm": 1.3851842880249023,
"learning_rate": 0.0019863013698630137,
"loss": 0.1908,
"step": 440
},
{
"epoch": 5.9932203389830505,
"eval_accuracy": 0.9599236641221374,
"eval_f1": 0.9496169994551458,
"eval_loss": 0.10439441353082657,
"eval_precision": 0.9348118279569892,
"eval_recall": 0.9681709987622584,
"eval_runtime": 2.3199,
"eval_samples_per_second": 225.87,
"eval_steps_per_second": 14.225,
"step": 442
},
{
"epoch": 6.101694915254237,
"grad_norm": 0.6037909388542175,
"learning_rate": 0.0019178082191780822,
"loss": 0.1326,
"step": 450
},
{
"epoch": 6.237288135593221,
"grad_norm": 0.6797343492507935,
"learning_rate": 0.0018493150684931506,
"loss": 0.1909,
"step": 460
},
{
"epoch": 6.372881355932203,
"grad_norm": 0.5503271222114563,
"learning_rate": 0.0017876712328767123,
"loss": 0.1328,
"step": 470
},
{
"epoch": 6.508474576271187,
"grad_norm": 0.46364396810531616,
"learning_rate": 0.0017191780821917808,
"loss": 0.1586,
"step": 480
},
{
"epoch": 6.6440677966101696,
"grad_norm": 0.7383654713630676,
"learning_rate": 0.0016506849315068492,
"loss": 0.1393,
"step": 490
},
{
"epoch": 6.779661016949152,
"grad_norm": 0.7454331517219543,
"learning_rate": 0.0015821917808219177,
"loss": 0.163,
"step": 500
},
{
"epoch": 6.915254237288136,
"grad_norm": 0.7943591475486755,
"learning_rate": 0.0015136986301369862,
"loss": 0.1637,
"step": 510
},
{
"epoch": 6.996610169491525,
"eval_accuracy": 0.9675572519083969,
"eval_f1": 0.9582894361020196,
"eval_loss": 0.07422558218240738,
"eval_precision": 0.9512019762554385,
"eval_recall": 0.9660573169570599,
"eval_runtime": 2.67,
"eval_samples_per_second": 196.256,
"eval_steps_per_second": 12.36,
"step": 516
},
{
"epoch": 7.0508474576271185,
"grad_norm": 0.5350046753883362,
"learning_rate": 0.0014452054794520546,
"loss": 0.1415,
"step": 520
},
{
"epoch": 7.186440677966102,
"grad_norm": 0.9918954372406006,
"learning_rate": 0.001376712328767123,
"loss": 0.1548,
"step": 530
},
{
"epoch": 7.322033898305085,
"grad_norm": 2.7371268272399902,
"learning_rate": 0.0013082191780821918,
"loss": 0.1583,
"step": 540
},
{
"epoch": 7.4576271186440675,
"grad_norm": 1.9455125331878662,
"learning_rate": 0.0012397260273972603,
"loss": 0.1399,
"step": 550
},
{
"epoch": 7.593220338983051,
"grad_norm": 1.1810232400894165,
"learning_rate": 0.0011712328767123287,
"loss": 0.1535,
"step": 560
},
{
"epoch": 7.728813559322034,
"grad_norm": 0.5082597136497498,
"learning_rate": 0.0011027397260273974,
"loss": 0.1395,
"step": 570
},
{
"epoch": 7.864406779661017,
"grad_norm": 0.435004860162735,
"learning_rate": 0.001034246575342466,
"loss": 0.1443,
"step": 580
},
{
"epoch": 8.0,
"grad_norm": 0.4963846504688263,
"learning_rate": 0.0009657534246575344,
"loss": 0.1385,
"step": 590
},
{
"epoch": 8.0,
"eval_accuracy": 0.9312977099236641,
"eval_f1": 0.9169177104804185,
"eval_loss": 0.18427740037441254,
"eval_precision": 0.8947368421052632,
"eval_recall": 0.9537275064267352,
"eval_runtime": 2.338,
"eval_samples_per_second": 224.121,
"eval_steps_per_second": 14.115,
"step": 590
},
{
"epoch": 8.135593220338983,
"grad_norm": 0.5249541997909546,
"learning_rate": 0.0008972602739726028,
"loss": 0.1574,
"step": 600
},
{
"epoch": 8.271186440677965,
"grad_norm": 0.4605633020401001,
"learning_rate": 0.0008287671232876713,
"loss": 0.1543,
"step": 610
},
{
"epoch": 8.40677966101695,
"grad_norm": 0.3483382761478424,
"learning_rate": 0.0007602739726027398,
"loss": 0.131,
"step": 620
},
{
"epoch": 8.542372881355933,
"grad_norm": 0.21878492832183838,
"learning_rate": 0.0006917808219178081,
"loss": 0.1158,
"step": 630
},
{
"epoch": 8.677966101694915,
"grad_norm": 0.6566260457038879,
"learning_rate": 0.0006232876712328767,
"loss": 0.1018,
"step": 640
},
{
"epoch": 8.813559322033898,
"grad_norm": 0.3624425232410431,
"learning_rate": 0.0005547945205479452,
"loss": 0.1137,
"step": 650
},
{
"epoch": 8.94915254237288,
"grad_norm": 0.8725977540016174,
"learning_rate": 0.0004863013698630137,
"loss": 0.1335,
"step": 660
},
{
"epoch": 8.989830508474576,
"eval_accuracy": 0.9751908396946565,
"eval_f1": 0.9679548394684229,
"eval_loss": 0.0676569938659668,
"eval_precision": 0.9626488095238095,
"eval_recall": 0.9736170617918689,
"eval_runtime": 2.3252,
"eval_samples_per_second": 225.358,
"eval_steps_per_second": 14.192,
"step": 663
},
{
"epoch": 9.084745762711865,
"grad_norm": 0.46945634484291077,
"learning_rate": 0.00041780821917808224,
"loss": 0.1348,
"step": 670
},
{
"epoch": 9.220338983050848,
"grad_norm": 0.33705276250839233,
"learning_rate": 0.0003493150684931507,
"loss": 0.0968,
"step": 680
},
{
"epoch": 9.35593220338983,
"grad_norm": 0.554704487323761,
"learning_rate": 0.0002808219178082192,
"loss": 0.1379,
"step": 690
},
{
"epoch": 9.491525423728813,
"grad_norm": 0.5419002175331116,
"learning_rate": 0.00021232876712328768,
"loss": 0.128,
"step": 700
},
{
"epoch": 9.627118644067796,
"grad_norm": 0.32422158122062683,
"learning_rate": 0.00014383561643835618,
"loss": 0.1065,
"step": 710
},
{
"epoch": 9.76271186440678,
"grad_norm": 0.3378284275531769,
"learning_rate": 7.534246575342466e-05,
"loss": 0.1354,
"step": 720
},
{
"epoch": 9.898305084745763,
"grad_norm": 0.2171776294708252,
"learning_rate": 6.849315068493151e-06,
"loss": 0.1186,
"step": 730
},
{
"epoch": 9.898305084745763,
"eval_accuracy": 0.9751908396946565,
"eval_f1": 0.9679548394684229,
"eval_loss": 0.0765310674905777,
"eval_precision": 0.9626488095238095,
"eval_recall": 0.9736170617918689,
"eval_runtime": 2.3291,
"eval_samples_per_second": 224.984,
"eval_steps_per_second": 14.169,
"step": 730
},
{
"epoch": 9.898305084745763,
"step": 730,
"total_flos": 3.6369520534486057e+18,
"train_loss": 0.18177251358554788,
"train_runtime": 459.6522,
"train_samples_per_second": 102.425,
"train_steps_per_second": 1.588
}
],
"logging_steps": 10,
"max_steps": 730,
"num_input_tokens_seen": 0,
"num_train_epochs": 10,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 3.6369520534486057e+18,
"train_batch_size": 16,
"trial_name": null,
"trial_params": null
}