Qwen2-7B_Function_Call_tiny_lora / trainer_state.json
svjack's picture
Upload folder using huggingface_hub
d16fbc5 verified
raw
history blame
No virus
16 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.0,
"eval_steps": 500,
"global_step": 375,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.04,
"grad_norm": 0.4841386675834656,
"learning_rate": 4.997807075247146e-05,
"loss": 0.688,
"num_input_tokens_seen": 46776,
"step": 5
},
{
"epoch": 0.08,
"grad_norm": 0.3365215063095093,
"learning_rate": 4.991232148123761e-05,
"loss": 0.763,
"num_input_tokens_seen": 85136,
"step": 10
},
{
"epoch": 0.12,
"grad_norm": 0.5198595523834229,
"learning_rate": 4.980286753286195e-05,
"loss": 0.6882,
"num_input_tokens_seen": 126584,
"step": 15
},
{
"epoch": 0.16,
"grad_norm": 0.40807807445526123,
"learning_rate": 4.964990092676263e-05,
"loss": 0.6951,
"num_input_tokens_seen": 167968,
"step": 20
},
{
"epoch": 0.2,
"grad_norm": 0.439756840467453,
"learning_rate": 4.9453690018345144e-05,
"loss": 0.5008,
"num_input_tokens_seen": 206832,
"step": 25
},
{
"epoch": 0.24,
"grad_norm": 0.45932796597480774,
"learning_rate": 4.9214579028215776e-05,
"loss": 0.542,
"num_input_tokens_seen": 243656,
"step": 30
},
{
"epoch": 0.28,
"grad_norm": 0.3281826376914978,
"learning_rate": 4.893298743830168e-05,
"loss": 0.5369,
"num_input_tokens_seen": 284016,
"step": 35
},
{
"epoch": 0.32,
"grad_norm": 0.2551577389240265,
"learning_rate": 4.860940925593703e-05,
"loss": 0.4948,
"num_input_tokens_seen": 327408,
"step": 40
},
{
"epoch": 0.36,
"grad_norm": 0.5936368107795715,
"learning_rate": 4.8244412147206284e-05,
"loss": 0.5244,
"num_input_tokens_seen": 363376,
"step": 45
},
{
"epoch": 0.4,
"grad_norm": 0.2480110377073288,
"learning_rate": 4.783863644106502e-05,
"loss": 0.421,
"num_input_tokens_seen": 398840,
"step": 50
},
{
"epoch": 0.44,
"grad_norm": 0.4601157307624817,
"learning_rate": 4.7392794005985326e-05,
"loss": 0.4517,
"num_input_tokens_seen": 436136,
"step": 55
},
{
"epoch": 0.48,
"grad_norm": 0.6387960314750671,
"learning_rate": 4.690766700109659e-05,
"loss": 0.4661,
"num_input_tokens_seen": 475856,
"step": 60
},
{
"epoch": 0.52,
"grad_norm": 0.4365377128124237,
"learning_rate": 4.638410650401267e-05,
"loss": 0.4928,
"num_input_tokens_seen": 514496,
"step": 65
},
{
"epoch": 0.56,
"grad_norm": 0.2992519438266754,
"learning_rate": 4.5823031017752485e-05,
"loss": 0.5424,
"num_input_tokens_seen": 554424,
"step": 70
},
{
"epoch": 0.6,
"grad_norm": 0.828787088394165,
"learning_rate": 4.522542485937369e-05,
"loss": 0.5419,
"num_input_tokens_seen": 593264,
"step": 75
},
{
"epoch": 0.64,
"grad_norm": 0.2943418622016907,
"learning_rate": 4.4592336433146e-05,
"loss": 0.4558,
"num_input_tokens_seen": 630264,
"step": 80
},
{
"epoch": 0.68,
"grad_norm": 0.3707512617111206,
"learning_rate": 4.3924876391293915e-05,
"loss": 0.5656,
"num_input_tokens_seen": 668864,
"step": 85
},
{
"epoch": 0.72,
"grad_norm": 0.4809654653072357,
"learning_rate": 4.3224215685535294e-05,
"loss": 0.4832,
"num_input_tokens_seen": 712504,
"step": 90
},
{
"epoch": 0.76,
"grad_norm": 0.516292929649353,
"learning_rate": 4.249158351283414e-05,
"loss": 0.4626,
"num_input_tokens_seen": 748872,
"step": 95
},
{
"epoch": 0.8,
"grad_norm": 0.5604212284088135,
"learning_rate": 4.172826515897146e-05,
"loss": 0.4837,
"num_input_tokens_seen": 788408,
"step": 100
},
{
"epoch": 0.84,
"grad_norm": 0.7355071902275085,
"learning_rate": 4.093559974371725e-05,
"loss": 0.5144,
"num_input_tokens_seen": 828448,
"step": 105
},
{
"epoch": 0.88,
"grad_norm": 0.46999862790107727,
"learning_rate": 4.011497787155938e-05,
"loss": 0.493,
"num_input_tokens_seen": 864680,
"step": 110
},
{
"epoch": 0.92,
"grad_norm": 0.5488362312316895,
"learning_rate": 3.92678391921108e-05,
"loss": 0.4083,
"num_input_tokens_seen": 902568,
"step": 115
},
{
"epoch": 0.96,
"grad_norm": 0.5053719282150269,
"learning_rate": 3.8395669874474915e-05,
"loss": 0.5172,
"num_input_tokens_seen": 944752,
"step": 120
},
{
"epoch": 1.0,
"grad_norm": 0.44234323501586914,
"learning_rate": 3.7500000000000003e-05,
"loss": 0.5843,
"num_input_tokens_seen": 988656,
"step": 125
},
{
"epoch": 1.04,
"grad_norm": 0.4518623352050781,
"learning_rate": 3.6582400877996546e-05,
"loss": 0.4567,
"num_input_tokens_seen": 1028584,
"step": 130
},
{
"epoch": 1.08,
"grad_norm": 0.39792686700820923,
"learning_rate": 3.564448228912682e-05,
"loss": 0.418,
"num_input_tokens_seen": 1065104,
"step": 135
},
{
"epoch": 1.12,
"grad_norm": 0.5141780376434326,
"learning_rate": 3.4687889661302576e-05,
"loss": 0.3785,
"num_input_tokens_seen": 1106744,
"step": 140
},
{
"epoch": 1.16,
"grad_norm": 0.5755515098571777,
"learning_rate": 3.3714301183045385e-05,
"loss": 0.4097,
"num_input_tokens_seen": 1140672,
"step": 145
},
{
"epoch": 1.2,
"grad_norm": 0.4626426696777344,
"learning_rate": 3.272542485937369e-05,
"loss": 0.4507,
"num_input_tokens_seen": 1177280,
"step": 150
},
{
"epoch": 1.24,
"grad_norm": 0.49699896574020386,
"learning_rate": 3.172299551538164e-05,
"loss": 0.368,
"num_input_tokens_seen": 1215744,
"step": 155
},
{
"epoch": 1.28,
"grad_norm": 0.5842998027801514,
"learning_rate": 3.0708771752766394e-05,
"loss": 0.4301,
"num_input_tokens_seen": 1257040,
"step": 160
},
{
"epoch": 1.32,
"grad_norm": 0.46173322200775146,
"learning_rate": 2.9684532864643122e-05,
"loss": 0.4488,
"num_input_tokens_seen": 1295992,
"step": 165
},
{
"epoch": 1.3599999999999999,
"grad_norm": 0.4926168918609619,
"learning_rate": 2.8652075714060295e-05,
"loss": 0.4075,
"num_input_tokens_seen": 1334672,
"step": 170
},
{
"epoch": 1.4,
"grad_norm": 0.5110897421836853,
"learning_rate": 2.761321158169134e-05,
"loss": 0.4991,
"num_input_tokens_seen": 1379280,
"step": 175
},
{
"epoch": 1.44,
"grad_norm": 0.33892518281936646,
"learning_rate": 2.656976298823284e-05,
"loss": 0.4894,
"num_input_tokens_seen": 1419704,
"step": 180
},
{
"epoch": 1.48,
"grad_norm": 0.46542254090309143,
"learning_rate": 2.5523560497083926e-05,
"loss": 0.4967,
"num_input_tokens_seen": 1459256,
"step": 185
},
{
"epoch": 1.52,
"grad_norm": 0.9100029468536377,
"learning_rate": 2.447643950291608e-05,
"loss": 0.5297,
"num_input_tokens_seen": 1503576,
"step": 190
},
{
"epoch": 1.56,
"grad_norm": 0.7347203493118286,
"learning_rate": 2.3430237011767167e-05,
"loss": 0.3939,
"num_input_tokens_seen": 1539904,
"step": 195
},
{
"epoch": 1.6,
"grad_norm": 0.6078445315361023,
"learning_rate": 2.238678841830867e-05,
"loss": 0.461,
"num_input_tokens_seen": 1580560,
"step": 200
},
{
"epoch": 1.6400000000000001,
"grad_norm": 0.5878064632415771,
"learning_rate": 2.1347924285939714e-05,
"loss": 0.4622,
"num_input_tokens_seen": 1623608,
"step": 205
},
{
"epoch": 1.6800000000000002,
"grad_norm": 0.6516324877738953,
"learning_rate": 2.031546713535688e-05,
"loss": 0.4043,
"num_input_tokens_seen": 1660744,
"step": 210
},
{
"epoch": 1.72,
"grad_norm": 0.9136825203895569,
"learning_rate": 1.9291228247233605e-05,
"loss": 0.428,
"num_input_tokens_seen": 1698784,
"step": 215
},
{
"epoch": 1.76,
"grad_norm": 0.751567006111145,
"learning_rate": 1.827700448461836e-05,
"loss": 0.3779,
"num_input_tokens_seen": 1734792,
"step": 220
},
{
"epoch": 1.8,
"grad_norm": 0.8802843689918518,
"learning_rate": 1.7274575140626318e-05,
"loss": 0.4526,
"num_input_tokens_seen": 1776144,
"step": 225
},
{
"epoch": 1.8399999999999999,
"grad_norm": 0.6807656288146973,
"learning_rate": 1.6285698816954624e-05,
"loss": 0.4627,
"num_input_tokens_seen": 1815864,
"step": 230
},
{
"epoch": 1.88,
"grad_norm": 0.6222789883613586,
"learning_rate": 1.5312110338697426e-05,
"loss": 0.4873,
"num_input_tokens_seen": 1856168,
"step": 235
},
{
"epoch": 1.92,
"grad_norm": 0.5224676728248596,
"learning_rate": 1.4355517710873184e-05,
"loss": 0.3234,
"num_input_tokens_seen": 1899120,
"step": 240
},
{
"epoch": 1.96,
"grad_norm": 0.4191940724849701,
"learning_rate": 1.3417599122003464e-05,
"loss": 0.4438,
"num_input_tokens_seen": 1936808,
"step": 245
},
{
"epoch": 2.0,
"grad_norm": 0.6064372062683105,
"learning_rate": 1.2500000000000006e-05,
"loss": 0.4407,
"num_input_tokens_seen": 1977312,
"step": 250
},
{
"epoch": 2.04,
"grad_norm": 0.5188178420066833,
"learning_rate": 1.1604330125525079e-05,
"loss": 0.4401,
"num_input_tokens_seen": 2018336,
"step": 255
},
{
"epoch": 2.08,
"grad_norm": 0.5036932826042175,
"learning_rate": 1.0732160807889211e-05,
"loss": 0.3771,
"num_input_tokens_seen": 2057776,
"step": 260
},
{
"epoch": 2.12,
"grad_norm": 0.5732894539833069,
"learning_rate": 9.88502212844063e-06,
"loss": 0.4043,
"num_input_tokens_seen": 2102312,
"step": 265
},
{
"epoch": 2.16,
"grad_norm": 0.6521058082580566,
"learning_rate": 9.064400256282757e-06,
"loss": 0.4018,
"num_input_tokens_seen": 2137304,
"step": 270
},
{
"epoch": 2.2,
"grad_norm": 0.698835015296936,
"learning_rate": 8.271734841028553e-06,
"loss": 0.4258,
"num_input_tokens_seen": 2176144,
"step": 275
},
{
"epoch": 2.24,
"grad_norm": 0.7500255107879639,
"learning_rate": 7.508416487165862e-06,
"loss": 0.3912,
"num_input_tokens_seen": 2211640,
"step": 280
},
{
"epoch": 2.2800000000000002,
"grad_norm": 0.7223864793777466,
"learning_rate": 6.775784314464717e-06,
"loss": 0.3458,
"num_input_tokens_seen": 2247272,
"step": 285
},
{
"epoch": 2.32,
"grad_norm": 0.6346307992935181,
"learning_rate": 6.075123608706093e-06,
"loss": 0.4255,
"num_input_tokens_seen": 2285776,
"step": 290
},
{
"epoch": 2.36,
"grad_norm": 0.6048356890678406,
"learning_rate": 5.4076635668540075e-06,
"loss": 0.4222,
"num_input_tokens_seen": 2326112,
"step": 295
},
{
"epoch": 2.4,
"grad_norm": 1.0743402242660522,
"learning_rate": 4.7745751406263165e-06,
"loss": 0.399,
"num_input_tokens_seen": 2362816,
"step": 300
},
{
"epoch": 2.44,
"grad_norm": 0.6533491015434265,
"learning_rate": 4.176968982247514e-06,
"loss": 0.3382,
"num_input_tokens_seen": 2402328,
"step": 305
},
{
"epoch": 2.48,
"grad_norm": 0.7778833508491516,
"learning_rate": 3.6158934959873353e-06,
"loss": 0.4465,
"num_input_tokens_seen": 2445600,
"step": 310
},
{
"epoch": 2.52,
"grad_norm": 0.7358531355857849,
"learning_rate": 3.092332998903416e-06,
"loss": 0.325,
"num_input_tokens_seen": 2485840,
"step": 315
},
{
"epoch": 2.56,
"grad_norm": 0.7607108354568481,
"learning_rate": 2.6072059940146775e-06,
"loss": 0.392,
"num_input_tokens_seen": 2524760,
"step": 320
},
{
"epoch": 2.6,
"grad_norm": 0.7286326289176941,
"learning_rate": 2.1613635589349756e-06,
"loss": 0.3672,
"num_input_tokens_seen": 2562040,
"step": 325
},
{
"epoch": 2.64,
"grad_norm": 0.5720709562301636,
"learning_rate": 1.7555878527937164e-06,
"loss": 0.3554,
"num_input_tokens_seen": 2600056,
"step": 330
},
{
"epoch": 2.68,
"grad_norm": 0.9434547424316406,
"learning_rate": 1.3905907440629752e-06,
"loss": 0.3801,
"num_input_tokens_seen": 2633392,
"step": 335
},
{
"epoch": 2.7199999999999998,
"grad_norm": 0.9704461693763733,
"learning_rate": 1.067012561698319e-06,
"loss": 0.435,
"num_input_tokens_seen": 2677016,
"step": 340
},
{
"epoch": 2.76,
"grad_norm": 0.6400224566459656,
"learning_rate": 7.854209717842231e-07,
"loss": 0.4063,
"num_input_tokens_seen": 2719536,
"step": 345
},
{
"epoch": 2.8,
"grad_norm": 0.46599817276000977,
"learning_rate": 5.463099816548579e-07,
"loss": 0.4894,
"num_input_tokens_seen": 2767424,
"step": 350
},
{
"epoch": 2.84,
"grad_norm": 0.4850185215473175,
"learning_rate": 3.5009907323737825e-07,
"loss": 0.3822,
"num_input_tokens_seen": 2805832,
"step": 355
},
{
"epoch": 2.88,
"grad_norm": 0.8774814605712891,
"learning_rate": 1.9713246713805588e-07,
"loss": 0.4028,
"num_input_tokens_seen": 2845912,
"step": 360
},
{
"epoch": 2.92,
"grad_norm": 0.6378630995750427,
"learning_rate": 8.767851876239074e-08,
"loss": 0.4293,
"num_input_tokens_seen": 2888128,
"step": 365
},
{
"epoch": 2.96,
"grad_norm": 0.8157021403312683,
"learning_rate": 2.192924752854042e-08,
"loss": 0.428,
"num_input_tokens_seen": 2925024,
"step": 370
},
{
"epoch": 3.0,
"grad_norm": 0.5457362532615662,
"learning_rate": 0.0,
"loss": 0.4766,
"num_input_tokens_seen": 2965968,
"step": 375
},
{
"epoch": 3.0,
"num_input_tokens_seen": 2965968,
"step": 375,
"total_flos": 1.2618658999382835e+17,
"train_loss": 0.45766232585906985,
"train_runtime": 3803.7297,
"train_samples_per_second": 1.577,
"train_steps_per_second": 0.099
}
],
"logging_steps": 5,
"max_steps": 375,
"num_input_tokens_seen": 2965968,
"num_train_epochs": 3,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 1.2618658999382835e+17,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}