|
{ |
|
"best_metric": 0.9108088053425674, |
|
"best_model_checkpoint": "qqp_#1_glue_cased/checkpoint-34113", |
|
"epoch": 3.0, |
|
"global_step": 34113, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.04397150646381145, |
|
"learning_rate": 1.9824113974144757e-05, |
|
"loss": 0.43561614990234376, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.0879430129276229, |
|
"learning_rate": 1.9648227948289512e-05, |
|
"loss": 0.3685419006347656, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.13191451939143434, |
|
"learning_rate": 1.9472341922434264e-05, |
|
"loss": 0.3377880249023438, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.1758860258552458, |
|
"learning_rate": 1.929645589657902e-05, |
|
"loss": 0.327943115234375, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.21985753231905725, |
|
"learning_rate": 1.9120569870723774e-05, |
|
"loss": 0.3118376159667969, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.2638290387828687, |
|
"learning_rate": 1.894468384486853e-05, |
|
"loss": 0.3047716064453125, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.30780054524668016, |
|
"learning_rate": 1.876879781901328e-05, |
|
"loss": 0.3038143005371094, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.3517720517104916, |
|
"learning_rate": 1.8592911793158036e-05, |
|
"loss": 0.2949424438476563, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.39574355817430307, |
|
"learning_rate": 1.8417025767302788e-05, |
|
"loss": 0.2912021789550781, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.4397150646381145, |
|
"learning_rate": 1.8241139741447543e-05, |
|
"loss": 0.2867056884765625, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.483686571101926, |
|
"learning_rate": 1.80652537155923e-05, |
|
"loss": 0.27814862060546874, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.5276580775657373, |
|
"learning_rate": 1.788936768973705e-05, |
|
"loss": 0.2777033996582031, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.5716295840295489, |
|
"learning_rate": 1.7713481663881805e-05, |
|
"loss": 0.2808096618652344, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.6156010904933603, |
|
"learning_rate": 1.753759563802656e-05, |
|
"loss": 0.2740480041503906, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.6595725969571717, |
|
"learning_rate": 1.7361709612171312e-05, |
|
"loss": 0.2704828796386719, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.7035441034209832, |
|
"learning_rate": 1.7185823586316067e-05, |
|
"loss": 0.26128680419921874, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.7475156098847947, |
|
"learning_rate": 1.7009937560460823e-05, |
|
"loss": 0.26227166748046876, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.7914871163486061, |
|
"learning_rate": 1.6834051534605578e-05, |
|
"loss": 0.257682861328125, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.8354586228124176, |
|
"learning_rate": 1.665816550875033e-05, |
|
"loss": 0.25748046875, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.879430129276229, |
|
"learning_rate": 1.6482279482895085e-05, |
|
"loss": 0.26029693603515625, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.9234016357400404, |
|
"learning_rate": 1.630639345703984e-05, |
|
"loss": 0.25330767822265626, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.967373142203852, |
|
"learning_rate": 1.6130507431184595e-05, |
|
"loss": 0.24725408935546875, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_accuracy": 0.8963888201830325, |
|
"eval_f1": 0.8609460580912863, |
|
"eval_loss": 0.2415674775838852, |
|
"step": 11371 |
|
}, |
|
{ |
|
"epoch": 1.0113446486676634, |
|
"learning_rate": 1.5954621405329347e-05, |
|
"loss": 0.22799398803710938, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 1.0553161551314747, |
|
"learning_rate": 1.5778735379474102e-05, |
|
"loss": 0.1779903106689453, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 1.0992876615952862, |
|
"learning_rate": 1.5602849353618857e-05, |
|
"loss": 0.18363124084472657, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 1.1432591680590978, |
|
"learning_rate": 1.5426963327763612e-05, |
|
"loss": 0.183293212890625, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 1.187230674522909, |
|
"learning_rate": 1.5251077301908364e-05, |
|
"loss": 0.1881647186279297, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 1.2312021809867206, |
|
"learning_rate": 1.5075191276053118e-05, |
|
"loss": 0.193143798828125, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 1.2751736874505322, |
|
"learning_rate": 1.4899305250197873e-05, |
|
"loss": 0.186142333984375, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 1.3191451939143435, |
|
"learning_rate": 1.4723419224342628e-05, |
|
"loss": 0.18537522888183594, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 1.363116700378155, |
|
"learning_rate": 1.4547533198487383e-05, |
|
"loss": 0.18841236877441406, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 1.4070882068419663, |
|
"learning_rate": 1.4371647172632135e-05, |
|
"loss": 0.18419113159179687, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 1.4510597133057779, |
|
"learning_rate": 1.419576114677689e-05, |
|
"loss": 0.18361325073242188, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 1.4950312197695892, |
|
"learning_rate": 1.4019875120921644e-05, |
|
"loss": 0.18803219604492188, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 1.5390027262334007, |
|
"learning_rate": 1.3843989095066399e-05, |
|
"loss": 0.1816302032470703, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 1.5829742326972123, |
|
"learning_rate": 1.3668103069211152e-05, |
|
"loss": 0.1842498016357422, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 1.6269457391610236, |
|
"learning_rate": 1.3492217043355906e-05, |
|
"loss": 0.17726910400390625, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 1.6709172456248351, |
|
"learning_rate": 1.3316331017500661e-05, |
|
"loss": 0.17991775512695313, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 1.7148887520886467, |
|
"learning_rate": 1.3140444991645416e-05, |
|
"loss": 0.1825124053955078, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 1.758860258552458, |
|
"learning_rate": 1.2964558965790168e-05, |
|
"loss": 0.17683036804199218, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 1.8028317650162693, |
|
"learning_rate": 1.2788672939934923e-05, |
|
"loss": 0.1868966827392578, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 1.846803271480081, |
|
"learning_rate": 1.2612786914079678e-05, |
|
"loss": 0.18423307800292968, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 1.8907747779438924, |
|
"learning_rate": 1.2436900888224432e-05, |
|
"loss": 0.17814576721191405, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 1.9347462844077037, |
|
"learning_rate": 1.2261014862369185e-05, |
|
"loss": 0.17719683837890626, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 1.9787177908715152, |
|
"learning_rate": 1.208512883651394e-05, |
|
"loss": 0.17974095153808595, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_accuracy": 0.9078407123423201, |
|
"eval_f1": 0.8756839717069265, |
|
"eval_loss": 0.23735392093658447, |
|
"step": 22742 |
|
}, |
|
{ |
|
"epoch": 2.0226892973353268, |
|
"learning_rate": 1.1909242810658694e-05, |
|
"loss": 0.14513198852539064, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 2.066660803799138, |
|
"learning_rate": 1.1733356784803449e-05, |
|
"loss": 0.11556282806396484, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 2.1106323102629494, |
|
"learning_rate": 1.15574707589482e-05, |
|
"loss": 0.12064967346191406, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 2.154603816726761, |
|
"learning_rate": 1.1381584733092956e-05, |
|
"loss": 0.11743055725097656, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 2.1985753231905725, |
|
"learning_rate": 1.1205698707237711e-05, |
|
"loss": 0.11710655212402343, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 2.242546829654384, |
|
"learning_rate": 1.1029812681382466e-05, |
|
"loss": 0.1169861831665039, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 2.2865183361181955, |
|
"learning_rate": 1.0853926655527218e-05, |
|
"loss": 0.12172984313964844, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 2.330489842582007, |
|
"learning_rate": 1.0678040629671973e-05, |
|
"loss": 0.11213615417480469, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 2.374461349045818, |
|
"learning_rate": 1.0502154603816728e-05, |
|
"loss": 0.12687485504150392, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 2.41843285550963, |
|
"learning_rate": 1.0326268577961482e-05, |
|
"loss": 0.12601725006103515, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 2.4624043619734413, |
|
"learning_rate": 1.0150382552106235e-05, |
|
"loss": 0.11948597717285156, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 2.5063758684372526, |
|
"learning_rate": 9.974496526250989e-06, |
|
"loss": 0.12668427276611327, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 2.5503473749010643, |
|
"learning_rate": 9.798610500395744e-06, |
|
"loss": 0.12504535675048828, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 2.5943188813648757, |
|
"learning_rate": 9.6227244745405e-06, |
|
"loss": 0.12663780975341796, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 2.638290387828687, |
|
"learning_rate": 9.446838448685253e-06, |
|
"loss": 0.122175048828125, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 2.6822618942924983, |
|
"learning_rate": 9.270952422830008e-06, |
|
"loss": 0.12571288299560546, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 2.72623340075631, |
|
"learning_rate": 9.095066396974761e-06, |
|
"loss": 0.12774851989746094, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 2.7702049072201214, |
|
"learning_rate": 8.919180371119516e-06, |
|
"loss": 0.12356333923339843, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 2.8141764136839327, |
|
"learning_rate": 8.74329434526427e-06, |
|
"loss": 0.1209749755859375, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 2.8581479201477444, |
|
"learning_rate": 8.567408319409023e-06, |
|
"loss": 0.12022649383544921, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 2.9021194266115558, |
|
"learning_rate": 8.391522293553777e-06, |
|
"loss": 0.12490190887451172, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 2.946090933075367, |
|
"learning_rate": 8.215636267698532e-06, |
|
"loss": 0.12208123779296876, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 2.9900624395391784, |
|
"learning_rate": 8.039750241843286e-06, |
|
"loss": 0.12288145446777343, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_accuracy": 0.9108088053425674, |
|
"eval_f1": 0.8790176474535328, |
|
"eval_loss": 0.2841033935546875, |
|
"step": 34113 |
|
} |
|
], |
|
"max_steps": 56855, |
|
"num_train_epochs": 5, |
|
"total_flos": 49275990211670160, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|