|
{
|
|
"best_metric": 0.27312836050987244,
|
|
"best_model_checkpoint": "vit_epochs1_batch32_lr5e-05_size224_tiles10_seed1_q3_dropout_v2_test11\\checkpoint-469",
|
|
"epoch": 1.0,
|
|
"eval_steps": 500,
|
|
"global_step": 469,
|
|
"is_hyper_param_search": false,
|
|
"is_local_process_zero": true,
|
|
"is_world_process_zero": true,
|
|
"log_history": [
|
|
{
|
|
"epoch": 0.010660980810234541,
|
|
"grad_norm": 2.002079963684082,
|
|
"learning_rate": 4.9466950959488276e-05,
|
|
"loss": 0.3909,
|
|
"step": 5
|
|
},
|
|
{
|
|
"epoch": 0.021321961620469083,
|
|
"grad_norm": 0.9621203541755676,
|
|
"learning_rate": 4.893390191897655e-05,
|
|
"loss": 0.1535,
|
|
"step": 10
|
|
},
|
|
{
|
|
"epoch": 0.031982942430703626,
|
|
"grad_norm": 0.5087786912918091,
|
|
"learning_rate": 4.840085287846482e-05,
|
|
"loss": 0.0848,
|
|
"step": 15
|
|
},
|
|
{
|
|
"epoch": 0.042643923240938165,
|
|
"grad_norm": 0.3503257632255554,
|
|
"learning_rate": 4.78678038379531e-05,
|
|
"loss": 0.0567,
|
|
"step": 20
|
|
},
|
|
{
|
|
"epoch": 0.053304904051172705,
|
|
"grad_norm": 0.2416423112154007,
|
|
"learning_rate": 4.7334754797441364e-05,
|
|
"loss": 0.0344,
|
|
"step": 25
|
|
},
|
|
{
|
|
"epoch": 0.06396588486140725,
|
|
"grad_norm": 0.3675519824028015,
|
|
"learning_rate": 4.6801705756929645e-05,
|
|
"loss": 0.0263,
|
|
"step": 30
|
|
},
|
|
{
|
|
"epoch": 0.07462686567164178,
|
|
"grad_norm": 0.17501111328601837,
|
|
"learning_rate": 4.626865671641791e-05,
|
|
"loss": 0.0262,
|
|
"step": 35
|
|
},
|
|
{
|
|
"epoch": 0.08528784648187633,
|
|
"grad_norm": 1.7560172080993652,
|
|
"learning_rate": 4.5735607675906185e-05,
|
|
"loss": 0.0274,
|
|
"step": 40
|
|
},
|
|
{
|
|
"epoch": 0.09594882729211088,
|
|
"grad_norm": 0.12559843063354492,
|
|
"learning_rate": 4.520255863539446e-05,
|
|
"loss": 0.0159,
|
|
"step": 45
|
|
},
|
|
{
|
|
"epoch": 0.10660980810234541,
|
|
"grad_norm": 9.90837574005127,
|
|
"learning_rate": 4.466950959488273e-05,
|
|
"loss": 0.0229,
|
|
"step": 50
|
|
},
|
|
{
|
|
"epoch": 0.11727078891257996,
|
|
"grad_norm": 19.459095001220703,
|
|
"learning_rate": 4.4136460554371006e-05,
|
|
"loss": 0.0845,
|
|
"step": 55
|
|
},
|
|
{
|
|
"epoch": 0.1279317697228145,
|
|
"grad_norm": 0.15913185477256775,
|
|
"learning_rate": 4.360341151385928e-05,
|
|
"loss": 0.0649,
|
|
"step": 60
|
|
},
|
|
{
|
|
"epoch": 0.13859275053304904,
|
|
"grad_norm": 0.10663477331399918,
|
|
"learning_rate": 4.307036247334755e-05,
|
|
"loss": 0.0129,
|
|
"step": 65
|
|
},
|
|
{
|
|
"epoch": 0.14925373134328357,
|
|
"grad_norm": 0.09200357645750046,
|
|
"learning_rate": 4.253731343283582e-05,
|
|
"loss": 0.0284,
|
|
"step": 70
|
|
},
|
|
{
|
|
"epoch": 0.15991471215351813,
|
|
"grad_norm": 8.210641860961914,
|
|
"learning_rate": 4.2004264392324094e-05,
|
|
"loss": 0.0336,
|
|
"step": 75
|
|
},
|
|
{
|
|
"epoch": 0.17057569296375266,
|
|
"grad_norm": 0.12569181621074677,
|
|
"learning_rate": 4.147121535181237e-05,
|
|
"loss": 0.0281,
|
|
"step": 80
|
|
},
|
|
{
|
|
"epoch": 0.1812366737739872,
|
|
"grad_norm": 0.44192206859588623,
|
|
"learning_rate": 4.093816631130064e-05,
|
|
"loss": 0.0292,
|
|
"step": 85
|
|
},
|
|
{
|
|
"epoch": 0.19189765458422176,
|
|
"grad_norm": 0.15402592718601227,
|
|
"learning_rate": 4.0405117270788915e-05,
|
|
"loss": 0.0495,
|
|
"step": 90
|
|
},
|
|
{
|
|
"epoch": 0.2025586353944563,
|
|
"grad_norm": 17.383081436157227,
|
|
"learning_rate": 3.987206823027719e-05,
|
|
"loss": 0.0904,
|
|
"step": 95
|
|
},
|
|
{
|
|
"epoch": 0.21321961620469082,
|
|
"grad_norm": 0.07923204451799393,
|
|
"learning_rate": 3.9339019189765456e-05,
|
|
"loss": 0.0503,
|
|
"step": 100
|
|
},
|
|
{
|
|
"epoch": 0.22388059701492538,
|
|
"grad_norm": 0.07286439090967178,
|
|
"learning_rate": 3.8805970149253736e-05,
|
|
"loss": 0.0317,
|
|
"step": 105
|
|
},
|
|
{
|
|
"epoch": 0.2345415778251599,
|
|
"grad_norm": 1.0418739318847656,
|
|
"learning_rate": 3.8272921108742e-05,
|
|
"loss": 0.0346,
|
|
"step": 110
|
|
},
|
|
{
|
|
"epoch": 0.24520255863539445,
|
|
"grad_norm": 3.832359552383423,
|
|
"learning_rate": 3.7739872068230284e-05,
|
|
"loss": 0.0988,
|
|
"step": 115
|
|
},
|
|
{
|
|
"epoch": 0.255863539445629,
|
|
"grad_norm": 0.06542658805847168,
|
|
"learning_rate": 3.720682302771855e-05,
|
|
"loss": 0.0075,
|
|
"step": 120
|
|
},
|
|
{
|
|
"epoch": 0.26652452025586354,
|
|
"grad_norm": 9.610258102416992,
|
|
"learning_rate": 3.6673773987206824e-05,
|
|
"loss": 0.0269,
|
|
"step": 125
|
|
},
|
|
{
|
|
"epoch": 0.2771855010660981,
|
|
"grad_norm": 0.17276552319526672,
|
|
"learning_rate": 3.61407249466951e-05,
|
|
"loss": 0.0231,
|
|
"step": 130
|
|
},
|
|
{
|
|
"epoch": 0.2878464818763326,
|
|
"grad_norm": 0.08185400068759918,
|
|
"learning_rate": 3.560767590618337e-05,
|
|
"loss": 0.0755,
|
|
"step": 135
|
|
},
|
|
{
|
|
"epoch": 0.29850746268656714,
|
|
"grad_norm": 0.07418997585773468,
|
|
"learning_rate": 3.5074626865671645e-05,
|
|
"loss": 0.0424,
|
|
"step": 140
|
|
},
|
|
{
|
|
"epoch": 0.3091684434968017,
|
|
"grad_norm": 2.326040267944336,
|
|
"learning_rate": 3.454157782515991e-05,
|
|
"loss": 0.1041,
|
|
"step": 145
|
|
},
|
|
{
|
|
"epoch": 0.31982942430703626,
|
|
"grad_norm": 1.4266926050186157,
|
|
"learning_rate": 3.400852878464819e-05,
|
|
"loss": 0.0466,
|
|
"step": 150
|
|
},
|
|
{
|
|
"epoch": 0.3304904051172708,
|
|
"grad_norm": 25.644779205322266,
|
|
"learning_rate": 3.347547974413646e-05,
|
|
"loss": 0.0492,
|
|
"step": 155
|
|
},
|
|
{
|
|
"epoch": 0.3411513859275053,
|
|
"grad_norm": 4.534906387329102,
|
|
"learning_rate": 3.294243070362473e-05,
|
|
"loss": 0.033,
|
|
"step": 160
|
|
},
|
|
{
|
|
"epoch": 0.35181236673773986,
|
|
"grad_norm": 10.8761625289917,
|
|
"learning_rate": 3.240938166311301e-05,
|
|
"loss": 0.0232,
|
|
"step": 165
|
|
},
|
|
{
|
|
"epoch": 0.3624733475479744,
|
|
"grad_norm": 0.3756234645843506,
|
|
"learning_rate": 3.187633262260128e-05,
|
|
"loss": 0.0225,
|
|
"step": 170
|
|
},
|
|
{
|
|
"epoch": 0.373134328358209,
|
|
"grad_norm": 11.933363914489746,
|
|
"learning_rate": 3.1343283582089554e-05,
|
|
"loss": 0.0702,
|
|
"step": 175
|
|
},
|
|
{
|
|
"epoch": 0.3837953091684435,
|
|
"grad_norm": 2.681648015975952,
|
|
"learning_rate": 3.081023454157783e-05,
|
|
"loss": 0.1197,
|
|
"step": 180
|
|
},
|
|
{
|
|
"epoch": 0.39445628997867804,
|
|
"grad_norm": 0.5134532451629639,
|
|
"learning_rate": 3.0277185501066102e-05,
|
|
"loss": 0.0339,
|
|
"step": 185
|
|
},
|
|
{
|
|
"epoch": 0.4051172707889126,
|
|
"grad_norm": 13.620699882507324,
|
|
"learning_rate": 2.9744136460554372e-05,
|
|
"loss": 0.0276,
|
|
"step": 190
|
|
},
|
|
{
|
|
"epoch": 0.4157782515991471,
|
|
"grad_norm": 0.09190113842487335,
|
|
"learning_rate": 2.9211087420042642e-05,
|
|
"loss": 0.0086,
|
|
"step": 195
|
|
},
|
|
{
|
|
"epoch": 0.42643923240938164,
|
|
"grad_norm": 0.10281543433666229,
|
|
"learning_rate": 2.867803837953092e-05,
|
|
"loss": 0.0129,
|
|
"step": 200
|
|
},
|
|
{
|
|
"epoch": 0.43710021321961623,
|
|
"grad_norm": 6.414993762969971,
|
|
"learning_rate": 2.814498933901919e-05,
|
|
"loss": 0.0555,
|
|
"step": 205
|
|
},
|
|
{
|
|
"epoch": 0.44776119402985076,
|
|
"grad_norm": 0.06216156855225563,
|
|
"learning_rate": 2.7611940298507467e-05,
|
|
"loss": 0.0076,
|
|
"step": 210
|
|
},
|
|
{
|
|
"epoch": 0.4584221748400853,
|
|
"grad_norm": 12.08731746673584,
|
|
"learning_rate": 2.7078891257995737e-05,
|
|
"loss": 0.0403,
|
|
"step": 215
|
|
},
|
|
{
|
|
"epoch": 0.4690831556503198,
|
|
"grad_norm": 0.10379679501056671,
|
|
"learning_rate": 2.6545842217484007e-05,
|
|
"loss": 0.0404,
|
|
"step": 220
|
|
},
|
|
{
|
|
"epoch": 0.47974413646055436,
|
|
"grad_norm": 0.06435238569974899,
|
|
"learning_rate": 2.6012793176972285e-05,
|
|
"loss": 0.0072,
|
|
"step": 225
|
|
},
|
|
{
|
|
"epoch": 0.4904051172707889,
|
|
"grad_norm": 0.07799974828958511,
|
|
"learning_rate": 2.5479744136460555e-05,
|
|
"loss": 0.0209,
|
|
"step": 230
|
|
},
|
|
{
|
|
"epoch": 0.5010660980810234,
|
|
"grad_norm": 0.05462189391255379,
|
|
"learning_rate": 2.494669509594883e-05,
|
|
"loss": 0.006,
|
|
"step": 235
|
|
},
|
|
{
|
|
"epoch": 0.511727078891258,
|
|
"grad_norm": 0.0564899705350399,
|
|
"learning_rate": 2.44136460554371e-05,
|
|
"loss": 0.0277,
|
|
"step": 240
|
|
},
|
|
{
|
|
"epoch": 0.5223880597014925,
|
|
"grad_norm": 0.6151730418205261,
|
|
"learning_rate": 2.3880597014925373e-05,
|
|
"loss": 0.0118,
|
|
"step": 245
|
|
},
|
|
{
|
|
"epoch": 0.5330490405117271,
|
|
"grad_norm": 3.0278522968292236,
|
|
"learning_rate": 2.3347547974413646e-05,
|
|
"loss": 0.0635,
|
|
"step": 250
|
|
},
|
|
{
|
|
"epoch": 0.5437100213219617,
|
|
"grad_norm": 23.011789321899414,
|
|
"learning_rate": 2.281449893390192e-05,
|
|
"loss": 0.0231,
|
|
"step": 255
|
|
},
|
|
{
|
|
"epoch": 0.5543710021321961,
|
|
"grad_norm": 2.13313889503479,
|
|
"learning_rate": 2.2281449893390194e-05,
|
|
"loss": 0.0391,
|
|
"step": 260
|
|
},
|
|
{
|
|
"epoch": 0.5650319829424307,
|
|
"grad_norm": 0.14648886024951935,
|
|
"learning_rate": 2.1748400852878467e-05,
|
|
"loss": 0.0054,
|
|
"step": 265
|
|
},
|
|
{
|
|
"epoch": 0.5756929637526652,
|
|
"grad_norm": 0.046188462525606155,
|
|
"learning_rate": 2.1215351812366738e-05,
|
|
"loss": 0.0069,
|
|
"step": 270
|
|
},
|
|
{
|
|
"epoch": 0.5863539445628998,
|
|
"grad_norm": 0.047539375722408295,
|
|
"learning_rate": 2.068230277185501e-05,
|
|
"loss": 0.0067,
|
|
"step": 275
|
|
},
|
|
{
|
|
"epoch": 0.5970149253731343,
|
|
"grad_norm": 0.051974523812532425,
|
|
"learning_rate": 2.0149253731343285e-05,
|
|
"loss": 0.0055,
|
|
"step": 280
|
|
},
|
|
{
|
|
"epoch": 0.6076759061833689,
|
|
"grad_norm": 0.049149394035339355,
|
|
"learning_rate": 1.961620469083156e-05,
|
|
"loss": 0.0053,
|
|
"step": 285
|
|
},
|
|
{
|
|
"epoch": 0.6183368869936035,
|
|
"grad_norm": 0.041319601237773895,
|
|
"learning_rate": 1.9083155650319832e-05,
|
|
"loss": 0.0221,
|
|
"step": 290
|
|
},
|
|
{
|
|
"epoch": 0.6289978678038379,
|
|
"grad_norm": 0.05550704523921013,
|
|
"learning_rate": 1.8550106609808106e-05,
|
|
"loss": 0.0212,
|
|
"step": 295
|
|
},
|
|
{
|
|
"epoch": 0.6396588486140725,
|
|
"grad_norm": 0.04463819041848183,
|
|
"learning_rate": 1.8017057569296376e-05,
|
|
"loss": 0.0348,
|
|
"step": 300
|
|
},
|
|
{
|
|
"epoch": 0.650319829424307,
|
|
"grad_norm": 2.7232251167297363,
|
|
"learning_rate": 1.7484008528784647e-05,
|
|
"loss": 0.0339,
|
|
"step": 305
|
|
},
|
|
{
|
|
"epoch": 0.6609808102345416,
|
|
"grad_norm": 0.04294276237487793,
|
|
"learning_rate": 1.695095948827292e-05,
|
|
"loss": 0.007,
|
|
"step": 310
|
|
},
|
|
{
|
|
"epoch": 0.6716417910447762,
|
|
"grad_norm": 0.04751422628760338,
|
|
"learning_rate": 1.6417910447761194e-05,
|
|
"loss": 0.0058,
|
|
"step": 315
|
|
},
|
|
{
|
|
"epoch": 0.6823027718550106,
|
|
"grad_norm": 0.06318726390600204,
|
|
"learning_rate": 1.5884861407249468e-05,
|
|
"loss": 0.0351,
|
|
"step": 320
|
|
},
|
|
{
|
|
"epoch": 0.6929637526652452,
|
|
"grad_norm": 0.04593655467033386,
|
|
"learning_rate": 1.535181236673774e-05,
|
|
"loss": 0.0084,
|
|
"step": 325
|
|
},
|
|
{
|
|
"epoch": 0.7036247334754797,
|
|
"grad_norm": 0.04337684437632561,
|
|
"learning_rate": 1.4818763326226012e-05,
|
|
"loss": 0.0048,
|
|
"step": 330
|
|
},
|
|
{
|
|
"epoch": 0.7142857142857143,
|
|
"grad_norm": 1.3650192022323608,
|
|
"learning_rate": 1.4285714285714285e-05,
|
|
"loss": 0.0054,
|
|
"step": 335
|
|
},
|
|
{
|
|
"epoch": 0.7249466950959488,
|
|
"grad_norm": 0.04110950231552124,
|
|
"learning_rate": 1.3752665245202559e-05,
|
|
"loss": 0.0048,
|
|
"step": 340
|
|
},
|
|
{
|
|
"epoch": 0.7356076759061834,
|
|
"grad_norm": 0.05691000819206238,
|
|
"learning_rate": 1.3219616204690833e-05,
|
|
"loss": 0.005,
|
|
"step": 345
|
|
},
|
|
{
|
|
"epoch": 0.746268656716418,
|
|
"grad_norm": 18.876996994018555,
|
|
"learning_rate": 1.2686567164179105e-05,
|
|
"loss": 0.031,
|
|
"step": 350
|
|
},
|
|
{
|
|
"epoch": 0.7569296375266524,
|
|
"grad_norm": 0.8067087531089783,
|
|
"learning_rate": 1.2153518123667377e-05,
|
|
"loss": 0.0403,
|
|
"step": 355
|
|
},
|
|
{
|
|
"epoch": 0.767590618336887,
|
|
"grad_norm": 0.04626353457570076,
|
|
"learning_rate": 1.162046908315565e-05,
|
|
"loss": 0.0049,
|
|
"step": 360
|
|
},
|
|
{
|
|
"epoch": 0.7782515991471215,
|
|
"grad_norm": 0.04179658740758896,
|
|
"learning_rate": 1.1087420042643924e-05,
|
|
"loss": 0.0047,
|
|
"step": 365
|
|
},
|
|
{
|
|
"epoch": 0.7889125799573561,
|
|
"grad_norm": 0.05049879848957062,
|
|
"learning_rate": 1.0554371002132196e-05,
|
|
"loss": 0.0049,
|
|
"step": 370
|
|
},
|
|
{
|
|
"epoch": 0.7995735607675906,
|
|
"grad_norm": 28.65748405456543,
|
|
"learning_rate": 1.002132196162047e-05,
|
|
"loss": 0.0207,
|
|
"step": 375
|
|
},
|
|
{
|
|
"epoch": 0.8102345415778252,
|
|
"grad_norm": 0.05166960135102272,
|
|
"learning_rate": 9.488272921108744e-06,
|
|
"loss": 0.0268,
|
|
"step": 380
|
|
},
|
|
{
|
|
"epoch": 0.8208955223880597,
|
|
"grad_norm": 0.03881421312689781,
|
|
"learning_rate": 8.955223880597016e-06,
|
|
"loss": 0.0206,
|
|
"step": 385
|
|
},
|
|
{
|
|
"epoch": 0.8315565031982942,
|
|
"grad_norm": 0.0383550263941288,
|
|
"learning_rate": 8.422174840085288e-06,
|
|
"loss": 0.0043,
|
|
"step": 390
|
|
},
|
|
{
|
|
"epoch": 0.8422174840085288,
|
|
"grad_norm": 0.04295238107442856,
|
|
"learning_rate": 7.889125799573561e-06,
|
|
"loss": 0.0042,
|
|
"step": 395
|
|
},
|
|
{
|
|
"epoch": 0.8528784648187633,
|
|
"grad_norm": 1.7127622365951538,
|
|
"learning_rate": 7.356076759061833e-06,
|
|
"loss": 0.0053,
|
|
"step": 400
|
|
},
|
|
{
|
|
"epoch": 0.8635394456289979,
|
|
"grad_norm": 0.049339987337589264,
|
|
"learning_rate": 6.823027718550107e-06,
|
|
"loss": 0.0044,
|
|
"step": 405
|
|
},
|
|
{
|
|
"epoch": 0.8742004264392325,
|
|
"grad_norm": 0.06354337185621262,
|
|
"learning_rate": 6.28997867803838e-06,
|
|
"loss": 0.0045,
|
|
"step": 410
|
|
},
|
|
{
|
|
"epoch": 0.8848614072494669,
|
|
"grad_norm": 0.1981351524591446,
|
|
"learning_rate": 5.756929637526653e-06,
|
|
"loss": 0.0054,
|
|
"step": 415
|
|
},
|
|
{
|
|
"epoch": 0.8955223880597015,
|
|
"grad_norm": 0.03931286185979843,
|
|
"learning_rate": 5.2238805970149255e-06,
|
|
"loss": 0.0208,
|
|
"step": 420
|
|
},
|
|
{
|
|
"epoch": 0.906183368869936,
|
|
"grad_norm": 0.0348835252225399,
|
|
"learning_rate": 4.690831556503199e-06,
|
|
"loss": 0.0157,
|
|
"step": 425
|
|
},
|
|
{
|
|
"epoch": 0.9168443496801706,
|
|
"grad_norm": 0.03531981259584427,
|
|
"learning_rate": 4.157782515991471e-06,
|
|
"loss": 0.0063,
|
|
"step": 430
|
|
},
|
|
{
|
|
"epoch": 0.9275053304904051,
|
|
"grad_norm": 4.256789684295654,
|
|
"learning_rate": 3.624733475479744e-06,
|
|
"loss": 0.0051,
|
|
"step": 435
|
|
},
|
|
{
|
|
"epoch": 0.9381663113006397,
|
|
"grad_norm": 0.03756505623459816,
|
|
"learning_rate": 3.0916844349680173e-06,
|
|
"loss": 0.0045,
|
|
"step": 440
|
|
},
|
|
{
|
|
"epoch": 0.9488272921108742,
|
|
"grad_norm": 0.16175320744514465,
|
|
"learning_rate": 2.55863539445629e-06,
|
|
"loss": 0.0075,
|
|
"step": 445
|
|
},
|
|
{
|
|
"epoch": 0.9594882729211087,
|
|
"grad_norm": 0.03726266324520111,
|
|
"learning_rate": 2.025586353944563e-06,
|
|
"loss": 0.0041,
|
|
"step": 450
|
|
},
|
|
{
|
|
"epoch": 0.9701492537313433,
|
|
"grad_norm": 19.503908157348633,
|
|
"learning_rate": 1.4925373134328358e-06,
|
|
"loss": 0.0224,
|
|
"step": 455
|
|
},
|
|
{
|
|
"epoch": 0.9808102345415778,
|
|
"grad_norm": 0.04280271381139755,
|
|
"learning_rate": 9.594882729211088e-07,
|
|
"loss": 0.0602,
|
|
"step": 460
|
|
},
|
|
{
|
|
"epoch": 0.9914712153518124,
|
|
"grad_norm": 0.03506707400083542,
|
|
"learning_rate": 4.264392324093817e-07,
|
|
"loss": 0.0233,
|
|
"step": 465
|
|
},
|
|
{
|
|
"epoch": 1.0,
|
|
"eval_accuracy": 0.9381333333333334,
|
|
"eval_loss": 0.27312836050987244,
|
|
"eval_runtime": 18.3733,
|
|
"eval_samples_per_second": 204.1,
|
|
"eval_steps_per_second": 6.422,
|
|
"step": 469
|
|
},
|
|
{
|
|
"epoch": 1.0,
|
|
"step": 469,
|
|
"total_flos": 1.16237984421888e+18,
|
|
"train_loss": 0.03354339535508964,
|
|
"train_runtime": 172.9262,
|
|
"train_samples_per_second": 86.742,
|
|
"train_steps_per_second": 2.712
|
|
}
|
|
],
|
|
"logging_steps": 5,
|
|
"max_steps": 469,
|
|
"num_input_tokens_seen": 0,
|
|
"num_train_epochs": 1,
|
|
"save_steps": 500,
|
|
"stateful_callbacks": {
|
|
"TrainerControl": {
|
|
"args": {
|
|
"should_epoch_stop": false,
|
|
"should_evaluate": false,
|
|
"should_log": false,
|
|
"should_save": true,
|
|
"should_training_stop": true
|
|
},
|
|
"attributes": {}
|
|
}
|
|
},
|
|
"total_flos": 1.16237984421888e+18,
|
|
"train_batch_size": 32,
|
|
"trial_name": null,
|
|
"trial_params": null
|
|
}
|
|
|