{
  "best_metric": null,
  "best_model_checkpoint": null,
  "epoch": 4.987212276214834,
  "eval_steps": 500,
  "global_step": 975,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.05115089514066496,
      "grad_norm": 214.7565900278413,
      "learning_rate": 1.0204081632653063e-06,
      "loss": 4.0075,
      "step": 10
    },
    {
      "epoch": 0.10230179028132992,
      "grad_norm": 22.211069433232247,
      "learning_rate": 2.0408163265306125e-06,
      "loss": 0.6282,
      "step": 20
    },
    {
      "epoch": 0.1534526854219949,
      "grad_norm": 12.11809889509627,
      "learning_rate": 3.0612244897959185e-06,
      "loss": 0.2107,
      "step": 30
    },
    {
      "epoch": 0.20460358056265984,
      "grad_norm": 2.8569999074326495,
      "learning_rate": 4.081632653061225e-06,
      "loss": 0.2135,
      "step": 40
    },
    {
      "epoch": 0.2557544757033248,
      "grad_norm": 6.867087399363433,
      "learning_rate": 5.1020408163265315e-06,
      "loss": 0.2072,
      "step": 50
    },
    {
      "epoch": 0.3069053708439898,
      "grad_norm": 3.4564712619293663,
      "learning_rate": 6.122448979591837e-06,
      "loss": 0.2002,
      "step": 60
    },
    {
      "epoch": 0.35805626598465473,
      "grad_norm": 3.69055204043271,
      "learning_rate": 7.1428571428571436e-06,
      "loss": 0.2,
      "step": 70
    },
    {
      "epoch": 0.4092071611253197,
      "grad_norm": 1.0696324041329028,
      "learning_rate": 8.16326530612245e-06,
      "loss": 0.1964,
      "step": 80
    },
    {
      "epoch": 0.46035805626598464,
      "grad_norm": 2.964347091843066,
      "learning_rate": 9.183673469387756e-06,
      "loss": 0.1952,
      "step": 90
    },
    {
      "epoch": 0.5115089514066496,
      "grad_norm": 3.646664998250867,
      "learning_rate": 9.99987167871469e-06,
      "loss": 0.2045,
      "step": 100
    },
    {
      "epoch": 0.5626598465473146,
      "grad_norm": 1.6191665925446022,
      "learning_rate": 9.995381125277747e-06,
      "loss": 0.2018,
      "step": 110
    },
    {
      "epoch": 0.6138107416879796,
      "grad_norm": 4.017961040145252,
      "learning_rate": 9.984481092596683e-06,
      "loss": 0.1975,
      "step": 120
    },
    {
      "epoch": 0.6649616368286445,
      "grad_norm": 2.9108813585819817,
      "learning_rate": 9.967185566297713e-06,
      "loss": 0.199,
      "step": 130
    },
    {
      "epoch": 0.7161125319693095,
      "grad_norm": 3.1021824472509767,
      "learning_rate": 9.94351673794421e-06,
      "loss": 0.1978,
      "step": 140
    },
    {
      "epoch": 0.7672634271099744,
      "grad_norm": 3.497183421035605,
      "learning_rate": 9.913504976563138e-06,
      "loss": 0.1972,
      "step": 150
    },
    {
      "epoch": 0.8184143222506394,
      "grad_norm": 2.667271306285725,
      "learning_rate": 9.877188789679111e-06,
      "loss": 0.1972,
      "step": 160
    },
    {
      "epoch": 0.8695652173913043,
      "grad_norm": 1.939938871403963,
      "learning_rate": 9.834614773906125e-06,
      "loss": 0.1983,
      "step": 170
    },
    {
      "epoch": 0.9207161125319693,
      "grad_norm": 3.277095189836375,
      "learning_rate": 9.785837555160322e-06,
      "loss": 0.2001,
      "step": 180
    },
    {
      "epoch": 0.9718670076726342,
      "grad_norm": 2.1349632530669784,
      "learning_rate": 9.730919718570501e-06,
      "loss": 0.1985,
      "step": 190
    },
    {
      "epoch": 1.0230179028132993,
      "grad_norm": 1.9921171293582758,
      "learning_rate": 9.669931728176329e-06,
      "loss": 0.2044,
      "step": 200
    },
    {
      "epoch": 1.0741687979539642,
      "grad_norm": 1.6800643640667052,
      "learning_rate": 9.602951836517255e-06,
      "loss": 0.1968,
      "step": 210
    },
    {
      "epoch": 1.1253196930946292,
      "grad_norm": 3.6566429458754515,
      "learning_rate": 9.530065984228156e-06,
      "loss": 0.1972,
      "step": 220
    },
    {
      "epoch": 1.1764705882352942,
      "grad_norm": 1.0633803952953227,
      "learning_rate": 9.451367689770532e-06,
      "loss": 0.1957,
      "step": 230
    },
    {
      "epoch": 1.227621483375959,
      "grad_norm": 1.4368952344952126,
      "learning_rate": 9.366957929440731e-06,
      "loss": 0.1942,
      "step": 240
    },
    {
      "epoch": 1.278772378516624,
      "grad_norm": 2.8115853641591015,
      "learning_rate": 9.276945007809167e-06,
      "loss": 0.1975,
      "step": 250
    },
    {
      "epoch": 1.329923273657289,
      "grad_norm": 2.2368726208053515,
      "learning_rate": 9.181444418756774e-06,
      "loss": 0.1941,
      "step": 260
    },
    {
      "epoch": 1.381074168797954,
      "grad_norm": 1.718287417414397,
      "learning_rate": 9.080578697286986e-06,
      "loss": 0.1955,
      "step": 270
    },
    {
      "epoch": 1.432225063938619,
      "grad_norm": 2.6480547516752373,
      "learning_rate": 8.974477262303382e-06,
      "loss": 0.1965,
      "step": 280
    },
    {
      "epoch": 1.4833759590792839,
      "grad_norm": 2.1323772112759856,
      "learning_rate": 8.86327625055474e-06,
      "loss": 0.1928,
      "step": 290
    },
    {
      "epoch": 1.5345268542199488,
      "grad_norm": 1.3695013623025927,
      "learning_rate": 8.747118341960542e-06,
      "loss": 0.1964,
      "step": 300
    },
    {
      "epoch": 1.5856777493606138,
      "grad_norm": 2.816490216879556,
      "learning_rate": 8.626152576541059e-06,
      "loss": 0.1926,
      "step": 310
    },
    {
      "epoch": 1.6368286445012787,
      "grad_norm": 2.63475775763493,
      "learning_rate": 8.500534163186914e-06,
      "loss": 0.1944,
      "step": 320
    },
    {
      "epoch": 1.6879795396419437,
      "grad_norm": 1.2976999470451909,
      "learning_rate": 8.37042428051349e-06,
      "loss": 0.1935,
      "step": 330
    },
    {
      "epoch": 1.7391304347826086,
      "grad_norm": 3.0528789638369314,
      "learning_rate": 8.235989870055684e-06,
      "loss": 0.1931,
      "step": 340
    },
    {
      "epoch": 1.7902813299232738,
      "grad_norm": 2.761510505813076,
      "learning_rate": 8.097403422068377e-06,
      "loss": 0.1901,
      "step": 350
    },
    {
      "epoch": 1.8414322250639388,
      "grad_norm": 1.2860400398022556,
      "learning_rate": 7.954842754207442e-06,
      "loss": 0.1846,
      "step": 360
    },
    {
      "epoch": 1.8925831202046037,
      "grad_norm": 1.640012696939897,
      "learning_rate": 7.80849078337527e-06,
      "loss": 0.1801,
      "step": 370
    },
    {
      "epoch": 1.9437340153452687,
      "grad_norm": 2.959712367273353,
      "learning_rate": 7.658535291023533e-06,
      "loss": 0.1766,
      "step": 380
    },
    {
      "epoch": 1.9948849104859336,
      "grad_norm": 2.582572305335923,
      "learning_rate": 7.5051686822143614e-06,
      "loss": 0.1592,
      "step": 390
    },
    {
      "epoch": 2.0460358056265986,
      "grad_norm": 2.8473557994991165,
      "learning_rate": 7.348587738749037e-06,
      "loss": 0.1467,
      "step": 400
    },
    {
      "epoch": 2.0971867007672635,
      "grad_norm": 2.246956202126538,
      "learning_rate": 7.18899336668097e-06,
      "loss": 0.1102,
      "step": 410
    },
    {
      "epoch": 2.1483375959079285,
      "grad_norm": 2.3922923060664685,
      "learning_rate": 7.0265903385369385e-06,
      "loss": 0.0799,
      "step": 420
    },
    {
      "epoch": 2.1994884910485935,
      "grad_norm": 2.235940967501667,
      "learning_rate": 6.861587030577326e-06,
      "loss": 0.0643,
      "step": 430
    },
    {
      "epoch": 2.2506393861892584,
      "grad_norm": 2.1749048219884615,
      "learning_rate": 6.694195155432473e-06,
      "loss": 0.039,
      "step": 440
    },
    {
      "epoch": 2.3017902813299234,
      "grad_norm": 2.4706549250013494,
      "learning_rate": 6.524629490458193e-06,
      "loss": 0.0287,
      "step": 450
    },
    {
      "epoch": 2.3529411764705883,
      "grad_norm": 2.1875953818358975,
      "learning_rate": 6.353107602159005e-06,
      "loss": 0.0211,
      "step": 460
    },
    {
      "epoch": 2.4040920716112533,
      "grad_norm": 1.1932819949506583,
      "learning_rate": 6.179849567032658e-06,
      "loss": 0.0199,
      "step": 470
    },
    {
      "epoch": 2.455242966751918,
      "grad_norm": 3.043082138284078,
      "learning_rate": 6.0050776891941266e-06,
      "loss": 0.0138,
      "step": 480
    },
    {
      "epoch": 2.506393861892583,
      "grad_norm": 1.4066811981696432,
      "learning_rate": 5.8290162151414e-06,
      "loss": 0.0102,
      "step": 490
    },
    {
      "epoch": 2.557544757033248,
      "grad_norm": 1.089275292286627,
      "learning_rate": 5.651891046029031e-06,
      "loss": 0.0091,
      "step": 500
    },
    {
      "epoch": 2.608695652173913,
      "grad_norm": 1.4386597224188378,
      "learning_rate": 5.473929447818628e-06,
      "loss": 0.0093,
      "step": 510
    },
    {
      "epoch": 2.659846547314578,
      "grad_norm": 1.6800741407957291,
      "learning_rate": 5.295359759678187e-06,
      "loss": 0.0062,
      "step": 520
    },
    {
      "epoch": 2.710997442455243,
      "grad_norm": 2.0143218617779004,
      "learning_rate": 5.1164111010044225e-06,
      "loss": 0.0058,
      "step": 530
    },
    {
      "epoch": 2.762148337595908,
      "grad_norm": 2.927032625013264,
      "learning_rate": 4.937313077443985e-06,
      "loss": 0.0071,
      "step": 540
    },
    {
      "epoch": 2.813299232736573,
      "grad_norm": 0.7049312278321461,
      "learning_rate": 4.7582954862907896e-06,
      "loss": 0.0041,
      "step": 550
    },
    {
      "epoch": 2.864450127877238,
      "grad_norm": 0.39393926121790496,
      "learning_rate": 4.579588021637448e-06,
      "loss": 0.0029,
      "step": 560
    },
    {
      "epoch": 2.915601023017903,
      "grad_norm": 0.7023810428810136,
      "learning_rate": 4.401419979659119e-06,
      "loss": 0.0036,
      "step": 570
    },
    {
      "epoch": 2.9667519181585678,
      "grad_norm": 1.7272328840467275,
      "learning_rate": 4.224019964407902e-06,
      "loss": 0.0028,
      "step": 580
    },
    {
      "epoch": 3.0179028132992327,
      "grad_norm": 0.19616140979782443,
      "learning_rate": 4.047615594495299e-06,
      "loss": 0.0017,
      "step": 590
    },
    {
      "epoch": 3.0690537084398977,
      "grad_norm": 0.060475506150692124,
      "learning_rate": 3.872433211039067e-06,
      "loss": 0.0013,
      "step": 600
    },
    {
      "epoch": 3.1202046035805626,
      "grad_norm": 0.060879802064906625,
      "learning_rate": 3.698697587249208e-06,
      "loss": 0.0012,
      "step": 610
    },
    {
      "epoch": 3.1713554987212276,
      "grad_norm": 0.8953617552425924,
      "learning_rate": 3.526631640025697e-06,
      "loss": 0.0009,
      "step": 620
    },
    {
      "epoch": 3.2225063938618925,
      "grad_norm": 0.400415376801392,
      "learning_rate": 3.3564561439380206e-06,
      "loss": 0.0006,
      "step": 630
    },
    {
      "epoch": 3.2736572890025575,
      "grad_norm": 0.13846989650131333,
      "learning_rate": 3.1883894479534882e-06,
      "loss": 0.0005,
      "step": 640
    },
    {
      "epoch": 3.3248081841432224,
      "grad_norm": 0.03190740929655757,
      "learning_rate": 3.022647195277799e-06,
      "loss": 0.0009,
      "step": 650
    },
    {
      "epoch": 3.3759590792838874,
      "grad_norm": 0.01044371613769382,
      "learning_rate": 2.8594420466673133e-06,
      "loss": 0.0001,
      "step": 660
    },
    {
      "epoch": 3.4271099744245523,
      "grad_norm": 0.021747528690115973,
      "learning_rate": 2.6989834075680452e-06,
      "loss": 0.0003,
      "step": 670
    },
    {
      "epoch": 3.4782608695652173,
      "grad_norm": 0.003734147207667446,
      "learning_rate": 2.541477159431489e-06,
      "loss": 0.0002,
      "step": 680
    },
    {
      "epoch": 3.5294117647058822,
      "grad_norm": 0.02578812155709838,
      "learning_rate": 2.3871253955520163e-06,
      "loss": 0.0003,
      "step": 690
    },
    {
      "epoch": 3.580562659846547,
      "grad_norm": 0.07582199264749429,
      "learning_rate": 2.2361261617647873e-06,
      "loss": 0.0001,
      "step": 700
    },
    {
      "epoch": 3.631713554987212,
      "grad_norm": 0.003064177461855074,
      "learning_rate": 2.0886732023368637e-06,
      "loss": 0.0001,
      "step": 710
    },
    {
      "epoch": 3.682864450127877,
      "grad_norm": 0.0034307231208645776,
      "learning_rate": 1.9449557113776157e-06,
      "loss": 0.0,
      "step": 720
    },
    {
      "epoch": 3.734015345268542,
      "grad_norm": 0.003406179275735569,
      "learning_rate": 1.8051580900873112e-06,
      "loss": 0.0003,
      "step": 730
    },
    {
      "epoch": 3.785166240409207,
      "grad_norm": 0.0017568341440027282,
      "learning_rate": 1.6694597101554354e-06,
      "loss": 0.0001,
      "step": 740
    },
    {
      "epoch": 3.836317135549872,
      "grad_norm": 0.001639728689373477,
      "learning_rate": 1.5380346836122429e-06,
      "loss": 0.0001,
      "step": 750
    },
    {
      "epoch": 3.887468030690537,
      "grad_norm": 0.001901581956312772,
      "learning_rate": 1.4110516394289103e-06,
      "loss": 0.0001,
      "step": 760
    },
    {
      "epoch": 3.938618925831202,
      "grad_norm": 0.001484483455691581,
      "learning_rate": 1.2886735071528829e-06,
      "loss": 0.0001,
      "step": 770
    },
    {
      "epoch": 3.9897698209718673,
      "grad_norm": 0.0018431338087995083,
      "learning_rate": 1.1710573078560406e-06,
      "loss": 0.0001,
      "step": 780
    },
    {
      "epoch": 4.040920716112532,
      "grad_norm": 0.06550643685551072,
      "learning_rate": 1.058353952663937e-06,
      "loss": 0.0002,
      "step": 790
    },
    {
      "epoch": 4.092071611253197,
      "grad_norm": 0.06509818369816833,
      "learning_rate": 9.50708049124568e-07,
      "loss": 0.0001,
      "step": 800
    },
    {
      "epoch": 4.143222506393862,
      "grad_norm": 0.001965866467798876,
      "learning_rate": 8.482577156651683e-07,
      "loss": 0.0001,
      "step": 810
    },
    {
      "epoch": 4.194373401534527,
      "grad_norm": 0.003659181883344629,
      "learning_rate": 7.511344043750546e-07,
      "loss": 0.0001,
      "step": 820
    },
    {
      "epoch": 4.245524296675192,
      "grad_norm": 0.0008421352975079428,
      "learning_rate": 6.594627323419383e-07,
      "loss": 0.0001,
      "step": 830
    },
    {
      "epoch": 4.296675191815857,
      "grad_norm": 0.0015741296233894467,
      "learning_rate": 5.733603217580885e-07,
      "loss": 0.0001,
      "step": 840
    },
    {
      "epoch": 4.3478260869565215,
      "grad_norm": 0.07134647602151516,
      "learning_rate": 4.929376490015148e-07,
      "loss": 0.0001,
      "step": 850
    },
    {
      "epoch": 4.398976982097187,
      "grad_norm": 0.0009193314183040581,
      "learning_rate": 4.1829790288581694e-07,
      "loss": 0.0,
      "step": 860
    },
    {
      "epoch": 4.450127877237851,
      "grad_norm": 0.0011078081696583465,
      "learning_rate": 3.495368522605602e-07,
      "loss": 0.0,
      "step": 870
    },
    {
      "epoch": 4.501278772378517,
      "grad_norm": 0.002913655514720279,
      "learning_rate": 2.867427231320774e-07,
      "loss": 0.0001,
      "step": 880
    },
    {
      "epoch": 4.552429667519181,
      "grad_norm": 0.0011670444227254474,
      "learning_rate": 2.2999608546234063e-07,
      "loss": 0.0001,
      "step": 890
    },
    {
      "epoch": 4.603580562659847,
      "grad_norm": 0.0010703222073159705,
      "learning_rate": 1.7936974979116505e-07,
      "loss": 0.0001,
      "step": 900
    },
    {
      "epoch": 4.654731457800511,
      "grad_norm": 0.0010341753002907296,
      "learning_rate": 1.349286738143829e-07,
      "loss": 0.0001,
      "step": 910
    },
    {
      "epoch": 4.705882352941177,
      "grad_norm": 0.12688243780881078,
      "learning_rate": 9.672987903784237e-08,
      "loss": 0.0002,
      "step": 920
    },
    {
      "epoch": 4.757033248081841,
      "grad_norm": 0.0013506462774850975,
      "learning_rate": 6.482237761419652e-08,
      "loss": 0.0001,
      "step": 930
    },
    {
      "epoch": 4.8081841432225065,
      "grad_norm": 0.009021779442771162,
      "learning_rate": 3.9247109456332925e-08,
      "loss": 0.0002,
      "step": 940
    },
    {
      "epoch": 4.859335038363171,
      "grad_norm": 0.0010536719662307653,
      "learning_rate": 2.003688970814377e-08,
      "loss": 0.0001,
      "step": 950
    },
    {
      "epoch": 4.910485933503836,
      "grad_norm": 0.0010263273253712944,
      "learning_rate": 7.216366640032668e-09,
      "loss": 0.0001,
      "step": 960
    },
    {
      "epoch": 4.961636828644501,
      "grad_norm": 0.060715491871115294,
      "learning_rate": 8.019900231881483e-10,
      "loss": 0.0001,
      "step": 970
    },
    {
      "epoch": 4.987212276214834,
      "step": 975,
      "total_flos": 261426857902080.0,
      "train_loss": 0.12791604029011908,
      "train_runtime": 13376.8379,
      "train_samples_per_second": 18.689,
      "train_steps_per_second": 0.073
    }
  ],
  "logging_steps": 10,
  "max_steps": 975,
  "num_input_tokens_seen": 0,
  "num_train_epochs": 5,
  "save_steps": 500,
  "stateful_callbacks": {
    "TrainerControl": {
      "args": {
        "should_epoch_stop": false,
        "should_evaluate": false,
        "should_log": false,
        "should_save": true,
        "should_training_stop": true
      },
      "attributes": {}
    }
  },
  "total_flos": 261426857902080.0,
  "train_batch_size": 16,
  "trial_name": null,
  "trial_params": null
}