|
{ |
|
"best_metric": 0.48128727078437805, |
|
"best_model_checkpoint": "data/Llama-31-8B_task-3_180-samples_config-3/checkpoint-340", |
|
"epoch": 27.0, |
|
"eval_steps": 500, |
|
"global_step": 459, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.058823529411764705, |
|
"grad_norm": 2.980715036392212, |
|
"learning_rate": 3.9215686274509804e-08, |
|
"loss": 2.6305, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.11764705882352941, |
|
"grad_norm": 1.5815191268920898, |
|
"learning_rate": 7.843137254901961e-08, |
|
"loss": 2.0227, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.23529411764705882, |
|
"grad_norm": 3.0497844219207764, |
|
"learning_rate": 1.5686274509803921e-07, |
|
"loss": 2.8101, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.35294117647058826, |
|
"grad_norm": 1.5842593908309937, |
|
"learning_rate": 2.3529411764705883e-07, |
|
"loss": 2.5918, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.47058823529411764, |
|
"grad_norm": 2.9371345043182373, |
|
"learning_rate": 3.1372549019607843e-07, |
|
"loss": 2.3864, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.5882352941176471, |
|
"grad_norm": 2.9547200202941895, |
|
"learning_rate": 3.921568627450981e-07, |
|
"loss": 2.6168, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.7058823529411765, |
|
"grad_norm": 2.5302109718322754, |
|
"learning_rate": 4.7058823529411767e-07, |
|
"loss": 2.2036, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.8235294117647058, |
|
"grad_norm": 3.466365337371826, |
|
"learning_rate": 5.490196078431373e-07, |
|
"loss": 2.7139, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.9411764705882353, |
|
"grad_norm": 1.7582124471664429, |
|
"learning_rate": 6.274509803921569e-07, |
|
"loss": 2.3708, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 2.4981932640075684, |
|
"eval_runtime": 31.8694, |
|
"eval_samples_per_second": 1.13, |
|
"eval_steps_per_second": 1.13, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 1.0588235294117647, |
|
"grad_norm": 1.3897138833999634, |
|
"learning_rate": 7.058823529411766e-07, |
|
"loss": 2.2296, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 1.1764705882352942, |
|
"grad_norm": 2.683563470840454, |
|
"learning_rate": 7.843137254901962e-07, |
|
"loss": 2.5726, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 1.2941176470588236, |
|
"grad_norm": 2.160468816757202, |
|
"learning_rate": 8.627450980392157e-07, |
|
"loss": 2.4421, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 1.4117647058823528, |
|
"grad_norm": 3.092349052429199, |
|
"learning_rate": 9.411764705882353e-07, |
|
"loss": 2.7441, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 1.5294117647058822, |
|
"grad_norm": 3.0546014308929443, |
|
"learning_rate": 1.019607843137255e-06, |
|
"loss": 2.6233, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 1.6470588235294117, |
|
"grad_norm": 2.6419599056243896, |
|
"learning_rate": 1.0980392156862745e-06, |
|
"loss": 2.6529, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 1.7647058823529411, |
|
"grad_norm": 1.2905603647232056, |
|
"learning_rate": 1.1764705882352942e-06, |
|
"loss": 2.1026, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 1.8823529411764706, |
|
"grad_norm": 2.6626884937286377, |
|
"learning_rate": 1.2549019607843137e-06, |
|
"loss": 2.5973, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 3.445452928543091, |
|
"learning_rate": 1.3333333333333334e-06, |
|
"loss": 2.4065, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 2.439713954925537, |
|
"eval_runtime": 31.87, |
|
"eval_samples_per_second": 1.13, |
|
"eval_steps_per_second": 1.13, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 2.1176470588235294, |
|
"grad_norm": 2.4814882278442383, |
|
"learning_rate": 1.4117647058823531e-06, |
|
"loss": 2.3601, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 2.235294117647059, |
|
"grad_norm": 3.0113446712493896, |
|
"learning_rate": 1.4901960784313726e-06, |
|
"loss": 2.3377, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 2.3529411764705883, |
|
"grad_norm": 2.5856285095214844, |
|
"learning_rate": 1.5686274509803923e-06, |
|
"loss": 2.24, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 2.4705882352941178, |
|
"grad_norm": 3.1034915447235107, |
|
"learning_rate": 1.6470588235294118e-06, |
|
"loss": 2.5231, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 2.588235294117647, |
|
"grad_norm": 3.499271869659424, |
|
"learning_rate": 1.7254901960784315e-06, |
|
"loss": 2.4463, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 2.7058823529411766, |
|
"grad_norm": 2.990689992904663, |
|
"learning_rate": 1.8039215686274512e-06, |
|
"loss": 2.4596, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 2.8235294117647056, |
|
"grad_norm": 2.712009906768799, |
|
"learning_rate": 1.8823529411764707e-06, |
|
"loss": 2.2987, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 2.9411764705882355, |
|
"grad_norm": 3.5523183345794678, |
|
"learning_rate": 1.96078431372549e-06, |
|
"loss": 2.3549, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 2.31467604637146, |
|
"eval_runtime": 31.8743, |
|
"eval_samples_per_second": 1.129, |
|
"eval_steps_per_second": 1.129, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 3.0588235294117645, |
|
"grad_norm": 3.0498082637786865, |
|
"learning_rate": 2.03921568627451e-06, |
|
"loss": 2.4567, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 3.176470588235294, |
|
"grad_norm": 2.239677906036377, |
|
"learning_rate": 2.1176470588235296e-06, |
|
"loss": 2.0235, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 3.2941176470588234, |
|
"grad_norm": 2.1120095252990723, |
|
"learning_rate": 2.196078431372549e-06, |
|
"loss": 2.2331, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 3.411764705882353, |
|
"grad_norm": 3.549705743789673, |
|
"learning_rate": 2.274509803921569e-06, |
|
"loss": 2.1111, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 3.5294117647058822, |
|
"grad_norm": 4.140646934509277, |
|
"learning_rate": 2.3529411764705885e-06, |
|
"loss": 2.4917, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 3.6470588235294117, |
|
"grad_norm": 3.1374101638793945, |
|
"learning_rate": 2.431372549019608e-06, |
|
"loss": 2.2287, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 3.764705882352941, |
|
"grad_norm": 2.198146104812622, |
|
"learning_rate": 2.5098039215686274e-06, |
|
"loss": 1.9659, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 3.8823529411764706, |
|
"grad_norm": 2.7118916511535645, |
|
"learning_rate": 2.5882352941176473e-06, |
|
"loss": 2.3314, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"grad_norm": 3.456117868423462, |
|
"learning_rate": 2.666666666666667e-06, |
|
"loss": 2.0578, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_loss": 2.085047483444214, |
|
"eval_runtime": 31.8792, |
|
"eval_samples_per_second": 1.129, |
|
"eval_steps_per_second": 1.129, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 4.117647058823529, |
|
"grad_norm": 2.2668004035949707, |
|
"learning_rate": 2.7450980392156867e-06, |
|
"loss": 2.1099, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 4.235294117647059, |
|
"grad_norm": 2.0983450412750244, |
|
"learning_rate": 2.8235294117647062e-06, |
|
"loss": 2.0241, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 4.352941176470588, |
|
"grad_norm": 1.6723711490631104, |
|
"learning_rate": 2.901960784313726e-06, |
|
"loss": 1.8905, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 4.470588235294118, |
|
"grad_norm": 3.0844905376434326, |
|
"learning_rate": 2.980392156862745e-06, |
|
"loss": 1.8637, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 4.588235294117647, |
|
"grad_norm": 3.0651304721832275, |
|
"learning_rate": 3.058823529411765e-06, |
|
"loss": 1.8883, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 4.705882352941177, |
|
"grad_norm": 2.8657171726226807, |
|
"learning_rate": 3.1372549019607846e-06, |
|
"loss": 1.8309, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 4.823529411764706, |
|
"grad_norm": 2.580749750137329, |
|
"learning_rate": 3.2156862745098045e-06, |
|
"loss": 1.7815, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 4.9411764705882355, |
|
"grad_norm": 2.4597244262695312, |
|
"learning_rate": 3.2941176470588236e-06, |
|
"loss": 1.8089, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_loss": 1.7080037593841553, |
|
"eval_runtime": 31.8717, |
|
"eval_samples_per_second": 1.13, |
|
"eval_steps_per_second": 1.13, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 5.0588235294117645, |
|
"grad_norm": 1.954211950302124, |
|
"learning_rate": 3.3725490196078435e-06, |
|
"loss": 1.6625, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 5.176470588235294, |
|
"grad_norm": 1.9714858531951904, |
|
"learning_rate": 3.450980392156863e-06, |
|
"loss": 1.5611, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 5.294117647058823, |
|
"grad_norm": 1.9127044677734375, |
|
"learning_rate": 3.529411764705883e-06, |
|
"loss": 1.5702, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 5.411764705882353, |
|
"grad_norm": 4.0653977394104, |
|
"learning_rate": 3.6078431372549024e-06, |
|
"loss": 1.5627, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 5.529411764705882, |
|
"grad_norm": 1.8669339418411255, |
|
"learning_rate": 3.6862745098039223e-06, |
|
"loss": 1.4645, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 5.647058823529412, |
|
"grad_norm": 2.6541359424591064, |
|
"learning_rate": 3.7647058823529414e-06, |
|
"loss": 1.3323, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 5.764705882352941, |
|
"grad_norm": 1.7167555093765259, |
|
"learning_rate": 3.843137254901962e-06, |
|
"loss": 1.3833, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 5.882352941176471, |
|
"grad_norm": 2.953925132751465, |
|
"learning_rate": 3.92156862745098e-06, |
|
"loss": 1.333, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"grad_norm": 1.2611362934112549, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 1.3018, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_loss": 1.2346911430358887, |
|
"eval_runtime": 31.8703, |
|
"eval_samples_per_second": 1.13, |
|
"eval_steps_per_second": 1.13, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 6.117647058823529, |
|
"grad_norm": 1.9918830394744873, |
|
"learning_rate": 4.07843137254902e-06, |
|
"loss": 1.1196, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 6.235294117647059, |
|
"grad_norm": 1.8971362113952637, |
|
"learning_rate": 4.15686274509804e-06, |
|
"loss": 1.1518, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 6.352941176470588, |
|
"grad_norm": 2.2451581954956055, |
|
"learning_rate": 4.235294117647059e-06, |
|
"loss": 1.0684, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 6.470588235294118, |
|
"grad_norm": 2.1816394329071045, |
|
"learning_rate": 4.313725490196079e-06, |
|
"loss": 0.9634, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 6.588235294117647, |
|
"grad_norm": 1.7623378038406372, |
|
"learning_rate": 4.392156862745098e-06, |
|
"loss": 0.9115, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 6.705882352941177, |
|
"grad_norm": 2.0872268676757812, |
|
"learning_rate": 4.4705882352941184e-06, |
|
"loss": 0.7509, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 6.823529411764706, |
|
"grad_norm": 1.5808390378952026, |
|
"learning_rate": 4.549019607843138e-06, |
|
"loss": 0.8618, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 6.9411764705882355, |
|
"grad_norm": 1.1898610591888428, |
|
"learning_rate": 4.627450980392157e-06, |
|
"loss": 1.0212, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_loss": 0.8016352653503418, |
|
"eval_runtime": 31.8718, |
|
"eval_samples_per_second": 1.13, |
|
"eval_steps_per_second": 1.13, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 7.0588235294117645, |
|
"grad_norm": 1.153451919555664, |
|
"learning_rate": 4.705882352941177e-06, |
|
"loss": 0.9469, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 7.176470588235294, |
|
"grad_norm": 1.1448363065719604, |
|
"learning_rate": 4.784313725490196e-06, |
|
"loss": 0.9065, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 7.294117647058823, |
|
"grad_norm": 1.0351287126541138, |
|
"learning_rate": 4.862745098039216e-06, |
|
"loss": 0.6538, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 7.411764705882353, |
|
"grad_norm": 0.9266816973686218, |
|
"learning_rate": 4.941176470588236e-06, |
|
"loss": 0.7569, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 7.529411764705882, |
|
"grad_norm": 0.869621992111206, |
|
"learning_rate": 5.019607843137255e-06, |
|
"loss": 0.6096, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 7.647058823529412, |
|
"grad_norm": 0.8324370980262756, |
|
"learning_rate": 5.098039215686274e-06, |
|
"loss": 0.4993, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 7.764705882352941, |
|
"grad_norm": 0.8301700353622437, |
|
"learning_rate": 5.176470588235295e-06, |
|
"loss": 0.4293, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 7.882352941176471, |
|
"grad_norm": 0.3732304275035858, |
|
"learning_rate": 5.254901960784314e-06, |
|
"loss": 0.632, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"grad_norm": 0.5899373888969421, |
|
"learning_rate": 5.333333333333334e-06, |
|
"loss": 0.4899, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_loss": 0.6475194096565247, |
|
"eval_runtime": 31.8748, |
|
"eval_samples_per_second": 1.129, |
|
"eval_steps_per_second": 1.129, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 8.117647058823529, |
|
"grad_norm": 0.6668117642402649, |
|
"learning_rate": 5.411764705882353e-06, |
|
"loss": 0.468, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 8.235294117647058, |
|
"grad_norm": 1.0659523010253906, |
|
"learning_rate": 5.4901960784313735e-06, |
|
"loss": 0.6599, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 8.352941176470589, |
|
"grad_norm": 0.787837028503418, |
|
"learning_rate": 5.568627450980393e-06, |
|
"loss": 0.8816, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 8.470588235294118, |
|
"grad_norm": 0.9166098237037659, |
|
"learning_rate": 5.6470588235294125e-06, |
|
"loss": 0.4608, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 8.588235294117647, |
|
"grad_norm": 0.8993943929672241, |
|
"learning_rate": 5.725490196078431e-06, |
|
"loss": 0.6013, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 8.705882352941176, |
|
"grad_norm": 0.5819871425628662, |
|
"learning_rate": 5.803921568627452e-06, |
|
"loss": 0.2645, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 8.823529411764707, |
|
"grad_norm": 0.6202341318130493, |
|
"learning_rate": 5.882352941176471e-06, |
|
"loss": 0.5653, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 8.941176470588236, |
|
"grad_norm": 0.737612247467041, |
|
"learning_rate": 5.96078431372549e-06, |
|
"loss": 0.6106, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_loss": 0.5890491008758545, |
|
"eval_runtime": 31.8673, |
|
"eval_samples_per_second": 1.13, |
|
"eval_steps_per_second": 1.13, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 9.058823529411764, |
|
"grad_norm": 0.8201161623001099, |
|
"learning_rate": 6.03921568627451e-06, |
|
"loss": 0.556, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 9.176470588235293, |
|
"grad_norm": 0.5118715167045593, |
|
"learning_rate": 6.11764705882353e-06, |
|
"loss": 0.521, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 9.294117647058824, |
|
"grad_norm": 2.472463369369507, |
|
"learning_rate": 6.19607843137255e-06, |
|
"loss": 0.5874, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 9.411764705882353, |
|
"grad_norm": 0.6849164366722107, |
|
"learning_rate": 6.274509803921569e-06, |
|
"loss": 0.5052, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 9.529411764705882, |
|
"grad_norm": 0.5705221891403198, |
|
"learning_rate": 6.352941176470589e-06, |
|
"loss": 0.4448, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 9.647058823529411, |
|
"grad_norm": 0.9986915588378906, |
|
"learning_rate": 6.431372549019609e-06, |
|
"loss": 0.4622, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 9.764705882352942, |
|
"grad_norm": 0.5936263203620911, |
|
"learning_rate": 6.5098039215686285e-06, |
|
"loss": 0.5221, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 9.882352941176471, |
|
"grad_norm": 0.4011266231536865, |
|
"learning_rate": 6.588235294117647e-06, |
|
"loss": 0.3669, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"grad_norm": 1.4033674001693726, |
|
"learning_rate": 6.666666666666667e-06, |
|
"loss": 0.5388, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_loss": 0.5729286670684814, |
|
"eval_runtime": 31.8736, |
|
"eval_samples_per_second": 1.129, |
|
"eval_steps_per_second": 1.129, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 10.117647058823529, |
|
"grad_norm": 0.7268795371055603, |
|
"learning_rate": 6.745098039215687e-06, |
|
"loss": 0.4123, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 10.235294117647058, |
|
"grad_norm": 0.534287691116333, |
|
"learning_rate": 6.8235294117647065e-06, |
|
"loss": 0.3577, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 10.352941176470589, |
|
"grad_norm": 0.46323052048683167, |
|
"learning_rate": 6.901960784313726e-06, |
|
"loss": 0.3157, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 10.470588235294118, |
|
"grad_norm": 0.4801871180534363, |
|
"learning_rate": 6.9803921568627454e-06, |
|
"loss": 0.5767, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 10.588235294117647, |
|
"grad_norm": 0.47244688868522644, |
|
"learning_rate": 7.058823529411766e-06, |
|
"loss": 0.5139, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 10.705882352941176, |
|
"grad_norm": 0.41083067655563354, |
|
"learning_rate": 7.137254901960785e-06, |
|
"loss": 0.5551, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 10.823529411764707, |
|
"grad_norm": 0.6211682558059692, |
|
"learning_rate": 7.215686274509805e-06, |
|
"loss": 0.5971, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 10.941176470588236, |
|
"grad_norm": 0.4872354567050934, |
|
"learning_rate": 7.294117647058823e-06, |
|
"loss": 0.7245, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"eval_loss": 0.5584969520568848, |
|
"eval_runtime": 31.8629, |
|
"eval_samples_per_second": 1.13, |
|
"eval_steps_per_second": 1.13, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 11.058823529411764, |
|
"grad_norm": 0.41652852296829224, |
|
"learning_rate": 7.372549019607845e-06, |
|
"loss": 0.3599, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 11.176470588235293, |
|
"grad_norm": 0.5096337795257568, |
|
"learning_rate": 7.450980392156863e-06, |
|
"loss": 0.5491, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 11.294117647058824, |
|
"grad_norm": 0.5131371021270752, |
|
"learning_rate": 7.529411764705883e-06, |
|
"loss": 0.6327, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 11.411764705882353, |
|
"grad_norm": 0.33539438247680664, |
|
"learning_rate": 7.607843137254902e-06, |
|
"loss": 0.4431, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 11.529411764705882, |
|
"grad_norm": 1.0883010625839233, |
|
"learning_rate": 7.686274509803923e-06, |
|
"loss": 0.5964, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 11.647058823529411, |
|
"grad_norm": 0.13403330743312836, |
|
"learning_rate": 7.764705882352941e-06, |
|
"loss": 0.1832, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 11.764705882352942, |
|
"grad_norm": 0.38453081250190735, |
|
"learning_rate": 7.84313725490196e-06, |
|
"loss": 0.5463, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 11.882352941176471, |
|
"grad_norm": 0.4325025677680969, |
|
"learning_rate": 7.92156862745098e-06, |
|
"loss": 0.3552, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"grad_norm": 0.26013830304145813, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 0.3568, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"eval_loss": 0.5533129572868347, |
|
"eval_runtime": 31.8552, |
|
"eval_samples_per_second": 1.13, |
|
"eval_steps_per_second": 1.13, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 12.117647058823529, |
|
"grad_norm": 0.5007168054580688, |
|
"learning_rate": 8.07843137254902e-06, |
|
"loss": 0.3615, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 12.235294117647058, |
|
"grad_norm": 0.4056757986545563, |
|
"learning_rate": 8.15686274509804e-06, |
|
"loss": 0.7152, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 12.352941176470589, |
|
"grad_norm": 0.2833678424358368, |
|
"learning_rate": 8.23529411764706e-06, |
|
"loss": 0.4289, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 12.470588235294118, |
|
"grad_norm": 0.3110792934894562, |
|
"learning_rate": 8.31372549019608e-06, |
|
"loss": 0.4519, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 12.588235294117647, |
|
"grad_norm": 0.3763163387775421, |
|
"learning_rate": 8.392156862745099e-06, |
|
"loss": 0.405, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 12.705882352941176, |
|
"grad_norm": 0.24176886677742004, |
|
"learning_rate": 8.470588235294118e-06, |
|
"loss": 0.4032, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 12.823529411764707, |
|
"grad_norm": 0.3594723343849182, |
|
"learning_rate": 8.549019607843138e-06, |
|
"loss": 0.3982, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 12.941176470588236, |
|
"grad_norm": 0.50929194688797, |
|
"learning_rate": 8.627450980392157e-06, |
|
"loss": 0.4165, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"eval_loss": 0.5352616310119629, |
|
"eval_runtime": 31.8489, |
|
"eval_samples_per_second": 1.13, |
|
"eval_steps_per_second": 1.13, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 13.058823529411764, |
|
"grad_norm": 0.32427483797073364, |
|
"learning_rate": 8.705882352941177e-06, |
|
"loss": 0.49, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 13.176470588235293, |
|
"grad_norm": 0.24497142434120178, |
|
"learning_rate": 8.784313725490196e-06, |
|
"loss": 0.3279, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 13.294117647058824, |
|
"grad_norm": 0.34544771909713745, |
|
"learning_rate": 8.862745098039216e-06, |
|
"loss": 0.4668, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 13.411764705882353, |
|
"grad_norm": 0.2216784805059433, |
|
"learning_rate": 8.941176470588237e-06, |
|
"loss": 0.3077, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 13.529411764705882, |
|
"grad_norm": 0.2850175201892853, |
|
"learning_rate": 9.019607843137256e-06, |
|
"loss": 0.3767, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 13.647058823529411, |
|
"grad_norm": 0.4109225571155548, |
|
"learning_rate": 9.098039215686276e-06, |
|
"loss": 0.4402, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 13.764705882352942, |
|
"grad_norm": 0.24095743894577026, |
|
"learning_rate": 9.176470588235294e-06, |
|
"loss": 0.3823, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 13.882352941176471, |
|
"grad_norm": 0.5017974972724915, |
|
"learning_rate": 9.254901960784315e-06, |
|
"loss": 0.5895, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"grad_norm": 0.45756474137306213, |
|
"learning_rate": 9.333333333333334e-06, |
|
"loss": 0.6226, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"eval_loss": 0.5420221090316772, |
|
"eval_runtime": 31.849, |
|
"eval_samples_per_second": 1.13, |
|
"eval_steps_per_second": 1.13, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 14.117647058823529, |
|
"grad_norm": 0.4417901039123535, |
|
"learning_rate": 9.411764705882354e-06, |
|
"loss": 0.545, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 14.235294117647058, |
|
"grad_norm": 0.3756580054759979, |
|
"learning_rate": 9.490196078431373e-06, |
|
"loss": 0.4739, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 14.352941176470589, |
|
"grad_norm": 0.4785820543766022, |
|
"learning_rate": 9.568627450980393e-06, |
|
"loss": 0.4149, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 14.470588235294118, |
|
"grad_norm": 0.444455087184906, |
|
"learning_rate": 9.647058823529412e-06, |
|
"loss": 0.508, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 14.588235294117647, |
|
"grad_norm": 0.3459950387477875, |
|
"learning_rate": 9.725490196078432e-06, |
|
"loss": 0.3752, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 14.705882352941176, |
|
"grad_norm": 0.276597797870636, |
|
"learning_rate": 9.803921568627451e-06, |
|
"loss": 0.3527, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 14.823529411764707, |
|
"grad_norm": 0.31315046548843384, |
|
"learning_rate": 9.882352941176472e-06, |
|
"loss": 0.396, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 14.941176470588236, |
|
"grad_norm": 0.009531126357614994, |
|
"learning_rate": 9.960784313725492e-06, |
|
"loss": 0.3284, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"eval_loss": 0.5025795102119446, |
|
"eval_runtime": 31.8517, |
|
"eval_samples_per_second": 1.13, |
|
"eval_steps_per_second": 1.13, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 15.058823529411764, |
|
"grad_norm": 0.5613787174224854, |
|
"learning_rate": 9.999995315380667e-06, |
|
"loss": 0.502, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 15.176470588235293, |
|
"grad_norm": 0.467040091753006, |
|
"learning_rate": 9.99995783847866e-06, |
|
"loss": 0.4182, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 15.294117647058824, |
|
"grad_norm": 0.41092485189437866, |
|
"learning_rate": 9.999882884955554e-06, |
|
"loss": 0.245, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 15.411764705882353, |
|
"grad_norm": 0.36648833751678467, |
|
"learning_rate": 9.99977045537315e-06, |
|
"loss": 0.3851, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 15.529411764705882, |
|
"grad_norm": 0.31992214918136597, |
|
"learning_rate": 9.999620550574155e-06, |
|
"loss": 0.4584, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 15.647058823529411, |
|
"grad_norm": 0.34673774242401123, |
|
"learning_rate": 9.999433171682158e-06, |
|
"loss": 0.3905, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 15.764705882352942, |
|
"grad_norm": 0.46017172932624817, |
|
"learning_rate": 9.999208320101643e-06, |
|
"loss": 0.4964, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 15.882352941176471, |
|
"grad_norm": 0.5312222838401794, |
|
"learning_rate": 9.998945997517957e-06, |
|
"loss": 0.3088, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"grad_norm": 0.5483973026275635, |
|
"learning_rate": 9.99864620589731e-06, |
|
"loss": 0.4813, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"eval_loss": 0.5214402079582214, |
|
"eval_runtime": 31.8457, |
|
"eval_samples_per_second": 1.13, |
|
"eval_steps_per_second": 1.13, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 16.11764705882353, |
|
"grad_norm": 0.40592437982559204, |
|
"learning_rate": 9.998308947486753e-06, |
|
"loss": 0.274, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 16.235294117647058, |
|
"grad_norm": 0.4894661009311676, |
|
"learning_rate": 9.997934224814173e-06, |
|
"loss": 0.361, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 16.352941176470587, |
|
"grad_norm": 0.5340806841850281, |
|
"learning_rate": 9.997522040688258e-06, |
|
"loss": 0.4403, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 16.470588235294116, |
|
"grad_norm": 0.6398447751998901, |
|
"learning_rate": 9.997072398198492e-06, |
|
"loss": 0.6416, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 16.58823529411765, |
|
"grad_norm": 0.4408299922943115, |
|
"learning_rate": 9.996585300715117e-06, |
|
"loss": 0.3814, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 16.705882352941178, |
|
"grad_norm": 0.45829370617866516, |
|
"learning_rate": 9.996060751889114e-06, |
|
"loss": 0.4494, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 16.823529411764707, |
|
"grad_norm": 0.5563424229621887, |
|
"learning_rate": 9.995498755652186e-06, |
|
"loss": 0.297, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 16.941176470588236, |
|
"grad_norm": 0.423210084438324, |
|
"learning_rate": 9.994899316216709e-06, |
|
"loss": 0.3015, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"eval_loss": 0.511587381362915, |
|
"eval_runtime": 31.8593, |
|
"eval_samples_per_second": 1.13, |
|
"eval_steps_per_second": 1.13, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 17.058823529411764, |
|
"grad_norm": 0.46335744857788086, |
|
"learning_rate": 9.994262438075713e-06, |
|
"loss": 0.4301, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 17.176470588235293, |
|
"grad_norm": 0.5282275676727295, |
|
"learning_rate": 9.993588126002848e-06, |
|
"loss": 0.3083, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 17.294117647058822, |
|
"grad_norm": 0.6960487961769104, |
|
"learning_rate": 9.992876385052346e-06, |
|
"loss": 0.5187, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 17.41176470588235, |
|
"grad_norm": 0.47753918170928955, |
|
"learning_rate": 9.992127220558976e-06, |
|
"loss": 0.2761, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 17.529411764705884, |
|
"grad_norm": 0.49851253628730774, |
|
"learning_rate": 9.991340638138022e-06, |
|
"loss": 0.2777, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 17.647058823529413, |
|
"grad_norm": 0.5102773904800415, |
|
"learning_rate": 9.990516643685222e-06, |
|
"loss": 0.4131, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 17.764705882352942, |
|
"grad_norm": 0.47633931040763855, |
|
"learning_rate": 9.98965524337673e-06, |
|
"loss": 0.5281, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 17.88235294117647, |
|
"grad_norm": 0.33498552441596985, |
|
"learning_rate": 9.988756443669081e-06, |
|
"loss": 0.2696, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"grad_norm": 0.5153418183326721, |
|
"learning_rate": 9.987820251299121e-06, |
|
"loss": 0.3513, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"eval_loss": 0.5070950388908386, |
|
"eval_runtime": 31.8477, |
|
"eval_samples_per_second": 1.13, |
|
"eval_steps_per_second": 1.13, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 18.11764705882353, |
|
"grad_norm": 0.3366803824901581, |
|
"learning_rate": 9.98684667328398e-06, |
|
"loss": 0.1873, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 18.235294117647058, |
|
"grad_norm": 0.6154054403305054, |
|
"learning_rate": 9.985835716921e-06, |
|
"loss": 0.391, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 18.352941176470587, |
|
"grad_norm": 0.6151386499404907, |
|
"learning_rate": 9.984787389787689e-06, |
|
"loss": 0.5524, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 18.470588235294116, |
|
"grad_norm": 0.34717944264411926, |
|
"learning_rate": 9.983701699741668e-06, |
|
"loss": 0.214, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 18.58823529411765, |
|
"grad_norm": 0.5958353877067566, |
|
"learning_rate": 9.982578654920601e-06, |
|
"loss": 0.4503, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 18.705882352941178, |
|
"grad_norm": 0.45278429985046387, |
|
"learning_rate": 9.981418263742148e-06, |
|
"loss": 0.2803, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 18.823529411764707, |
|
"grad_norm": 1.5949257612228394, |
|
"learning_rate": 9.980220534903889e-06, |
|
"loss": 0.3574, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 18.941176470588236, |
|
"grad_norm": 0.6842171549797058, |
|
"learning_rate": 9.978985477383264e-06, |
|
"loss": 0.3638, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 19.0, |
|
"eval_loss": 0.5485844016075134, |
|
"eval_runtime": 31.8529, |
|
"eval_samples_per_second": 1.13, |
|
"eval_steps_per_second": 1.13, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 19.058823529411764, |
|
"grad_norm": 0.5384336113929749, |
|
"learning_rate": 9.97771310043751e-06, |
|
"loss": 0.4372, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 19.176470588235293, |
|
"grad_norm": 18.550317764282227, |
|
"learning_rate": 9.97640341360358e-06, |
|
"loss": 0.3407, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 19.294117647058822, |
|
"grad_norm": 2.0581817626953125, |
|
"learning_rate": 9.975056426698094e-06, |
|
"loss": 0.2623, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 19.41176470588235, |
|
"grad_norm": 0.6005405187606812, |
|
"learning_rate": 9.973672149817232e-06, |
|
"loss": 0.258, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 19.529411764705884, |
|
"grad_norm": 0.8219459056854248, |
|
"learning_rate": 9.972250593336689e-06, |
|
"loss": 0.3986, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 19.647058823529413, |
|
"grad_norm": 1.8439754247665405, |
|
"learning_rate": 9.970791767911581e-06, |
|
"loss": 0.2887, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 19.764705882352942, |
|
"grad_norm": 2.892220973968506, |
|
"learning_rate": 9.96929568447637e-06, |
|
"loss": 0.3433, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 19.88235294117647, |
|
"grad_norm": 0.6804599761962891, |
|
"learning_rate": 9.967762354244778e-06, |
|
"loss": 0.3686, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"grad_norm": 0.8408982753753662, |
|
"learning_rate": 9.966191788709716e-06, |
|
"loss": 0.5246, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"eval_loss": 0.48128727078437805, |
|
"eval_runtime": 31.8515, |
|
"eval_samples_per_second": 1.13, |
|
"eval_steps_per_second": 1.13, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 20.11764705882353, |
|
"grad_norm": 0.5616675019264221, |
|
"learning_rate": 9.964583999643174e-06, |
|
"loss": 0.2985, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 20.235294117647058, |
|
"grad_norm": 0.4580283463001251, |
|
"learning_rate": 9.962938999096159e-06, |
|
"loss": 0.2537, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 20.352941176470587, |
|
"grad_norm": 0.857369601726532, |
|
"learning_rate": 9.961256799398584e-06, |
|
"loss": 0.2804, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 20.470588235294116, |
|
"grad_norm": 1.1619380712509155, |
|
"learning_rate": 9.95953741315919e-06, |
|
"loss": 0.519, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 20.58823529411765, |
|
"grad_norm": 0.5293999314308167, |
|
"learning_rate": 9.957780853265441e-06, |
|
"loss": 0.2812, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 20.705882352941178, |
|
"grad_norm": 0.6322659850120544, |
|
"learning_rate": 9.955987132883435e-06, |
|
"loss": 0.2476, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 20.823529411764707, |
|
"grad_norm": 0.6491472125053406, |
|
"learning_rate": 9.954156265457801e-06, |
|
"loss": 0.2483, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 20.941176470588236, |
|
"grad_norm": 0.869428813457489, |
|
"learning_rate": 9.952288264711601e-06, |
|
"loss": 0.4751, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 21.0, |
|
"eval_loss": 0.536905825138092, |
|
"eval_runtime": 31.8461, |
|
"eval_samples_per_second": 1.13, |
|
"eval_steps_per_second": 1.13, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 21.058823529411764, |
|
"grad_norm": 0.7692728042602539, |
|
"learning_rate": 9.950383144646221e-06, |
|
"loss": 0.4331, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 21.176470588235293, |
|
"grad_norm": 0.6134036779403687, |
|
"learning_rate": 9.948440919541277e-06, |
|
"loss": 0.3302, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 21.294117647058822, |
|
"grad_norm": 0.8637880682945251, |
|
"learning_rate": 9.946461603954499e-06, |
|
"loss": 0.5403, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 21.41176470588235, |
|
"grad_norm": 0.7665501832962036, |
|
"learning_rate": 9.944445212721619e-06, |
|
"loss": 0.2682, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 21.529411764705884, |
|
"grad_norm": 0.6994332671165466, |
|
"learning_rate": 9.942391760956277e-06, |
|
"loss": 0.4364, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 21.647058823529413, |
|
"grad_norm": 0.8093287944793701, |
|
"learning_rate": 9.940301264049885e-06, |
|
"loss": 0.2887, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 21.764705882352942, |
|
"grad_norm": 0.7530141472816467, |
|
"learning_rate": 9.938173737671531e-06, |
|
"loss": 0.2479, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 21.88235294117647, |
|
"grad_norm": 0.35253098607063293, |
|
"learning_rate": 9.936009197767847e-06, |
|
"loss": 0.1859, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 22.0, |
|
"grad_norm": 0.8288949728012085, |
|
"learning_rate": 9.933807660562898e-06, |
|
"loss": 0.2074, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 22.0, |
|
"eval_loss": 0.5176519751548767, |
|
"eval_runtime": 31.8467, |
|
"eval_samples_per_second": 1.13, |
|
"eval_steps_per_second": 1.13, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 22.11764705882353, |
|
"grad_norm": 0.7651407122612, |
|
"learning_rate": 9.931569142558057e-06, |
|
"loss": 0.3298, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 22.235294117647058, |
|
"grad_norm": 0.7411403059959412, |
|
"learning_rate": 9.929293660531889e-06, |
|
"loss": 0.3866, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 22.352941176470587, |
|
"grad_norm": 0.8898152709007263, |
|
"learning_rate": 9.926981231540007e-06, |
|
"loss": 0.3071, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 22.470588235294116, |
|
"grad_norm": 1.235049843788147, |
|
"learning_rate": 9.924631872914967e-06, |
|
"loss": 0.3644, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 22.58823529411765, |
|
"grad_norm": 0.9050986766815186, |
|
"learning_rate": 9.922245602266119e-06, |
|
"loss": 0.2616, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 22.705882352941178, |
|
"grad_norm": 0.986419141292572, |
|
"learning_rate": 9.919822437479488e-06, |
|
"loss": 0.2003, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 22.823529411764707, |
|
"grad_norm": 1.1197755336761475, |
|
"learning_rate": 9.91736239671763e-06, |
|
"loss": 0.3103, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 22.941176470588236, |
|
"grad_norm": 1.1346862316131592, |
|
"learning_rate": 9.91486549841951e-06, |
|
"loss": 0.2513, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 23.0, |
|
"eval_loss": 0.5108680129051208, |
|
"eval_runtime": 31.8499, |
|
"eval_samples_per_second": 1.13, |
|
"eval_steps_per_second": 1.13, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 23.058823529411764, |
|
"grad_norm": 0.3815372884273529, |
|
"learning_rate": 9.912331761300341e-06, |
|
"loss": 0.1238, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 23.176470588235293, |
|
"grad_norm": 0.004847167991101742, |
|
"learning_rate": 9.909761204351469e-06, |
|
"loss": 0.0666, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 23.294117647058822, |
|
"grad_norm": 0.9460958242416382, |
|
"learning_rate": 9.90715384684021e-06, |
|
"loss": 0.4942, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 23.41176470588235, |
|
"grad_norm": 1.0769481658935547, |
|
"learning_rate": 9.904509708309723e-06, |
|
"loss": 0.2983, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 23.529411764705884, |
|
"grad_norm": 0.9719020128250122, |
|
"learning_rate": 9.901828808578846e-06, |
|
"loss": 0.2081, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 23.647058823529413, |
|
"grad_norm": 1.7906395196914673, |
|
"learning_rate": 9.899111167741966e-06, |
|
"loss": 0.3697, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 23.764705882352942, |
|
"grad_norm": 1.5768994092941284, |
|
"learning_rate": 9.896356806168851e-06, |
|
"loss": 0.3432, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 23.88235294117647, |
|
"grad_norm": 0.9857346415519714, |
|
"learning_rate": 9.89356574450451e-06, |
|
"loss": 0.203, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 24.0, |
|
"grad_norm": 1.2476699352264404, |
|
"learning_rate": 9.890738003669029e-06, |
|
"loss": 0.3019, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 24.0, |
|
"eval_loss": 0.5099675059318542, |
|
"eval_runtime": 31.8826, |
|
"eval_samples_per_second": 1.129, |
|
"eval_steps_per_second": 1.129, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 24.11764705882353, |
|
"grad_norm": 1.1214749813079834, |
|
"learning_rate": 9.887873604857424e-06, |
|
"loss": 0.3208, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 24.235294117647058, |
|
"grad_norm": 1.2565882205963135, |
|
"learning_rate": 9.884972569539471e-06, |
|
"loss": 0.2897, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 24.352941176470587, |
|
"grad_norm": 0.9853368401527405, |
|
"learning_rate": 9.882034919459556e-06, |
|
"loss": 0.2316, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 24.470588235294116, |
|
"grad_norm": 1.3978081941604614, |
|
"learning_rate": 9.879060676636502e-06, |
|
"loss": 0.3282, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 24.58823529411765, |
|
"grad_norm": 1.6501458883285522, |
|
"learning_rate": 9.876049863363415e-06, |
|
"loss": 0.2489, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 24.705882352941178, |
|
"grad_norm": 1.0865875482559204, |
|
"learning_rate": 9.873002502207502e-06, |
|
"loss": 0.1472, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 24.823529411764707, |
|
"grad_norm": 1.1791576147079468, |
|
"learning_rate": 9.86991861600992e-06, |
|
"loss": 0.1987, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 24.941176470588236, |
|
"grad_norm": 1.4849443435668945, |
|
"learning_rate": 9.866798227885588e-06, |
|
"loss": 0.2039, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 25.0, |
|
"eval_loss": 0.542856752872467, |
|
"eval_runtime": 31.8504, |
|
"eval_samples_per_second": 1.13, |
|
"eval_steps_per_second": 1.13, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 25.058823529411764, |
|
"grad_norm": 1.064898133277893, |
|
"learning_rate": 9.863641361223025e-06, |
|
"loss": 0.2573, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 25.176470588235293, |
|
"grad_norm": 1.3044614791870117, |
|
"learning_rate": 9.860448039684169e-06, |
|
"loss": 0.2652, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 25.294117647058822, |
|
"grad_norm": 1.1536026000976562, |
|
"learning_rate": 9.857218287204204e-06, |
|
"loss": 0.1707, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 25.41176470588235, |
|
"grad_norm": 1.575131893157959, |
|
"learning_rate": 9.853952127991374e-06, |
|
"loss": 0.2226, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 25.529411764705884, |
|
"grad_norm": 1.8576124906539917, |
|
"learning_rate": 9.850649586526808e-06, |
|
"loss": 0.1639, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 25.647058823529413, |
|
"grad_norm": 1.8986772298812866, |
|
"learning_rate": 9.847310687564335e-06, |
|
"loss": 0.367, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 25.764705882352942, |
|
"grad_norm": 1.0811642408370972, |
|
"learning_rate": 9.843935456130295e-06, |
|
"loss": 0.2656, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 25.88235294117647, |
|
"grad_norm": 1.5638922452926636, |
|
"learning_rate": 9.840523917523354e-06, |
|
"loss": 0.1376, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 26.0, |
|
"grad_norm": 1.4766254425048828, |
|
"learning_rate": 9.83707609731432e-06, |
|
"loss": 0.228, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 26.0, |
|
"eval_loss": 0.5160675644874573, |
|
"eval_runtime": 31.8614, |
|
"eval_samples_per_second": 1.13, |
|
"eval_steps_per_second": 1.13, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 26.11764705882353, |
|
"grad_norm": 0.983343780040741, |
|
"learning_rate": 9.833592021345938e-06, |
|
"loss": 0.1742, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 26.235294117647058, |
|
"grad_norm": 1.3315422534942627, |
|
"learning_rate": 9.830071715732708e-06, |
|
"loss": 0.1853, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 26.352941176470587, |
|
"grad_norm": 0.6733668446540833, |
|
"learning_rate": 9.826515206860683e-06, |
|
"loss": 0.1493, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 26.470588235294116, |
|
"grad_norm": 1.4821362495422363, |
|
"learning_rate": 9.822922521387277e-06, |
|
"loss": 0.2319, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 26.58823529411765, |
|
"grad_norm": 3.0113203525543213, |
|
"learning_rate": 9.819293686241057e-06, |
|
"loss": 0.253, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 26.705882352941178, |
|
"grad_norm": 2.0522031784057617, |
|
"learning_rate": 9.81562872862155e-06, |
|
"loss": 0.1756, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 26.823529411764707, |
|
"grad_norm": 2.2314300537109375, |
|
"learning_rate": 9.811927675999035e-06, |
|
"loss": 0.1977, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 26.941176470588236, |
|
"grad_norm": 1.571890115737915, |
|
"learning_rate": 9.808190556114333e-06, |
|
"loss": 0.2127, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 27.0, |
|
"eval_loss": 0.5205699801445007, |
|
"eval_runtime": 31.8543, |
|
"eval_samples_per_second": 1.13, |
|
"eval_steps_per_second": 1.13, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 27.0, |
|
"step": 459, |
|
"total_flos": 2.8798901486092288e+17, |
|
"train_loss": 0.8010637635823689, |
|
"train_runtime": 10567.3661, |
|
"train_samples_per_second": 1.93, |
|
"train_steps_per_second": 0.241 |
|
} |
|
], |
|
"logging_steps": 2, |
|
"max_steps": 2550, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 150, |
|
"save_steps": 25, |
|
"stateful_callbacks": { |
|
"EarlyStoppingCallback": { |
|
"args": { |
|
"early_stopping_patience": 7, |
|
"early_stopping_threshold": 0.0 |
|
}, |
|
"attributes": { |
|
"early_stopping_patience_counter": 0 |
|
} |
|
}, |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 2.8798901486092288e+17, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|