zephyr7b_prompts / training_graph.json
Jon
Upload 8 files
eb03016
raw
history blame
28.1 kB
[
{
"current_steps": 0,
"loss": 3.6926,
"learning_rate": 0.0,
"epoch": 0.01
},
{
"current_steps": 1,
"loss": 3.0107,
"learning_rate": 0.0,
"epoch": 0.03
},
{
"current_steps": 2,
"loss": 3.0688,
"learning_rate": 0.0,
"epoch": 0.04
},
{
"current_steps": 3,
"loss": 4.0176,
"learning_rate": 0.0,
"epoch": 0.06
},
{
"current_steps": 4,
"loss": 4.0609,
"learning_rate": 0.0,
"epoch": 0.07
},
{
"current_steps": 5,
"loss": 3.754,
"learning_rate": 0.0,
"epoch": 0.09
},
{
"current_steps": 6,
"loss": 3.1981,
"learning_rate": 0.0,
"epoch": 0.1
},
{
"current_steps": 7,
"loss": 2.7138,
"learning_rate": 0.0,
"epoch": 0.11
},
{
"current_steps": 8,
"loss": 3.8803,
"learning_rate": 0.0,
"epoch": 0.13
},
{
"current_steps": 9,
"loss": 3.5793,
"learning_rate": 2.9999999999999997e-06,
"epoch": 0.14
},
{
"current_steps": 10,
"loss": 2.8589,
"learning_rate": 5.999999999999999e-06,
"epoch": 0.16
},
{
"current_steps": 11,
"loss": 2.6536,
"learning_rate": 8.999999999999999e-06,
"epoch": 0.17
},
{
"current_steps": 12,
"loss": 4.1845,
"learning_rate": 1.1999999999999999e-05,
"epoch": 0.19
},
{
"current_steps": 13,
"loss": 3.7447,
"learning_rate": 1.4999999999999999e-05,
"epoch": 0.2
},
{
"current_steps": 14,
"loss": 2.9558,
"learning_rate": 1.7999999999999997e-05,
"epoch": 0.21
},
{
"current_steps": 15,
"loss": 3.7552,
"learning_rate": 2.1e-05,
"epoch": 0.23
},
{
"current_steps": 16,
"loss": 3.7336,
"learning_rate": 2.3999999999999997e-05,
"epoch": 0.24
},
{
"current_steps": 17,
"loss": 2.484,
"learning_rate": 2.6999999999999996e-05,
"epoch": 0.26
},
{
"current_steps": 18,
"loss": 3.7203,
"learning_rate": 2.9999999999999997e-05,
"epoch": 0.27
},
{
"current_steps": 19,
"loss": 3.6129,
"learning_rate": 2.9999999999999997e-05,
"epoch": 0.29
},
{
"current_steps": 20,
"loss": 2.2497,
"learning_rate": 3.2999999999999996e-05,
"epoch": 0.3
},
{
"current_steps": 21,
"loss": 3.2732,
"learning_rate": 3.5999999999999994e-05,
"epoch": 0.31
},
{
"current_steps": 22,
"loss": 2.4979,
"learning_rate": 3.9e-05,
"epoch": 0.33
},
{
"current_steps": 23,
"loss": 3.043,
"learning_rate": 4.2e-05,
"epoch": 0.34
},
{
"current_steps": 24,
"loss": 3.2435,
"learning_rate": 4.4999999999999996e-05,
"epoch": 0.36
},
{
"current_steps": 25,
"loss": 3.3493,
"learning_rate": 4.7999999999999994e-05,
"epoch": 0.37
},
{
"current_steps": 26,
"loss": 2.6339,
"learning_rate": 5.1e-05,
"epoch": 0.39
},
{
"current_steps": 27,
"loss": 2.9486,
"learning_rate": 5.399999999999999e-05,
"epoch": 0.4
},
{
"current_steps": 28,
"loss": 2.3865,
"learning_rate": 5.6999999999999996e-05,
"epoch": 0.41
},
{
"current_steps": 29,
"loss": 2.3687,
"learning_rate": 5.9999999999999995e-05,
"epoch": 0.43
},
{
"current_steps": 30,
"loss": 2.3417,
"learning_rate": 6.299999999999999e-05,
"epoch": 0.44
},
{
"current_steps": 31,
"loss": 2.5443,
"learning_rate": 6.599999999999999e-05,
"epoch": 0.46
},
{
"current_steps": 32,
"loss": 2.4717,
"learning_rate": 6.9e-05,
"epoch": 0.47
},
{
"current_steps": 33,
"loss": 2.2291,
"learning_rate": 7.199999999999999e-05,
"epoch": 0.49
},
{
"current_steps": 34,
"loss": 2.4815,
"learning_rate": 7.5e-05,
"epoch": 0.5
},
{
"current_steps": 35,
"loss": 2.4468,
"learning_rate": 7.8e-05,
"epoch": 0.51
},
{
"current_steps": 36,
"loss": 2.4246,
"learning_rate": 8.1e-05,
"epoch": 0.53
},
{
"current_steps": 37,
"loss": 2.3154,
"learning_rate": 8.4e-05,
"epoch": 0.54
},
{
"current_steps": 38,
"loss": 1.9939,
"learning_rate": 8.699999999999999e-05,
"epoch": 0.56
},
{
"current_steps": 39,
"loss": 1.8099,
"learning_rate": 8.999999999999999e-05,
"epoch": 0.57
},
{
"current_steps": 40,
"loss": 1.865,
"learning_rate": 9.3e-05,
"epoch": 0.59
},
{
"current_steps": 41,
"loss": 1.4322,
"learning_rate": 9.599999999999999e-05,
"epoch": 0.6
},
{
"current_steps": 42,
"loss": 2.3252,
"learning_rate": 9.9e-05,
"epoch": 0.61
},
{
"current_steps": 43,
"loss": 1.7521,
"learning_rate": 0.000102,
"epoch": 0.63
},
{
"current_steps": 44,
"loss": 2.145,
"learning_rate": 0.00010499999999999999,
"epoch": 0.64
},
{
"current_steps": 45,
"loss": 1.6641,
"learning_rate": 0.00010799999999999998,
"epoch": 0.66
},
{
"current_steps": 46,
"loss": 1.8938,
"learning_rate": 0.00011099999999999999,
"epoch": 0.67
},
{
"current_steps": 47,
"loss": 1.5194,
"learning_rate": 0.00011399999999999999,
"epoch": 0.69
},
{
"current_steps": 48,
"loss": 1.4478,
"learning_rate": 0.000117,
"epoch": 0.7
},
{
"current_steps": 49,
"loss": 1.9414,
"learning_rate": 0.00011999999999999999,
"epoch": 0.71
},
{
"current_steps": 50,
"loss": 1.5601,
"learning_rate": 0.00012299999999999998,
"epoch": 0.73
},
{
"current_steps": 51,
"loss": 1.9015,
"learning_rate": 0.00012599999999999997,
"epoch": 0.74
},
{
"current_steps": 52,
"loss": 1.7384,
"learning_rate": 0.000129,
"epoch": 0.76
},
{
"current_steps": 53,
"loss": 1.7282,
"learning_rate": 0.00013199999999999998,
"epoch": 0.77
},
{
"current_steps": 54,
"loss": 1.6877,
"learning_rate": 0.000135,
"epoch": 0.79
},
{
"current_steps": 55,
"loss": 1.0647,
"learning_rate": 0.000138,
"epoch": 0.8
},
{
"current_steps": 56,
"loss": 1.755,
"learning_rate": 0.00014099999999999998,
"epoch": 0.81
},
{
"current_steps": 57,
"loss": 1.516,
"learning_rate": 0.00014399999999999998,
"epoch": 0.83
},
{
"current_steps": 58,
"loss": 1.3498,
"learning_rate": 0.000147,
"epoch": 0.84
},
{
"current_steps": 59,
"loss": 1.6789,
"learning_rate": 0.00015,
"epoch": 0.86
},
{
"current_steps": 60,
"loss": 1.3651,
"learning_rate": 0.00015299999999999998,
"epoch": 0.87
},
{
"current_steps": 61,
"loss": 1.2967,
"learning_rate": 0.000156,
"epoch": 0.89
},
{
"current_steps": 62,
"loss": 1.3998,
"learning_rate": 0.000159,
"epoch": 0.9
},
{
"current_steps": 63,
"loss": 1.1989,
"learning_rate": 0.000162,
"epoch": 0.91
},
{
"current_steps": 64,
"loss": 1.029,
"learning_rate": 0.000165,
"epoch": 0.93
},
{
"current_steps": 65,
"loss": 1.4441,
"learning_rate": 0.000168,
"epoch": 0.94
},
{
"current_steps": 66,
"loss": 1.1804,
"learning_rate": 0.00017099999999999998,
"epoch": 0.96
},
{
"current_steps": 67,
"loss": 1.2868,
"learning_rate": 0.00017399999999999997,
"epoch": 0.97
},
{
"current_steps": 68,
"loss": 1.1724,
"learning_rate": 0.00017699999999999997,
"epoch": 0.99
},
{
"current_steps": 69,
"loss": 1.4346,
"learning_rate": 0.00017999999999999998,
"epoch": 1.0
},
{
"current_steps": 70,
"loss": 1.0183,
"learning_rate": 0.00018299999999999998,
"epoch": 1.01
},
{
"current_steps": 71,
"loss": 1.1259,
"learning_rate": 0.000186,
"epoch": 1.03
},
{
"current_steps": 72,
"loss": 1.1713,
"learning_rate": 0.00018899999999999999,
"epoch": 1.04
},
{
"current_steps": 73,
"loss": 1.0773,
"learning_rate": 0.00019199999999999998,
"epoch": 1.06
},
{
"current_steps": 74,
"loss": 1.2956,
"learning_rate": 0.000195,
"epoch": 1.07
},
{
"current_steps": 75,
"loss": 0.9359,
"learning_rate": 0.000198,
"epoch": 1.09
},
{
"current_steps": 76,
"loss": 0.9838,
"learning_rate": 0.000201,
"epoch": 1.1
},
{
"current_steps": 77,
"loss": 0.9488,
"learning_rate": 0.000204,
"epoch": 1.11
},
{
"current_steps": 78,
"loss": 1.0748,
"learning_rate": 0.00020699999999999996,
"epoch": 1.13
},
{
"current_steps": 79,
"loss": 1.0914,
"learning_rate": 0.00020999999999999998,
"epoch": 1.14
},
{
"current_steps": 80,
"loss": 0.5364,
"learning_rate": 0.00021299999999999997,
"epoch": 1.16
},
{
"current_steps": 81,
"loss": 0.4909,
"learning_rate": 0.00021599999999999996,
"epoch": 1.17
},
{
"current_steps": 82,
"loss": 1.0176,
"learning_rate": 0.00021899999999999998,
"epoch": 1.19
},
{
"current_steps": 83,
"loss": 1.0543,
"learning_rate": 0.00022199999999999998,
"epoch": 1.2
},
{
"current_steps": 84,
"loss": 0.9092,
"learning_rate": 0.000225,
"epoch": 1.21
},
{
"current_steps": 85,
"loss": 0.5678,
"learning_rate": 0.00022799999999999999,
"epoch": 1.23
},
{
"current_steps": 86,
"loss": 0.8385,
"learning_rate": 0.00023099999999999998,
"epoch": 1.24
},
{
"current_steps": 87,
"loss": 1.6241,
"learning_rate": 0.000234,
"epoch": 1.26
},
{
"current_steps": 88,
"loss": 0.8294,
"learning_rate": 0.000237,
"epoch": 1.27
},
{
"current_steps": 89,
"loss": 1.2953,
"learning_rate": 0.00023999999999999998,
"epoch": 1.29
},
{
"current_steps": 90,
"loss": 0.5297,
"learning_rate": 0.000243,
"epoch": 1.3
},
{
"current_steps": 91,
"loss": 1.1592,
"learning_rate": 0.00024599999999999996,
"epoch": 1.31
},
{
"current_steps": 92,
"loss": 1.1055,
"learning_rate": 0.000249,
"epoch": 1.33
},
{
"current_steps": 93,
"loss": 0.9804,
"learning_rate": 0.00025199999999999995,
"epoch": 1.34
},
{
"current_steps": 94,
"loss": 0.8683,
"learning_rate": 0.00025499999999999996,
"epoch": 1.36
},
{
"current_steps": 95,
"loss": 0.9698,
"learning_rate": 0.000258,
"epoch": 1.37
},
{
"current_steps": 96,
"loss": 1.037,
"learning_rate": 0.000261,
"epoch": 1.39
},
{
"current_steps": 97,
"loss": 0.5479,
"learning_rate": 0.00026399999999999997,
"epoch": 1.4
},
{
"current_steps": 98,
"loss": 0.598,
"learning_rate": 0.000267,
"epoch": 1.41
},
{
"current_steps": 99,
"loss": 0.3627,
"learning_rate": 0.00027,
"epoch": 1.43
},
{
"current_steps": 100,
"loss": 0.8042,
"learning_rate": 0.00027299999999999997,
"epoch": 1.44
},
{
"current_steps": 101,
"loss": 1.0378,
"learning_rate": 0.000276,
"epoch": 1.46
},
{
"current_steps": 102,
"loss": 1.0192,
"learning_rate": 0.000279,
"epoch": 1.47
},
{
"current_steps": 103,
"loss": 1.4654,
"learning_rate": 0.00028199999999999997,
"epoch": 1.49
},
{
"current_steps": 104,
"loss": 1.0614,
"learning_rate": 0.000285,
"epoch": 1.5
},
{
"current_steps": 105,
"loss": 0.779,
"learning_rate": 0.00028799999999999995,
"epoch": 1.51
},
{
"current_steps": 106,
"loss": 1.0255,
"learning_rate": 0.00029099999999999997,
"epoch": 1.53
},
{
"current_steps": 107,
"loss": 1.0924,
"learning_rate": 0.000294,
"epoch": 1.54
},
{
"current_steps": 108,
"loss": 1.1016,
"learning_rate": 0.00029699999999999996,
"epoch": 1.56
},
{
"current_steps": 109,
"loss": 0.8748,
"learning_rate": 0.0003,
"epoch": 1.57
},
{
"current_steps": 110,
"loss": 0.8048,
"learning_rate": 0.00029727272727272724,
"epoch": 1.59
},
{
"current_steps": 111,
"loss": 1.0229,
"learning_rate": 0.0002945454545454545,
"epoch": 1.6
},
{
"current_steps": 112,
"loss": 0.9677,
"learning_rate": 0.0002918181818181818,
"epoch": 1.61
},
{
"current_steps": 113,
"loss": 0.5605,
"learning_rate": 0.00028909090909090904,
"epoch": 1.63
},
{
"current_steps": 114,
"loss": 0.9392,
"learning_rate": 0.00028636363636363636,
"epoch": 1.64
},
{
"current_steps": 115,
"loss": 1.0068,
"learning_rate": 0.0002836363636363636,
"epoch": 1.66
},
{
"current_steps": 116,
"loss": 0.898,
"learning_rate": 0.0002809090909090909,
"epoch": 1.67
},
{
"current_steps": 117,
"loss": 1.1297,
"learning_rate": 0.00027818181818181815,
"epoch": 1.69
},
{
"current_steps": 118,
"loss": 0.8696,
"learning_rate": 0.0002754545454545454,
"epoch": 1.7
},
{
"current_steps": 119,
"loss": 0.975,
"learning_rate": 0.0002727272727272727,
"epoch": 1.71
},
{
"current_steps": 120,
"loss": 1.107,
"learning_rate": 0.00027,
"epoch": 1.73
},
{
"current_steps": 121,
"loss": 1.1696,
"learning_rate": 0.0002672727272727272,
"epoch": 1.74
},
{
"current_steps": 122,
"loss": 0.7181,
"learning_rate": 0.00026454545454545453,
"epoch": 1.76
},
{
"current_steps": 123,
"loss": 0.4008,
"learning_rate": 0.0002618181818181818,
"epoch": 1.77
},
{
"current_steps": 124,
"loss": 0.6227,
"learning_rate": 0.00025909090909090907,
"epoch": 1.79
},
{
"current_steps": 125,
"loss": 1.4091,
"learning_rate": 0.00025636363636363633,
"epoch": 1.8
},
{
"current_steps": 126,
"loss": 0.7953,
"learning_rate": 0.0002536363636363636,
"epoch": 1.81
},
{
"current_steps": 127,
"loss": 1.1635,
"learning_rate": 0.00025090909090909086,
"epoch": 1.83
},
{
"current_steps": 128,
"loss": 0.6109,
"learning_rate": 0.0002481818181818182,
"epoch": 1.84
},
{
"current_steps": 129,
"loss": 0.7244,
"learning_rate": 0.00024545454545454545,
"epoch": 1.86
},
{
"current_steps": 130,
"loss": 1.0868,
"learning_rate": 0.0002427272727272727,
"epoch": 1.87
},
{
"current_steps": 131,
"loss": 1.1968,
"learning_rate": 0.00023999999999999998,
"epoch": 1.89
},
{
"current_steps": 132,
"loss": 1.109,
"learning_rate": 0.00023727272727272724,
"epoch": 1.9
},
{
"current_steps": 133,
"loss": 0.8744,
"learning_rate": 0.00023454545454545454,
"epoch": 1.91
},
{
"current_steps": 134,
"loss": 0.6971,
"learning_rate": 0.0002318181818181818,
"epoch": 1.93
},
{
"current_steps": 135,
"loss": 0.827,
"learning_rate": 0.00022909090909090907,
"epoch": 1.94
},
{
"current_steps": 136,
"loss": 0.8474,
"learning_rate": 0.00022636363636363633,
"epoch": 1.96
},
{
"current_steps": 137,
"loss": 0.8051,
"learning_rate": 0.00022363636363636363,
"epoch": 1.97
},
{
"current_steps": 138,
"loss": 0.8532,
"learning_rate": 0.0002209090909090909,
"epoch": 1.99
},
{
"current_steps": 139,
"loss": 0.8277,
"learning_rate": 0.00021818181818181816,
"epoch": 2.0
},
{
"current_steps": 140,
"loss": 0.3926,
"learning_rate": 0.00021545454545454542,
"epoch": 2.01
},
{
"current_steps": 141,
"loss": 0.5717,
"learning_rate": 0.00021272727272727272,
"epoch": 2.03
},
{
"current_steps": 142,
"loss": 0.6956,
"learning_rate": 0.00020999999999999998,
"epoch": 2.04
},
{
"current_steps": 143,
"loss": 0.6353,
"learning_rate": 0.00020727272727272725,
"epoch": 2.06
},
{
"current_steps": 144,
"loss": 0.4248,
"learning_rate": 0.0002045454545454545,
"epoch": 2.07
},
{
"current_steps": 145,
"loss": 0.6299,
"learning_rate": 0.0002018181818181818,
"epoch": 2.09
},
{
"current_steps": 146,
"loss": 0.3415,
"learning_rate": 0.0001990909090909091,
"epoch": 2.1
},
{
"current_steps": 147,
"loss": 0.4788,
"learning_rate": 0.00019636363636363634,
"epoch": 2.11
},
{
"current_steps": 148,
"loss": 0.6898,
"learning_rate": 0.00019363636363636363,
"epoch": 2.13
},
{
"current_steps": 149,
"loss": 0.4212,
"learning_rate": 0.0001909090909090909,
"epoch": 2.14
},
{
"current_steps": 150,
"loss": 0.6855,
"learning_rate": 0.0001881818181818182,
"epoch": 2.16
},
{
"current_steps": 151,
"loss": 0.4411,
"learning_rate": 0.00018545454545454543,
"epoch": 2.17
},
{
"current_steps": 152,
"loss": 0.4706,
"learning_rate": 0.00018272727272727272,
"epoch": 2.19
},
{
"current_steps": 153,
"loss": 0.6222,
"learning_rate": 0.00017999999999999998,
"epoch": 2.2
},
{
"current_steps": 154,
"loss": 0.6584,
"learning_rate": 0.00017727272727272728,
"epoch": 2.21
},
{
"current_steps": 155,
"loss": 0.6037,
"learning_rate": 0.00017454545454545452,
"epoch": 2.23
},
{
"current_steps": 156,
"loss": 0.6478,
"learning_rate": 0.0001718181818181818,
"epoch": 2.24
},
{
"current_steps": 157,
"loss": 0.4591,
"learning_rate": 0.00016909090909090907,
"epoch": 2.26
},
{
"current_steps": 158,
"loss": 0.4085,
"learning_rate": 0.00016636363636363637,
"epoch": 2.27
},
{
"current_steps": 159,
"loss": 0.6747,
"learning_rate": 0.0001636363636363636,
"epoch": 2.29
},
{
"current_steps": 160,
"loss": 0.4008,
"learning_rate": 0.0001609090909090909,
"epoch": 2.3
},
{
"current_steps": 161,
"loss": 0.671,
"learning_rate": 0.00015818181818181816,
"epoch": 2.31
},
{
"current_steps": 162,
"loss": 0.593,
"learning_rate": 0.00015545454545454546,
"epoch": 2.33
},
{
"current_steps": 163,
"loss": 0.4881,
"learning_rate": 0.0001527272727272727,
"epoch": 2.34
},
{
"current_steps": 164,
"loss": 0.6749,
"learning_rate": 0.00015,
"epoch": 2.36
},
{
"current_steps": 165,
"loss": 0.7904,
"learning_rate": 0.00014727272727272725,
"epoch": 2.37
},
{
"current_steps": 166,
"loss": 0.7036,
"learning_rate": 0.00014454545454545452,
"epoch": 2.39
},
{
"current_steps": 167,
"loss": 0.5043,
"learning_rate": 0.0001418181818181818,
"epoch": 2.4
},
{
"current_steps": 168,
"loss": 0.7245,
"learning_rate": 0.00013909090909090908,
"epoch": 2.41
},
{
"current_steps": 169,
"loss": 0.7101,
"learning_rate": 0.00013636363636363634,
"epoch": 2.43
},
{
"current_steps": 170,
"loss": 0.3898,
"learning_rate": 0.0001336363636363636,
"epoch": 2.44
},
{
"current_steps": 171,
"loss": 0.4639,
"learning_rate": 0.0001309090909090909,
"epoch": 2.46
},
{
"current_steps": 172,
"loss": 0.686,
"learning_rate": 0.00012818181818181817,
"epoch": 2.47
},
{
"current_steps": 173,
"loss": 0.4952,
"learning_rate": 0.00012545454545454543,
"epoch": 2.49
},
{
"current_steps": 174,
"loss": 0.2727,
"learning_rate": 0.00012272727272727272,
"epoch": 2.5
},
{
"current_steps": 175,
"loss": 0.3428,
"learning_rate": 0.00011999999999999999,
"epoch": 2.51
},
{
"current_steps": 176,
"loss": 0.253,
"learning_rate": 0.00011727272727272727,
"epoch": 2.53
},
{
"current_steps": 177,
"loss": 0.5778,
"learning_rate": 0.00011454545454545453,
"epoch": 2.54
},
{
"current_steps": 178,
"loss": 0.639,
"learning_rate": 0.00011181818181818181,
"epoch": 2.56
},
{
"current_steps": 179,
"loss": 0.7327,
"learning_rate": 0.00010909090909090908,
"epoch": 2.57
},
{
"current_steps": 180,
"loss": 0.4956,
"learning_rate": 0.00010636363636363636,
"epoch": 2.59
},
{
"current_steps": 181,
"loss": 0.364,
"learning_rate": 0.00010363636363636362,
"epoch": 2.6
},
{
"current_steps": 182,
"loss": 0.4497,
"learning_rate": 0.0001009090909090909,
"epoch": 2.61
},
{
"current_steps": 183,
"loss": 0.7653,
"learning_rate": 9.818181818181817e-05,
"epoch": 2.63
},
{
"current_steps": 184,
"loss": 0.6586,
"learning_rate": 9.545454545454545e-05,
"epoch": 2.64
},
{
"current_steps": 185,
"loss": 0.4404,
"learning_rate": 9.272727272727271e-05,
"epoch": 2.66
},
{
"current_steps": 186,
"loss": 0.7484,
"learning_rate": 8.999999999999999e-05,
"epoch": 2.67
},
{
"current_steps": 187,
"loss": 0.6176,
"learning_rate": 8.727272727272726e-05,
"epoch": 2.69
},
{
"current_steps": 188,
"loss": 0.7404,
"learning_rate": 8.454545454545454e-05,
"epoch": 2.7
},
{
"current_steps": 189,
"loss": 0.8124,
"learning_rate": 8.18181818181818e-05,
"epoch": 2.71
},
{
"current_steps": 190,
"loss": 0.6525,
"learning_rate": 7.909090909090908e-05,
"epoch": 2.73
},
{
"current_steps": 191,
"loss": 0.5986,
"learning_rate": 7.636363636363635e-05,
"epoch": 2.74
},
{
"current_steps": 192,
"loss": 0.294,
"learning_rate": 7.363636363636363e-05,
"epoch": 2.76
},
{
"current_steps": 193,
"loss": 0.2973,
"learning_rate": 7.09090909090909e-05,
"epoch": 2.77
},
{
"current_steps": 194,
"loss": 0.3589,
"learning_rate": 6.818181818181817e-05,
"epoch": 2.79
},
{
"current_steps": 195,
"loss": 0.78,
"learning_rate": 6.545454545454545e-05,
"epoch": 2.8
},
{
"current_steps": 196,
"loss": 0.4056,
"learning_rate": 6.272727272727272e-05,
"epoch": 2.81
},
{
"current_steps": 197,
"loss": 0.5843,
"learning_rate": 5.9999999999999995e-05,
"epoch": 2.83
},
{
"current_steps": 198,
"loss": 0.5745,
"learning_rate": 5.727272727272727e-05,
"epoch": 2.84
},
{
"current_steps": 199,
"loss": 0.2763,
"learning_rate": 5.454545454545454e-05,
"epoch": 2.86
},
{
"current_steps": 200,
"loss": 0.5494,
"learning_rate": 5.181818181818181e-05,
"epoch": 2.87
},
{
"current_steps": 201,
"loss": 0.4962,
"learning_rate": 4.9090909090909084e-05,
"epoch": 2.89
},
{
"current_steps": 202,
"loss": 0.5864,
"learning_rate": 4.6363636363636356e-05,
"epoch": 2.9
},
{
"current_steps": 203,
"loss": 0.3165,
"learning_rate": 4.363636363636363e-05,
"epoch": 2.91
},
{
"current_steps": 204,
"loss": 0.5969,
"learning_rate": 4.09090909090909e-05,
"epoch": 2.93
},
{
"current_steps": 205,
"loss": 0.3003,
"learning_rate": 3.8181818181818174e-05,
"epoch": 2.94
},
{
"current_steps": 206,
"loss": 0.6398,
"learning_rate": 3.545454545454545e-05,
"epoch": 2.96
},
{
"current_steps": 207,
"loss": 0.604,
"learning_rate": 3.2727272727272725e-05,
"epoch": 2.97
},
{
"current_steps": 208,
"loss": 0.6389,
"learning_rate": 2.9999999999999997e-05,
"epoch": 2.99
},
{
"current_steps": 209,
"loss": 0.5865,
"learning_rate": 2.727272727272727e-05,
"epoch": 3.0
},
{
"current_steps": 209,
"loss": 0.5865,
"learning_rate": 2.727272727272727e-05,
"epoch": 3.0
}
]