|
{ |
|
"best_metric": 0.7286937832832336, |
|
"best_model_checkpoint": "/media/xzuyn/c41d93ed-83a7-4577-9d38-5062b7e14438/xzuyn/Documents/AI_ROCM_5-6/LLaMa-LoRA-Tuner/data/lora_models/LLaMa-2-WizardLM-Uncensored-Tulu-Format-50K-7B-LoRA/checkpoint-20000", |
|
"epoch": 0.4, |
|
"global_step": 20000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 0.0001, |
|
"loss": 1.1136, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 9.997999599919985e-05, |
|
"loss": 0.9595, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.995999199839968e-05, |
|
"loss": 0.8936, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.993998799759952e-05, |
|
"loss": 0.9237, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.992038407681536e-05, |
|
"loss": 0.9005, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.99003800760152e-05, |
|
"loss": 0.8433, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.988037607521505e-05, |
|
"loss": 0.863, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.98603720744149e-05, |
|
"loss": 0.8634, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.984036807361474e-05, |
|
"loss": 0.8463, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.982036407281457e-05, |
|
"loss": 0.8527, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.980036007201441e-05, |
|
"loss": 0.8259, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.978035607121424e-05, |
|
"loss": 0.8473, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.976035207041409e-05, |
|
"loss": 0.8462, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.974034806961393e-05, |
|
"loss": 0.8609, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.972034406881376e-05, |
|
"loss": 0.8238, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.970034006801361e-05, |
|
"loss": 0.8212, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.968033606721344e-05, |
|
"loss": 0.8391, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.966033206641328e-05, |
|
"loss": 0.774, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.964032806561313e-05, |
|
"loss": 0.8149, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.962032406481296e-05, |
|
"loss": 0.8794, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.96003200640128e-05, |
|
"loss": 0.7747, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.958031606321264e-05, |
|
"loss": 0.8191, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.956031206241248e-05, |
|
"loss": 0.8219, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.954030806161232e-05, |
|
"loss": 0.8191, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.952030406081217e-05, |
|
"loss": 0.8662, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.950030006001201e-05, |
|
"loss": 0.787, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.948029605921184e-05, |
|
"loss": 0.8668, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.946029205841169e-05, |
|
"loss": 0.7817, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.944028805761153e-05, |
|
"loss": 0.8021, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.942028405681136e-05, |
|
"loss": 0.8324, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.940028005601121e-05, |
|
"loss": 0.8631, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.938027605521105e-05, |
|
"loss": 0.797, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.936027205441088e-05, |
|
"loss": 0.8224, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.934026805361073e-05, |
|
"loss": 0.8164, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.932026405281056e-05, |
|
"loss": 0.7975, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.930046009201841e-05, |
|
"loss": 0.8468, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.928045609121824e-05, |
|
"loss": 0.8169, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.926045209041809e-05, |
|
"loss": 0.7761, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.924044808961793e-05, |
|
"loss": 0.8042, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.922044408881776e-05, |
|
"loss": 0.8146, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.920044008801761e-05, |
|
"loss": 0.8243, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.918043608721744e-05, |
|
"loss": 0.8315, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.916043208641728e-05, |
|
"loss": 0.8014, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.914042808561713e-05, |
|
"loss": 0.8241, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.912042408481696e-05, |
|
"loss": 0.8233, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.91004200840168e-05, |
|
"loss": 0.7748, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.908041608321665e-05, |
|
"loss": 0.8197, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.906041208241649e-05, |
|
"loss": 0.8195, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.904040808161634e-05, |
|
"loss": 0.7864, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.902040408081617e-05, |
|
"loss": 0.7812, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.900040008001601e-05, |
|
"loss": 0.7966, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.898039607921584e-05, |
|
"loss": 0.793, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.896039207841569e-05, |
|
"loss": 0.8568, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.894038807761553e-05, |
|
"loss": 0.7614, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.892038407681536e-05, |
|
"loss": 0.8171, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.890038007601521e-05, |
|
"loss": 0.7649, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.888037607521505e-05, |
|
"loss": 0.8074, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.886037207441488e-05, |
|
"loss": 0.7591, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.884036807361473e-05, |
|
"loss": 0.8102, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.882036407281456e-05, |
|
"loss": 0.784, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.88003600720144e-05, |
|
"loss": 0.8367, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.878035607121425e-05, |
|
"loss": 0.8418, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.876035207041409e-05, |
|
"loss": 0.798, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.874034806961394e-05, |
|
"loss": 0.7928, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.872034406881377e-05, |
|
"loss": 0.8262, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.870034006801361e-05, |
|
"loss": 0.7965, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.868033606721346e-05, |
|
"loss": 0.7952, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.866033206641329e-05, |
|
"loss": 0.7987, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.864032806561313e-05, |
|
"loss": 0.7939, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.862032406481296e-05, |
|
"loss": 0.7484, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.860032006401281e-05, |
|
"loss": 0.819, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.858031606321265e-05, |
|
"loss": 0.747, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.856031206241248e-05, |
|
"loss": 0.8029, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.854030806161233e-05, |
|
"loss": 0.9013, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.852030406081216e-05, |
|
"loss": 0.8337, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.8500300060012e-05, |
|
"loss": 0.7927, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.848029605921185e-05, |
|
"loss": 0.8725, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.846029205841169e-05, |
|
"loss": 0.7865, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.844028805761154e-05, |
|
"loss": 0.7726, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.842048409681936e-05, |
|
"loss": 0.7868, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.840048009601921e-05, |
|
"loss": 0.7918, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.838047609521905e-05, |
|
"loss": 0.7924, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.836047209441888e-05, |
|
"loss": 0.8036, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.834046809361873e-05, |
|
"loss": 0.7844, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.832046409281857e-05, |
|
"loss": 0.8124, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.830046009201842e-05, |
|
"loss": 0.837, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.828045609121826e-05, |
|
"loss": 0.8315, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.826045209041809e-05, |
|
"loss": 0.7738, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.824044808961794e-05, |
|
"loss": 0.8164, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.822044408881777e-05, |
|
"loss": 0.7529, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.820044008801761e-05, |
|
"loss": 0.8037, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.818043608721746e-05, |
|
"loss": 0.8537, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.816043208641729e-05, |
|
"loss": 0.8241, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.814042808561713e-05, |
|
"loss": 0.7729, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.812042408481696e-05, |
|
"loss": 0.8116, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.81004200840168e-05, |
|
"loss": 0.8254, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.808041608321665e-05, |
|
"loss": 0.8285, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.806041208241648e-05, |
|
"loss": 0.7577, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.804040808161633e-05, |
|
"loss": 0.8024, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.802040408081616e-05, |
|
"loss": 0.8078, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"eval_loss": 0.740916907787323, |
|
"eval_runtime": 689.7578, |
|
"eval_samples_per_second": 7.211, |
|
"eval_steps_per_second": 0.902, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.8000400080016e-05, |
|
"loss": 0.7849, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.798039607921585e-05, |
|
"loss": 0.7931, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.796039207841569e-05, |
|
"loss": 0.7884, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.794038807761554e-05, |
|
"loss": 0.7949, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.792038407681537e-05, |
|
"loss": 0.7916, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.790038007601521e-05, |
|
"loss": 0.8072, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.788037607521506e-05, |
|
"loss": 0.7965, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.786037207441489e-05, |
|
"loss": 0.8039, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.784036807361473e-05, |
|
"loss": 0.8011, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.782036407281456e-05, |
|
"loss": 0.8014, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.78003600720144e-05, |
|
"loss": 0.7835, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.778035607121425e-05, |
|
"loss": 0.8469, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.776035207041408e-05, |
|
"loss": 0.8018, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.774034806961393e-05, |
|
"loss": 0.7839, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.772034406881376e-05, |
|
"loss": 0.8312, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.770054010802161e-05, |
|
"loss": 0.7886, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.768053610722146e-05, |
|
"loss": 0.8096, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.766053210642129e-05, |
|
"loss": 0.7806, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.764052810562113e-05, |
|
"loss": 0.8512, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.762052410482096e-05, |
|
"loss": 0.8475, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.76005201040208e-05, |
|
"loss": 0.749, |
|
"step": 12100 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.758051610322065e-05, |
|
"loss": 0.8245, |
|
"step": 12200 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.756051210242048e-05, |
|
"loss": 0.7966, |
|
"step": 12300 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.754050810162033e-05, |
|
"loss": 0.773, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.752050410082017e-05, |
|
"loss": 0.7876, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.750050010002001e-05, |
|
"loss": 0.7982, |
|
"step": 12600 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.748049609921986e-05, |
|
"loss": 0.7509, |
|
"step": 12700 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.746049209841969e-05, |
|
"loss": 0.7815, |
|
"step": 12800 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.744048809761953e-05, |
|
"loss": 0.8516, |
|
"step": 12900 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.742048409681937e-05, |
|
"loss": 0.7684, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.740048009601921e-05, |
|
"loss": 0.7889, |
|
"step": 13100 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.738047609521905e-05, |
|
"loss": 0.7882, |
|
"step": 13200 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.736047209441889e-05, |
|
"loss": 0.8196, |
|
"step": 13300 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.734046809361873e-05, |
|
"loss": 0.7889, |
|
"step": 13400 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.732046409281856e-05, |
|
"loss": 0.7931, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.73004600920184e-05, |
|
"loss": 0.8172, |
|
"step": 13600 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.728045609121825e-05, |
|
"loss": 0.8364, |
|
"step": 13700 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.726045209041808e-05, |
|
"loss": 0.822, |
|
"step": 13800 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.724044808961793e-05, |
|
"loss": 0.8546, |
|
"step": 13900 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.722044408881777e-05, |
|
"loss": 0.7977, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.720044008801761e-05, |
|
"loss": 0.7591, |
|
"step": 14100 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.718063612722545e-05, |
|
"loss": 0.7787, |
|
"step": 14200 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.716063212642529e-05, |
|
"loss": 0.8361, |
|
"step": 14300 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.714062812562513e-05, |
|
"loss": 0.8153, |
|
"step": 14400 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.712062412482496e-05, |
|
"loss": 0.776, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.71006201240248e-05, |
|
"loss": 0.7491, |
|
"step": 14600 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.708061612322465e-05, |
|
"loss": 0.7975, |
|
"step": 14700 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.70606121224245e-05, |
|
"loss": 0.7637, |
|
"step": 14800 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.704060812162434e-05, |
|
"loss": 0.7977, |
|
"step": 14900 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.702060412082417e-05, |
|
"loss": 0.8041, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.700060012002401e-05, |
|
"loss": 0.7997, |
|
"step": 15100 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.698059611922386e-05, |
|
"loss": 0.8221, |
|
"step": 15200 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.696059211842369e-05, |
|
"loss": 0.7657, |
|
"step": 15300 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.694058811762353e-05, |
|
"loss": 0.8091, |
|
"step": 15400 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.692058411682336e-05, |
|
"loss": 0.7692, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.690058011602321e-05, |
|
"loss": 0.7988, |
|
"step": 15600 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.688057611522305e-05, |
|
"loss": 0.7457, |
|
"step": 15700 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.686057211442288e-05, |
|
"loss": 0.8223, |
|
"step": 15800 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.684056811362273e-05, |
|
"loss": 0.8366, |
|
"step": 15900 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.682056411282256e-05, |
|
"loss": 0.7959, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.680076015203041e-05, |
|
"loss": 0.7944, |
|
"step": 16100 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.678075615123024e-05, |
|
"loss": 0.7645, |
|
"step": 16200 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.676075215043009e-05, |
|
"loss": 0.8111, |
|
"step": 16300 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.674074814962993e-05, |
|
"loss": 0.7907, |
|
"step": 16400 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.672074414882976e-05, |
|
"loss": 0.8108, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.670074014802961e-05, |
|
"loss": 0.8119, |
|
"step": 16600 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.668073614722945e-05, |
|
"loss": 0.8109, |
|
"step": 16700 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.666073214642928e-05, |
|
"loss": 0.7862, |
|
"step": 16800 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.664072814562913e-05, |
|
"loss": 0.7886, |
|
"step": 16900 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.662072414482897e-05, |
|
"loss": 0.786, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.660072014402882e-05, |
|
"loss": 0.8014, |
|
"step": 17100 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.658071614322865e-05, |
|
"loss": 0.7486, |
|
"step": 17200 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.65607121424285e-05, |
|
"loss": 0.7688, |
|
"step": 17300 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.654070814162834e-05, |
|
"loss": 0.7815, |
|
"step": 17400 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.652070414082817e-05, |
|
"loss": 0.7857, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.650070014002801e-05, |
|
"loss": 0.777, |
|
"step": 17600 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.648069613922786e-05, |
|
"loss": 0.8186, |
|
"step": 17700 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.646069213842769e-05, |
|
"loss": 0.7765, |
|
"step": 17800 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.644068813762753e-05, |
|
"loss": 0.8268, |
|
"step": 17900 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.642068413682736e-05, |
|
"loss": 0.7864, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.640068013602721e-05, |
|
"loss": 0.7741, |
|
"step": 18100 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.638067613522705e-05, |
|
"loss": 0.831, |
|
"step": 18200 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 9.636067213442688e-05, |
|
"loss": 0.7912, |
|
"step": 18300 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 9.634066813362673e-05, |
|
"loss": 0.7864, |
|
"step": 18400 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 9.632066413282656e-05, |
|
"loss": 0.7639, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 9.63006601320264e-05, |
|
"loss": 0.7874, |
|
"step": 18600 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 9.628065613122625e-05, |
|
"loss": 0.7642, |
|
"step": 18700 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.626065213042609e-05, |
|
"loss": 0.756, |
|
"step": 18800 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.624064812962594e-05, |
|
"loss": 0.8036, |
|
"step": 18900 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.622064412882577e-05, |
|
"loss": 0.7537, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.620064012802561e-05, |
|
"loss": 0.8193, |
|
"step": 19100 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.618063612722546e-05, |
|
"loss": 0.8036, |
|
"step": 19200 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.616063212642529e-05, |
|
"loss": 0.784, |
|
"step": 19300 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.614062812562513e-05, |
|
"loss": 0.7861, |
|
"step": 19400 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.612062412482496e-05, |
|
"loss": 0.7882, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.610062012402481e-05, |
|
"loss": 0.8048, |
|
"step": 19600 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.608061612322465e-05, |
|
"loss": 0.8544, |
|
"step": 19700 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 9.606061212242448e-05, |
|
"loss": 0.7827, |
|
"step": 19800 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 9.604060812162433e-05, |
|
"loss": 0.7641, |
|
"step": 19900 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 9.602060412082416e-05, |
|
"loss": 0.7146, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"eval_loss": 0.7286937832832336, |
|
"eval_runtime": 689.1627, |
|
"eval_samples_per_second": 7.217, |
|
"eval_steps_per_second": 0.903, |
|
"step": 20000 |
|
} |
|
], |
|
"max_steps": 500000, |
|
"num_train_epochs": 10, |
|
"total_flos": 2.798206695804764e+17, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|